{"commit":"47a3c103344a5d6a558d8a5fa581c3440ca791e6","old_file":"erpnext\/patches\/4_0\/countrywise_coa.py","new_file":"erpnext\/patches\/4_0\/countrywise_coa.py","old_contents":"# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors\n# License: GNU General Public License v3. See license.txt\n\nfrom __future__ import unicode_literals\nimport frappe\n\ndef execute():\n\tfrappe.reload_doc(\"setup\", 'doctype', \"company\")\n\tfrappe.reload_doc(\"accounts\", 'doctype', \"account\")\n\t\n\tfrappe.db.sql(\"\"\"update tabAccount set account_type='Fixed Asset' \n\t\twhere account_type='Fixed Asset Account'\"\"\")\n\t\t\n\tfor d in (('Asset', 'Debit', 'No'), ('Liability', 'Credit', 'No'), ('Expense', 'Debit', 'Yes'), \n\t\t\t('Income', 'Credit', 'Yes')):\n\t\tfrappe.db.sql(\"\"\"update `tabAccount` set root_type = %s \n\t\t\twhere debit_or_credit=%s and is_pl_account=%s\"\"\", d)\n\t\t\t\n\tfrappe.db.sql(\"\"\"update `tabAccount` set balance_must_be=debit_or_credit \n\t\twhere ifnull(allow_negative_balance, 0) = 0\"\"\")","new_contents":"# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors\n# License: GNU General Public License v3. See license.txt\n\nfrom __future__ import unicode_literals\nimport frappe\n\ndef execute():\n\tfrappe.reload_doc(\"setup\", 'doctype', \"company\")\n\tfrappe.reload_doc(\"accounts\", 'doctype', \"account\")\n\t\n\tfrappe.db.sql(\"\"\"update tabAccount set account_type='Fixed Asset' \n\t\twhere account_type='Fixed Asset Account'\"\"\")\n\n\tfrappe.db.sql(\"\"\"update `tabAccount` set report_type = \n\t\tif(is_pl_account=='Yes', 'Profit and Loss', 'Balance Sheet')\"\"\")\n\t\t\t\n\tfrappe.db.sql(\"\"\"update `tabAccount` set balance_must_be=debit_or_credit \n\t\twhere ifnull(allow_negative_balance, 0) = 0\"\"\")","subject":"Patch to update old accounts property","message":"Patch to update old accounts property\n","lang":"Python","license":"agpl-3.0","repos":"anandpdoshi\/erpnext,anandpdoshi\/erpnext,pombredanne\/erpnext,gmarke\/erpnext,Drooids\/erpnext,suyashphadtare\/test,suyashphadtare\/vestasi-erp-final,mbauskar\/omnitech-erpnext,treejames\/erpnext,tmimori\/erpnext,susuchina\/ERPNEXT,gangadharkadam\/smrterp,suyashphadtare\/vestasi-update-erp,fuhongliang\/erpnext,gsnbng\/erpnext,hatwar\/buyback-erpnext,Tejal011089\/paypal_erpnext,mahabuber\/erpnext,mahabuber\/erpnext,gangadhar-kadam\/verve_test_erp,mbauskar\/omnitech-demo-erpnext,Tejal011089\/trufil-erpnext,geekroot\/erpnext,gangadharkadam\/v5_erp,BhupeshGupta\/erpnext,gangadhar-kadam\/verve_live_erp,mbauskar\/omnitech-erpnext,netfirms\/erpnext,Tejal011089\/huntercamp_erpnext,suyashphadtare\/vestasi-erp-jan-end,hernad\/erpnext,aruizramon\/alec_erpnext,suyashphadtare\/vestasi-erp-jan-end,BhupeshGupta\/erpnext,indictranstech\/tele-erpnext,suyashphadtare\/vestasi-erp-jan-end,shft117\/SteckerApp,suyashphadtare\/vestasi-erp-1,gmarke\/erpnext,gangadhar-kadam\/laganerp,SPKian\/Testing,suyashphadtare\/sajil-erp,anandpdoshi\/erpnext,sheafferusa\/erpnext,indictranstech\/internal-erpnext,indictranstech\/phrerp,shft117\/SteckerApp,indictranstech\/erpnext,SPKian\/Testing,indictranstech\/fbd_erpnext,SPKian\/Testing2,suyashphadtare\/test,Tejal011089\/huntercamp_erpnext,MartinEnder\/erpnext-de,tmimori\/erpnext,gangadharkadam\/sher,gangadhar-kadam\/verve-erp,hernad\/erpnext,Suninus\/erpnext,gangadharkadam\/v6_erp,pawaranand\/phrerp,ThiagoGarciaAlves\/erpnext,indictranstech\/focal-erpnext,suyashphadtare\/vestasi-update-erp,mbauskar\/helpdesk-erpnext,rohitwaghchaure\/digitales_erpnext,gangadhar-kadam\/latestchurcherp,indictranstech\/fbd_erpnext,indictranstech\/reciphergroup-erpnext,rohitwaghchaure\/erpnext_smart,gangadharkadam\/v5_erp,gangadharkadam\/office_erp,indictranstech\/biggift-erpnext,gmarke\/erpnext,gangadhar-kadam\/hrerp,hatwar\/focal-erpnext,Tejal011089\/huntercamp_erpnext,indictranstech\/phrerp,rohitwaghchaure\/GenieManager-erpnext,meisterkleister\/erpnext,njmube\/erpnext,gangadhar-kadam\/verve_erp,gangadharkadam\/contributionerp,gangadhar-kadam\/smrterp,ShashaQin\/erpnext,indictranstech\/trufil-erpnext,netfirms\/erpnext,treejames\/erpnext,gangadharkadam\/contributionerp,ThiagoGarciaAlves\/erpnext,hatwar\/focal-erpnext,gangadharkadam\/saloon_erp_install,mbauskar\/internal-hr,indictranstech\/buyback-erp,indictranstech\/trufil-erpnext,indictranstech\/focal-erpnext,rohitwaghchaure\/GenieManager-erpnext,indictranstech\/reciphergroup-erpnext,Tejal011089\/digitales_erpnext,gangadhar-kadam\/helpdesk-erpnext,suyashphadtare\/sajil-final-erp,rohitwaghchaure\/erpnext-receipher,indictranstech\/focal-erpnext,mbauskar\/Das_Erpnext,gangadharkadam\/vlinkerp,suyashphadtare\/vestasi-erp-1,gangadhar-kadam\/hrerp,gangadharkadam\/letzerp,pawaranand\/phrerp,geekroot\/erpnext,hanselke\/erpnext-1,rohitwaghchaure\/erpnext_smart,indictranstech\/Das_Erpnext,gangadharkadam\/v6_erp,Tejal011089\/fbd_erpnext,mbauskar\/phrerp,SPKian\/Testing,gsnbng\/erpnext,Drooids\/erpnext,gangadharkadam\/saloon_erp_install,ShashaQin\/erpnext,gangadhar-kadam\/verve_erp,Tejal011089\/digitales_erpnext,rohitwaghchaure\/erpnext-receipher,4commerce-technologies-AG\/erpnext,saurabh6790\/test-erp,mbauskar\/sapphire-erpnext,mbauskar\/alec_frappe5_erpnext,gangadharkadam\/v6_erp,gangadhar-kadam\/laganerp,susuchina\/ERPNEXT,gangadhar-kadam\/verve_erp,indictranstech\/tele-erpnext,rohitwaghchaure\/New_Theme_Erp,gangadharkadam\/vlinkerp,mbauskar\/alec_frappe5_erpnext,fuhongliang\/erpnext,hatwar\/Das_erpnext,mahabuber\/erpnext,suyashphadtare\/vestasi-erp-jan-end,gangadhar-kadam\/verve-erp,indictranstech\/phrerp,Tejal011089\/paypal_erpnext,Suninus\/erpnext,gangadharkadam\/saloon_erp_install,SPKian\/Testing2,BhupeshGupta\/erpnext,indictranstech\/internal-erpnext,gangadharkadam\/office_erp,suyashphadtare\/vestasi-update-erp,rohitwaghchaure\/New_Theme_Erp,mbauskar\/omnitech-demo-erpnext,gangadharkadam\/v4_erp,indictranstech\/biggift-erpnext,gangadhar-kadam\/verve_erp,indictranstech\/tele-erpnext,rohitwaghchaure\/digitales_erpnext,indictranstech\/internal-erpnext,ShashaQin\/erpnext,njmube\/erpnext,gangadharkadam\/v4_erp,gangadhar-kadam\/laganerp,indictranstech\/buyback-erp,mbauskar\/omnitech-erpnext,pombredanne\/erpnext,suyashphadtare\/gd-erp,pawaranand\/phrerp,sagar30051991\/ozsmart-erp,hanselke\/erpnext-1,hatwar\/buyback-erpnext,sheafferusa\/erpnext,Tejal011089\/trufil-erpnext,indictranstech\/osmosis-erpnext,MartinEnder\/erpnext-de,anandpdoshi\/erpnext,Drooids\/erpnext,gangadharkadam\/vlinkerp,meisterkleister\/erpnext,gangadharkadam\/tailorerp,indictranstech\/vestasi-erpnext,gangadharkadam\/letzerp,gangadharkadam\/saloon_erp,gangadharkadam\/johnerp,Tejal011089\/paypal_erpnext,gangadharkadam\/letzerp,gangadharkadam\/contributionerp,hatwar\/focal-erpnext,mbauskar\/phrerp,mbauskar\/internal-hr,tmimori\/erpnext,mbauskar\/helpdesk-erpnext,hatwar\/Das_erpnext,indictranstech\/tele-erpnext,4commerce-technologies-AG\/erpnext,shitolepriya\/test-erp,fuhongliang\/erpnext,gangadharkadam\/v4_erp,indictranstech\/erpnext,gsnbng\/erpnext,susuchina\/ERPNEXT,Drooids\/erpnext,indictranstech\/vestasi-erpnext,gangadharkadam\/saloon_erp,gangadhar-kadam\/helpdesk-erpnext,suyashphadtare\/vestasi-erp-final,mbauskar\/omnitech-erpnext,gangadharkadam\/sher,saurabh6790\/test-erp,njmube\/erpnext,pombredanne\/erpnext,suyashphadtare\/sajil-final-erp,rohitwaghchaure\/New_Theme_Erp,hatwar\/buyback-erpnext,aruizramon\/alec_erpnext,gangadharkadam\/verveerp,gangadharkadam\/smrterp,rohitwaghchaure\/erpnext-receipher,SPKian\/Testing,gangadharkadam\/office_erp,gangadharkadam\/tailorerp,pombredanne\/erpnext,Tejal011089\/digitales_erpnext,SPKian\/Testing2,Tejal011089\/fbd_erpnext,rohitwaghchaure\/New_Theme_Erp,gangadhar-kadam\/latestchurcherp,geekroot\/erpnext,njmube\/erpnext,mbauskar\/omnitech-demo-erpnext,mbauskar\/phrerp,suyashphadtare\/sajil-final-erp,rohitwaghchaure\/GenieManager-erpnext,gangadharkadam\/letzerp,hatwar\/buyback-erpnext,gangadharkadam\/saloon_erp,aruizramon\/alec_erpnext,BhupeshGupta\/erpnext,netfirms\/erpnext,Tejal011089\/fbd_erpnext,rohitwaghchaure\/erpnext_smart,indictranstech\/phrerp,rohitwaghchaure\/digitales_erpnext,indictranstech\/osmosis-erpnext,indictranstech\/biggift-erpnext,Tejal011089\/osmosis_erpnext,indictranstech\/buyback-erp,saurabh6790\/test-erp,gangadharkadam\/sterp,Tejal011089\/trufil-erpnext,suyashphadtare\/sajil-erp,mbauskar\/internal-hr,indictranstech\/fbd_erpnext,dieface\/erpnext,gangadharkadam\/verveerp,suyashphadtare\/vestasi-erp-final,indictranstech\/reciphergroup-erpnext,gangadharkadam\/v4_erp,mbauskar\/Das_Erpnext,gmarke\/erpnext,gangadharkadam\/saloon_erp,sheafferusa\/erpnext,gangadharkadam\/v5_erp,hernad\/erpnext,treejames\/erpnext,saurabh6790\/test-erp,netfirms\/erpnext,indictranstech\/osmosis-erpnext,Suninus\/erpnext,rohitwaghchaure\/erpnext-receipher,gangadhar-kadam\/smrterp,indictranstech\/Das_Erpnext,indictranstech\/buyback-erp,shitolepriya\/test-erp,gangadhar-kadam\/verve-erp,mbauskar\/sapphire-erpnext,hatwar\/Das_erpnext,Tejal011089\/osmosis_erpnext,sagar30051991\/ozsmart-erp,indictranstech\/Das_Erpnext,Tejal011089\/paypal_erpnext,4commerce-technologies-AG\/erpnext,SPKian\/Testing2,gangadhar-kadam\/verve_test_erp,indictranstech\/reciphergroup-erpnext,Tejal011089\/osmosis_erpnext,indictranstech\/osmosis-erpnext,shitolepriya\/test-erp,indictranstech\/trufil-erpnext,rohitwaghchaure\/GenieManager-erpnext,gangadharkadam\/johnerp,gangadhar-kadam\/helpdesk-erpnext,rohitwaghchaure\/digitales_erpnext,gangadharkadam\/vlinkerp,mbauskar\/omnitech-demo-erpnext,treejames\/erpnext,gangadhar-kadam\/verve_test_erp,MartinEnder\/erpnext-de,suyashphadtare\/gd-erp,shft117\/SteckerApp,gangadharkadam\/verveerp,gangadharkadam\/verveerp,fuhongliang\/erpnext,indictranstech\/trufil-erpnext,indictranstech\/internal-erpnext,sheafferusa\/erpnext,indictranstech\/biggift-erpnext,gangadharkadam\/v6_erp,dieface\/erpnext,Aptitudetech\/ERPNext,indictranstech\/vestasi-erpnext,mbauskar\/Das_Erpnext,mbauskar\/alec_frappe5_erpnext,gangadhar-kadam\/verve_live_erp,mbauskar\/alec_frappe5_erpnext,meisterkleister\/erpnext,Tejal011089\/digitales_erpnext,suyashphadtare\/vestasi-erp-1,mbauskar\/sapphire-erpnext,mbauskar\/sapphire-erpnext,hanselke\/erpnext-1,gangadharkadam\/sterp,Suninus\/erpnext,sagar30051991\/ozsmart-erp,gangadhar-kadam\/verve_live_erp,gangadhar-kadam\/helpdesk-erpnext,indictranstech\/fbd_erpnext,mbauskar\/helpdesk-erpnext,susuchina\/ERPNEXT,mbauskar\/phrerp,mahabuber\/erpnext,hanselke\/erpnext-1,gangadharkadam\/contributionerp,mbauskar\/Das_Erpnext,indictranstech\/erpnext,aruizramon\/alec_erpnext,gangadhar-kadam\/verve_live_erp,gangadhar-kadam\/latestchurcherp,geekroot\/erpnext,ThiagoGarciaAlves\/erpnext,Tejal011089\/fbd_erpnext,Tejal011089\/trufil-erpnext,gangadhar-kadam\/latestchurcherp,gangadhar-kadam\/verve_test_erp,hernad\/erpnext,pawaranand\/phrerp,tmimori\/erpnext,indictranstech\/focal-erpnext,suyashphadtare\/test,dieface\/erpnext,ShashaQin\/erpnext,indictranstech\/erpnext,indictranstech\/Das_Erpnext,gangadharkadam\/saloon_erp_install,hatwar\/Das_erpnext,gsnbng\/erpnext,hatwar\/focal-erpnext,Tejal011089\/huntercamp_erpnext,shitolepriya\/test-erp,sagar30051991\/ozsmart-erp,Tejal011089\/osmosis_erpnext,shft117\/SteckerApp,suyashphadtare\/gd-erp,meisterkleister\/erpnext,ThiagoGarciaAlves\/erpnext,suyashphadtare\/sajil-erp,dieface\/erpnext,gangadharkadam\/v5_erp,suyashphadtare\/gd-erp,indictranstech\/vestasi-erpnext,mbauskar\/helpdesk-erpnext,MartinEnder\/erpnext-de"} {"commit":"eaa99e12ef4b868e825ffe01f4eb9319e439827b","old_file":"examples\/face_detection\/face_detect.py","new_file":"examples\/face_detection\/face_detect.py","old_contents":"from scannerpy import Database, DeviceType, Job\nfrom scannerpy.stdlib import pipelines\nimport subprocess\nimport cv2\nimport sys\nimport os.path\nsys.path.append(os.path.dirname(os.path.abspath(__file__)) + '\/..')\nimport util\n\nwith Database() as db:\n print('Ingesting video into Scanner ...')\n [input_table], _ = db.ingest_videos(\n [('example', util.download_video())], force=True)\n\n print('Detecting faces...')\n bboxes_table = pipelines.detect_faces(\n db, input_table, lambda t: t.all(), 'example_bboxes')\n\n print('Drawing faces onto video...')\n frame = input_table.as_op().all()\n bboxes = bboxes_table.as_op().all()\n out_frame = db.ops.DrawBox(frame = frame, bboxes = bboxes)\n job = Job(columns = [out_frame], name = 'example_bboxes_overlay')\n out_table = db.run(job, force=True)\n out_table.column('frame').save_mp4('example_faces')\n\n print('Successfully generated example_faces.mp4')\n","new_contents":"from scannerpy import Database, DeviceType, Job\nfrom scannerpy.stdlib import pipelines\nimport subprocess\nimport cv2\nimport sys\nimport os.path\nsys.path.append(os.path.dirname(os.path.abspath(__file__)) + '\/..')\nimport util\n\nmovie_path = util.download_video() if len(sys.argv) <= 1 else sys.argv[1]\nprint('Detecting faces in movie {}'.format(movie_path))\nmovie_name = os.path.splitext(os.path.basename(movie_path))[0]\n\nwith Database() as db:\n print('Ingesting video into Scanner ...')\n [input_table], _ = db.ingest_videos(\n [(movie_name, movie_path)], force=True)\n\n print('Detecting faces...')\n bboxes_table = pipelines.detect_faces(\n db, input_table, lambda t: t.all(),\n movie_name + '_bboxes')\n\n print('Drawing faces onto video...')\n frame = input_table.as_op().all()\n bboxes = bboxes_table.as_op().all()\n out_frame = db.ops.DrawBox(frame = frame, bboxes = bboxes)\n job = Job(columns = [out_frame], name = movie_name + '_bboxes_overlay')\n out_table = db.run(job, force=True)\n out_table.column('frame').save_mp4(movie_name + '_faces')\n\n print('Successfully generated {:s}_faces.mp4'.format(movie_name))\n","subject":"Update face detect example to take a path argument","message":"Update face detect example to take a path argument\n","lang":"Python","license":"apache-2.0","repos":"scanner-research\/scanner,scanner-research\/scanner,scanner-research\/scanner,scanner-research\/scanner"} {"commit":"421fd2d6fc0ffeaf35a99d426c7a1f9914e1da4e","old_file":"weaveserver\/services\/plugins\/__init__.py","new_file":"weaveserver\/services\/plugins\/__init__.py","old_contents":"from .service import PluginService\n\n__meta__ = {\n \"name\": \"Plugin Manager\",\n \"class\": PluginService,\n \"deps\": [\"messaging\", \"simpledb\", \"appmanager\"],\n \"config\": [\n {\n \"name\": \"plugins\",\n \"loaders\": [\n {\"type\": \"env\"},\n {\"type\": \"sysvarfile\"}\n ]\n }\n ]\n}\n","new_contents":"from .service import PluginService\n\n__meta__ = {\n \"name\": \"Plugin Manager\",\n \"class\": PluginService,\n \"deps\": [\"messaging\", \"simpledb\", \"appmanager\", \"http\"],\n \"config\": [\n {\n \"name\": \"plugins\",\n \"loaders\": [\n {\"type\": \"env\"},\n {\"type\": \"sysvarfile\"}\n ]\n }\n ]\n}\n","subject":"Add http dependency for plugins.","message":"Add http dependency for plugins.\n","lang":"Python","license":"mit","repos":"supersaiyanmode\/HomePiServer,supersaiyanmode\/HomePiServer,supersaiyanmode\/HomePiServer"} {"commit":"627d79ae4950338c8a5a0d75bae244c9c0374d4a","old_file":"friendlyurls\/admin.py","new_file":"friendlyurls\/admin.py","old_contents":"from django.contrib import admin\nfrom friendlyurls.models import *\n\nclass UrlMappingAdmin(admin.ModelAdmin):\n list_display = ('friendly_path', 'resolved_url', 'content_type', 'object')\nadmin.site.register(UrlMapping, UrlMappingAdmin)\n\n","new_contents":"from django.contrib import admin\nfrom friendlyurls.models import *\n\nclass UrlMappingAdmin(admin.ModelAdmin):\n list_display = ('friendly_path', 'resolved_url', 'content_type', 'object')\n search_fields = ('friendly_path','content_type__name')\nadmin.site.register(UrlMapping, UrlMappingAdmin)\n\n","subject":"Allow searching of vanity urls","message":"Allow searching of vanity urls\n","lang":"Python","license":"bsd-3-clause","repos":"tachang\/django_friendlyurls"} {"commit":"d0126b7e31c69ea7dd5cda4b9a3e931f5b8a8fbf","old_file":"rest_framework\/authtoken\/views.py","new_file":"rest_framework\/authtoken\/views.py","old_contents":"from rest_framework.views import APIView\nfrom rest_framework import parsers\nfrom rest_framework import renderers\nfrom rest_framework.response import Response\nfrom rest_framework.authtoken.models import Token\nfrom rest_framework.authtoken.serializers import AuthTokenSerializer\n\n\nclass ObtainAuthToken(APIView):\n throttle_classes = ()\n permission_classes = ()\n parser_classes = (parsers.FormParser, parsers.MultiPartParser, parsers.JSONParser,)\n renderer_classes = (renderers.JSONRenderer,)\n\n def post(self, request):\n serializer = AuthTokenSerializer(data=request.data)\n serializer.is_valid(raise_exception=True)\n user = serializer.validated_data['user']\n token, created = Token.objects.get_or_create(user=user)\n return Response({'token': token.key})\n\n\nobtain_auth_token = ObtainAuthToken.as_view()\n","new_contents":"from rest_framework.views import APIView\nfrom rest_framework import parsers\nfrom rest_framework import renderers\nfrom rest_framework.response import Response\nfrom rest_framework.authtoken.models import Token\nfrom rest_framework.authtoken.serializers import AuthTokenSerializer\n\n\nclass ObtainAuthToken(APIView):\n throttle_classes = ()\n permission_classes = ()\n parser_classes = (parsers.FormParser, parsers.MultiPartParser, parsers.JSONParser,)\n renderer_classes = (renderers.JSONRenderer,)\n serializer_class = AuthTokenSerializer\n\n def post(self, request):\n serializer = self.serializer_class(data=request.data)\n serializer.is_valid(raise_exception=True)\n user = serializer.validated_data['user']\n token, created = Token.objects.get_or_create(user=user)\n return Response({'token': token.key})\n\n\nobtain_auth_token = ObtainAuthToken.as_view()\n","subject":"Set serializer_class on ObtainAuthToken view","message":"Set serializer_class on ObtainAuthToken view\n","lang":"Python","license":"bsd-2-clause","repos":"kennydude\/django-rest-framework,mgaitan\/django-rest-framework,wedaly\/django-rest-framework,arpheno\/django-rest-framework,callorico\/django-rest-framework,johnraz\/django-rest-framework,nryoung\/django-rest-framework,krinart\/django-rest-framework,gregmuellegger\/django-rest-framework,akalipetis\/django-rest-framework,tigeraniya\/django-rest-framework,ebsaral\/django-rest-framework,cheif\/django-rest-framework,jpulec\/django-rest-framework,ajaali\/django-rest-framework,brandoncazander\/django-rest-framework,leeahoward\/django-rest-framework,atombrella\/django-rest-framework,jtiai\/django-rest-framework,fishky\/django-rest-framework,kgeorgy\/django-rest-framework,gregmuellegger\/django-rest-framework,ambivalentno\/django-rest-framework,tomchristie\/django-rest-framework,lubomir\/django-rest-framework,abdulhaq-e\/django-rest-framework,johnraz\/django-rest-framework,d0ugal\/django-rest-framework,iheitlager\/django-rest-framework,justanr\/django-rest-framework,jtiai\/django-rest-framework,simudream\/django-rest-framework,sehmaschine\/django-rest-framework,fishky\/django-rest-framework,rubendura\/django-rest-framework,kylefox\/django-rest-framework,rafaelang\/django-rest-framework,ashishfinoit\/django-rest-framework,simudream\/django-rest-framework,rhblind\/django-rest-framework,callorico\/django-rest-framework,uruz\/django-rest-framework,jpadilla\/django-rest-framework,andriy-s\/django-rest-framework,jerryhebert\/django-rest-framework,hunter007\/django-rest-framework,jpulec\/django-rest-framework,elim\/django-rest-framework,kgeorgy\/django-rest-framework,wzbozon\/django-rest-framework,callorico\/django-rest-framework,linovia\/django-rest-framework,ticosax\/django-rest-framework,wzbozon\/django-rest-framework,wwj718\/django-rest-framework,HireAnEsquire\/django-rest-framework,maryokhin\/django-rest-framework,HireAnEsquire\/django-rest-framework,damycra\/django-rest-framework,jness\/django-rest-framework,jpadilla\/django-rest-framework,agconti\/django-rest-framework,justanr\/django-rest-framework,werthen\/django-rest-framework,wangpanjun\/django-rest-framework,cheif\/django-rest-framework,mgaitan\/django-rest-framework,kylefox\/django-rest-framework,sbellem\/django-rest-framework,zeldalink0515\/django-rest-framework,tomchristie\/django-rest-framework,justanr\/django-rest-framework,qsorix\/django-rest-framework,delinhabit\/django-rest-framework,cyberj\/django-rest-framework,brandoncazander\/django-rest-framework,yiyocx\/django-rest-framework,lubomir\/django-rest-framework,tcroiset\/django-rest-framework,kgeorgy\/django-rest-framework,edx\/django-rest-framework,ebsaral\/django-rest-framework,dmwyatt\/django-rest-framework,brandoncazander\/django-rest-framework,yiyocx\/django-rest-framework,uploadcare\/django-rest-framework,potpath\/django-rest-framework,waytai\/django-rest-framework,damycra\/django-rest-framework,agconti\/django-rest-framework,AlexandreProenca\/django-rest-framework,alacritythief\/django-rest-framework,ebsaral\/django-rest-framework,akalipetis\/django-rest-framework,douwevandermeij\/django-rest-framework,James1345\/django-rest-framework,sehmaschine\/django-rest-framework,potpath\/django-rest-framework,uruz\/django-rest-framework,jerryhebert\/django-rest-framework,delinhabit\/django-rest-framework,hunter007\/django-rest-framework,davesque\/django-rest-framework,leeahoward\/django-rest-framework,thedrow\/django-rest-framework-1,edx\/django-rest-framework,maryokhin\/django-rest-framework,linovia\/django-rest-framework,thedrow\/django-rest-framework-1,abdulhaq-e\/django-rest-framework,ticosax\/django-rest-framework,kylefox\/django-rest-framework,jerryhebert\/django-rest-framework,wangpanjun\/django-rest-framework,iheitlager\/django-rest-framework,yiyocx\/django-rest-framework,dmwyatt\/django-rest-framework,douwevandermeij\/django-rest-framework,potpath\/django-rest-framework,canassa\/django-rest-framework,canassa\/django-rest-framework,arpheno\/django-rest-framework,d0ugal\/django-rest-framework,d0ugal\/django-rest-framework,hnakamur\/django-rest-framework,rhblind\/django-rest-framework,rafaelcaricio\/django-rest-framework,kezabelle\/django-rest-framework,thedrow\/django-rest-framework-1,bluedazzle\/django-rest-framework,waytai\/django-rest-framework,ambivalentno\/django-rest-framework,krinart\/django-rest-framework,damycra\/django-rest-framework,andriy-s\/django-rest-framework,raphaelmerx\/django-rest-framework,VishvajitP\/django-rest-framework,rafaelcaricio\/django-rest-framework,uploadcare\/django-rest-framework,rafaelcaricio\/django-rest-framework,cyberj\/django-rest-framework,ossanna16\/django-rest-framework,paolopaolopaolo\/django-rest-framework,rubendura\/django-rest-framework,sbellem\/django-rest-framework,aericson\/django-rest-framework,wwj718\/django-rest-framework,werthen\/django-rest-framework,krinart\/django-rest-framework,VishvajitP\/django-rest-framework,HireAnEsquire\/django-rest-framework,uruz\/django-rest-framework,kezabelle\/django-rest-framework,abdulhaq-e\/django-rest-framework,hnarayanan\/django-rest-framework,delinhabit\/django-rest-framework,akalipetis\/django-rest-framework,aericson\/django-rest-framework,tomchristie\/django-rest-framework,ossanna16\/django-rest-framework,ezheidtmann\/django-rest-framework,tigeraniya\/django-rest-framework,hnakamur\/django-rest-framework,buptlsl\/django-rest-framework,vstoykov\/django-rest-framework,tcroiset\/django-rest-framework,nryoung\/django-rest-framework,ambivalentno\/django-rest-framework,rhblind\/django-rest-framework,raphaelmerx\/django-rest-framework,vstoykov\/django-rest-framework,jpadilla\/django-rest-framework,bluedazzle\/django-rest-framework,paolopaolopaolo\/django-rest-framework,paolopaolopaolo\/django-rest-framework,zeldalink0515\/django-rest-framework,MJafarMashhadi\/django-rest-framework,antonyc\/django-rest-framework,MJafarMashhadi\/django-rest-framework,davesque\/django-rest-framework,raphaelmerx\/django-rest-framework,ticosax\/django-rest-framework,sehmaschine\/django-rest-framework,xiaotangyuan\/django-rest-framework,hunter007\/django-rest-framework,YBJAY00000\/django-rest-framework,iheitlager\/django-rest-framework,simudream\/django-rest-framework,hnakamur\/django-rest-framework,xiaotangyuan\/django-rest-framework,rubendura\/django-rest-framework,rafaelang\/django-rest-framework,wwj718\/django-rest-framework,tigeraniya\/django-rest-framework,ajaali\/django-rest-framework,jness\/django-rest-framework,MJafarMashhadi\/django-rest-framework,ashishfinoit\/django-rest-framework,cheif\/django-rest-framework,adambain-vokal\/django-rest-framework,ezheidtmann\/django-rest-framework,canassa\/django-rest-framework,sheppard\/django-rest-framework,YBJAY00000\/django-rest-framework,johnraz\/django-rest-framework,ezheidtmann\/django-rest-framework,lubomir\/django-rest-framework,andriy-s\/django-rest-framework,davesque\/django-rest-framework,uploadcare\/django-rest-framework,YBJAY00000\/django-rest-framework,buptlsl\/django-rest-framework,rafaelang\/django-rest-framework,ossanna16\/django-rest-framework,atombrella\/django-rest-framework,douwevandermeij\/django-rest-framework,sheppard\/django-rest-framework,VishvajitP\/django-rest-framework,sheppard\/django-rest-framework,werthen\/django-rest-framework,sbellem\/django-rest-framework,nhorelik\/django-rest-framework,kezabelle\/django-rest-framework,jpulec\/django-rest-framework,mgaitan\/django-rest-framework,jtiai\/django-rest-framework,alacritythief\/django-rest-framework,qsorix\/django-rest-framework,kennydude\/django-rest-framework,gregmuellegger\/django-rest-framework,jness\/django-rest-framework,vstoykov\/django-rest-framework,fishky\/django-rest-framework,nhorelik\/django-rest-framework,James1345\/django-rest-framework,alacritythief\/django-rest-framework,elim\/django-rest-framework,ajaali\/django-rest-framework,elim\/django-rest-framework,antonyc\/django-rest-framework,nryoung\/django-rest-framework,ashishfinoit\/django-rest-framework,wzbozon\/django-rest-framework,AlexandreProenca\/django-rest-framework,maryokhin\/django-rest-framework,pombredanne\/django-rest-framework,edx\/django-rest-framework,leeahoward\/django-rest-framework,buptlsl\/django-rest-framework,pombredanne\/django-rest-framework,pombredanne\/django-rest-framework,aericson\/django-rest-framework,linovia\/django-rest-framework,adambain-vokal\/django-rest-framework,hnarayanan\/django-rest-framework,bluedazzle\/django-rest-framework,xiaotangyuan\/django-rest-framework,agconti\/django-rest-framework,James1345\/django-rest-framework,qsorix\/django-rest-framework,cyberj\/django-rest-framework,dmwyatt\/django-rest-framework,zeldalink0515\/django-rest-framework,atombrella\/django-rest-framework,arpheno\/django-rest-framework,kennydude\/django-rest-framework,tcroiset\/django-rest-framework,waytai\/django-rest-framework,AlexandreProenca\/django-rest-framework,wangpanjun\/django-rest-framework,nhorelik\/django-rest-framework,adambain-vokal\/django-rest-framework,antonyc\/django-rest-framework,hnarayanan\/django-rest-framework,wedaly\/django-rest-framework,wedaly\/django-rest-framework"} {"commit":"4393740af93ae0ac1927e68c422e24735b0216c1","old_file":"infosystem\/subsystem\/policy\/entity.py","new_file":"infosystem\/subsystem\/policy\/entity.py","old_contents":"from sqlalchemy import UniqueConstraint\nfrom infosystem.common.subsystem import entity\nfrom infosystem.database import db\n\n\nclass Policy(entity.Entity, db.Model):\n\n attributes = ['id', 'capability_id', 'role_id', 'bypass']\n domain_id = db.Column(db.CHAR(32), db.ForeignKey(\"domain.id\"), nullable=False)\n capability_id = db.Column(db.CHAR(32), db.ForeignKey(\"capability.id\"), nullable=False)\n role_id = db.Column(db.CHAR(32), db.ForeignKey(\"role.id\"), nullable=True)\n bypass = db.Column(db.Boolean, nullable=False, default=False)\n __table_args__ = (UniqueConstraint('domain_id', 'capability_id', 'role_id', name='policy_uk'),)\n\n def __init__(self, id, domain_id, capability_id, role_id, bypass):\n super(Policy, self).__init__(id)\n self.domain_id = domain_id\n self.capability_id = capability_id\n self.role_id = role_id\n self.bypass = bypass\n","new_contents":"from sqlalchemy import UniqueConstraint\nfrom infosystem.common.subsystem import entity\nfrom infosystem.database import db\n\n\nclass Policy(entity.Entity, db.Model):\n\n attributes = ['id', 'capability_id', 'role_id', 'bypass']\n domain_id = db.Column(db.CHAR(32), db.ForeignKey(\"domain.id\"), nullable=False)\n capability_id = db.Column(db.CHAR(32), db.ForeignKey(\"capability.id\"), nullable=False)\n role_id = db.Column(db.CHAR(32), db.ForeignKey(\"role.id\"), nullable=True)\n bypass = db.Column(db.Boolean, nullable=False, default=False)\n __table_args__ = (UniqueConstraint('domain_id', 'capability_id', 'role_id', name='policy_uk'),)\n\n def __init__(self, id, domain_id, capability_id, role_id=None, bypass=False):\n super(Policy, self).__init__(id)\n self.domain_id = domain_id\n self.capability_id = capability_id\n self.role_id = role_id\n self.bypass = bypass\n","subject":"Make role_id & bypass opt args in Policy __init__","message":"Make role_id & bypass opt args in Policy __init__\n","lang":"Python","license":"apache-2.0","repos":"samueldmq\/infosystem"} {"commit":"e99a4aa3fab84e112f5d82eafe9012f7e2be9447","old_file":"problem-static\/Intro-Eval_50\/admin\/eval.py","new_file":"problem-static\/Intro-Eval_50\/admin\/eval.py","old_contents":"#!\/usr\/bin\/python2.7\n\ndel __builtins__.__dict__['__import__']\ndel __builtins__.__dict__['reload']\n\nflag = \"eval_is_fun\"\n\ndef main():\n print \"Hi, welcome to the flag database. We are under construction right now, so you cannot view the flags, or do anything.\"\n while True:\n command = raw_input(\"What would you like to do? \")\n try:\n result = eval(command)\n print \"Here is the result of your command: %s\" %(result)\n except:\n print \"Invalid command, try again\"\n\nmain()\n","new_contents":"#!\/usr\/bin\/python2.7\n\nimport sys\n\ndel __builtins__.__dict__['__import__']\ndel __builtins__.__dict__['reload']\n\nflag = \"eval_is_fun\"\n\nclass UnbufferedStream(object):\n def __init__(self, stream):\n self.stream = stream\n def write(self, data):\n self.stream.write(data)\n self.stream.flush()\n def __getattr__(self, attr):\n return getattr(self.stream, attr)\n\nsys.stdout = UnbufferedStream(sys.stdout)\ndef main():\n while True:\n print \"Welcome to the flag database! We are currently under construction. Please do not hack the flags.\"\n try:\n command = str(raw_input(\"What would you like to do? \"))\n result = str(eval(command))\n print \"This is the result: %s\" %(result)\n except Exception, e:\n print \"Invalid command!!!! EXITING!!!!!\"\n return\n\nmain()","subject":"Fix Intro Eval with Unbuffered Streams","message":"Fix Intro Eval with Unbuffered Streams\n","lang":"Python","license":"mit","repos":"james9909\/IntroCTF,james9909\/IntroCTF,james9909\/IntroCTF,james9909\/IntroCTF,james9909\/IntroCTF,james9909\/IntroCTF"} {"commit":"bdeb60d5e82e5eaaaaf805286bae29e9112af307","old_file":"us_ignite\/common\/management\/commands\/common_load_fixtures.py","new_file":"us_ignite\/common\/management\/commands\/common_load_fixtures.py","old_contents":"import urlparse\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.contrib.sites.models import Site\n\n\nclass Command(BaseCommand):\n\n def handle(self, *args, **options):\n parsed_url = urlparse.urlparse(settings.SITE_URL)\n Site.objects.all().update(domain=parsed_url.netloc,\n name=parsed_url.netloc)\n print \"Done!\"\n","new_contents":"import urlparse\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.contrib.sites.models import Site\n\nfrom us_ignite.profiles.models import Interest\n\n\nINTEREST_LIST = (\n ('SDN', 'sdn'),\n ('OpenFlow', 'openflow'),\n ('Ultra fast', 'ultra-fast'),\n ('Advanced wireless', 'advanced-wireless'),\n ('Low-latency', 'low-latency'),\n ('Local cloud\/edge computing', 'local-cloud-edge-computing'),\n ('Healthcare', 'healthcare'),\n ('Education & Workforce development', 'education-workforce-development'),\n ('Energy', 'energy'),\n ('Transportation', 'transportation'),\n ('Advanced Manufacturing', 'advanced-manufacturing'),\n ('Public Safety', 'public-safety'),\n)\n\n\nclass Command(BaseCommand):\n\n def handle(self, *args, **options):\n parsed_url = urlparse.urlparse(settings.SITE_URL)\n Site.objects.all().update(domain=parsed_url.netloc,\n name=parsed_url.netloc)\n print \"Updated site URL.\"\n for name, slug in INTEREST_LIST:\n interest, is_new = (Interest.objects\n .get_or_create(name=name, slug=slug))\n if is_new:\n print u'Imported interest: %s' % interest\n","subject":"Add initial data for the ``Interest`` model.","message":"Add initial data for the ``Interest`` model.\n","lang":"Python","license":"bsd-3-clause","repos":"us-ignite\/us_ignite,us-ignite\/us_ignite,us-ignite\/us_ignite,us-ignite\/us_ignite,us-ignite\/us_ignite"} {"commit":"4ce3a0d32c3b1399523483257ad1922705cb3dc9","old_file":"__openerp__.py","new_file":"__openerp__.py","old_contents":"# -*- coding: utf-8 -*-\n\n{\n \"name\": \"Account Credit Transfer\",\n \"version\": \"1.0.2\",\n \"author\": \"XCG Consulting\",\n \"website\": \"http:\/\/www.openerp-experts.com\",\n \"category\": 'Accounting',\n \"description\": \"\"\"Account Voucher Credit Transfer Payment.\n You need to set up some things before using it.\n A credit transfer config link a bank with a parser\n A credit transfer parser link a parser with a template that you can upload\n \"\"\",\n\n \"depends\": [\n 'base',\n 'account_streamline',\n ],\n\n \"data\": [\n \"security\/ir.model.access.csv\",\n \"views\/config.xml\",\n \"views\/parser.xml\",\n \"views\/res.bank.xml\",\n ],\n\n 'demo_xml': [],\n 'test': [],\n 'installable': True,\n 'active': False,\n 'external_dependencies': {\n 'python': ['genshi']\n }\n}\n","new_contents":"# -*- coding: utf-8 -*-\n\n{\n \"name\": \"Account Credit Transfer\",\n \"version\": \"1.0.3\",\n \"author\": \"XCG Consulting\",\n \"website\": \"http:\/\/www.openerp-experts.com\",\n \"category\": 'Accounting',\n \"description\": \"\"\"Account Voucher Credit Transfer Payment.\n You need to set up some things before using it.\n A credit transfer config link a bank with a parser\n A credit transfer parser link a parser with a template that you can upload\n \"\"\",\n\n \"depends\": [\n 'base',\n 'account_streamline',\n ],\n\n \"data\": [\n \"security\/ir.model.access.csv\",\n \"views\/config.xml\",\n \"views\/parser.xml\",\n \"views\/res.bank.xml\",\n ],\n\n 'demo_xml': [],\n 'test': [],\n 'installable': True,\n 'active': False,\n 'external_dependencies': {\n 'python': ['genshi']\n }\n}\n","subject":"Change version to 1.0.3 (dev)","message":"Change version to 1.0.3 (dev)\n","lang":"Python","license":"agpl-3.0","repos":"xcgd\/account_credit_transfer"} {"commit":"e73795b8ad016bba5b1ab5118a5153085a1e99b0","old_file":"nova\/tests\/functional\/api_sample_tests\/test_servers_ips.py","new_file":"nova\/tests\/functional\/api_sample_tests\/test_servers_ips.py","old_contents":"# Copyright 2012 Nebula, Inc.\n# Copyright 2013 IBM Corp.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom nova.tests.functional.api_sample_tests import test_servers\n\n\nclass ServersIpsJsonTest(test_servers.ServersSampleBase):\n sample_dir = 'server-ips'\n\n def test_get(self):\n # Test getting a server's IP information.\n uuid = self._post_server()\n response = self._do_get('servers\/%s\/ips' % uuid)\n self._verify_response('server-ips-resp', {}, response, 200)\n\n def test_get_by_network(self):\n # Test getting a server's IP information by network id.\n uuid = self._post_server()\n response = self._do_get('servers\/%s\/ips\/private' % uuid)\n self._verify_response('server-ips-network-resp', {}, response, 200)\n","new_contents":"# Copyright 2012 Nebula, Inc.\n# Copyright 2013 IBM Corp.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom nova.tests import fixtures as nova_fixtures\nfrom nova.tests.functional.api_sample_tests import test_servers\n\n\nclass ServersIpsJsonTest(test_servers.ServersSampleBase):\n sample_dir = 'server-ips'\n\n def test_get(self):\n # Test getting a server's IP information.\n uuid = self._post_server()\n response = self._do_get('servers\/%s\/ips' % uuid)\n self._verify_response('server-ips-resp', {}, response, 200)\n\n def test_get_by_network(self):\n # Test getting a server's IP information by network id.\n server_uuid = self._post_server()\n network_label = nova_fixtures.NeutronFixture.network_1['name']\n response = self._do_get('servers\/%s\/ips\/%s' % (\n server_uuid, network_label))\n self._verify_response('server-ips-network-resp', {}, response, 200)\n","subject":"Make it obvious where we're getting our names from","message":"trivial: Make it obvious where we're getting our names from\n\nChange-Id: Ib9aa790c8999e50a2a3587561604ff1e51666f38\nSigned-off-by: Stephen Finucane <492121341a95b3c3aab646bed44634f739dd019b@redhat.com>\n","lang":"Python","license":"apache-2.0","repos":"mahak\/nova,mahak\/nova,klmitch\/nova,klmitch\/nova,klmitch\/nova,rahulunair\/nova,mahak\/nova,rahulunair\/nova,openstack\/nova,openstack\/nova,rahulunair\/nova,openstack\/nova,klmitch\/nova"} {"commit":"9708e6d01835f597460967b9da47fa4c6d32a907","old_file":"test\/goldstandard\/benchmark_confirm.py","new_file":"test\/goldstandard\/benchmark_confirm.py","old_contents":"import sqlite3 as sql\nimport os\nimport sys\nimport logging\nimport benchmark\n\n# bmVerify(['final_r7', 'final_r8'], filepath=\"\/home\/ysun\/disambig\/newcode\/all\/\", outdir = \"\/home\/ayu\/results_v2\/\")\n \n# Text Files\ntxt_file = 'benchmark_errors.txt'\nopened_file = open(txt_file, 'U')\nlog_file = 'benchmark_results.log'\n\n# Logging\nlogging.basicConfig(filename=log_file, level=logging.DEBUG)\nopen(log_file, \"w\")\n\n# Set Up SQL Connections\ncon = sql.connect('invnum_N_zardoz_with_invpat.sqlite3') \n\nwith con:\n\n con_cur = con.cursor()\n logging.info(\"Beginning to query database\")\n con_cur.execute(\"CREATE INDEX IF NOT EXISTS index_invnum ON invpat (Invnum)\");\n con_cur.execute(\"CREATE INDEX IF NOT EXISTS index_lastname ON invpat (Lastname)\");\n con_cur.execute(\"CREATE INDEX IF NOT EXISTS index_firstname ON invpat (Firstname)\");\n count = 0\n errors = 0\n success = 0\n","new_contents":"import sqlite3 as sql\nimport os\nimport sys\nimport logging\nimport benchmark\n\n# bmVerify(['final_r7', 'final_r8'], filepath=\"\/home\/ysun\/disambig\/newcode\/all\/\", outdir = \"\/home\/ayu\/results_v2\/\")\n \n# Text Files\ntxt_file = 'benchmark_errors.txt'\nopened_file = open(txt_file, 'U')\nlog_file = 'benchmark_results.log'\n\n# Logging\nlogging.basicConfig(filename=log_file, level=logging.DEBUG)\nopen(log_file, \"w\")\n\n# Set Up SQL Connections\ncon = sql.connect('invnum_N_zardoz_with_invpat.sqlite3') \n\nwith con:\n\n con_cur = con.cursor()\n logging.info(\"Beginning to query database\")\n con_cur.execute(\"CREATE INDEX IF NOT EXISTS index_invnum ON invpat (Invnum)\");\n con_cur.execute(\"CREATE INDEX IF NOT EXISTS index_lastname ON invpat (Lastname)\");\n con_cur.execute(\"CREATE INDEX IF NOT EXISTS index_firstname ON invpat (Firstname)\");\n count = 0\n errors = 0\n success = 0\n\n while True:\n \n line_read = opened_file.readline()\n # print line_read\n \n if not line_read:\n print \"EXITING\"\n break\n count = count + 1\n if count%100 == 0:\n print \"starting patent\", count\n","subject":"Set up reading lines from file, print count for testing purposes","message":"Set up reading lines from file, print count for testing purposes\n","lang":"Python","license":"bsd-2-clause","repos":"yngcan\/patentprocessor,nikken1\/patentprocessor,funginstitute\/patentprocessor,nikken1\/patentprocessor,yngcan\/patentprocessor,nikken1\/patentprocessor,funginstitute\/patentprocessor,yngcan\/patentprocessor,funginstitute\/patentprocessor"} {"commit":"943d66dc04ec74a27ae12c46ae1559e585faa716","old_file":"scripts\/master\/factory\/dart\/channels.py","new_file":"scripts\/master\/factory\/dart\/channels.py","old_contents":"# Copyright 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nclass Channel(object):\n def __init__(self, name, branch, position, category_postfix, priority):\n self.branch = branch\n self.builder_postfix = '-' + name\n self.category_postfix = category_postfix\n self.name = name\n self.position = position\n self.priority = priority\n self.all_deps_path = '\/' + branch + '\/deps\/all.deps'\n self.standalone_deps_path = '\/' + branch + '\/deps\/standalone.deps'\n self.dartium_deps_path = '\/' + branch + '\/deps\/dartium.deps'\n\n# The channel names are replicated in the slave.cfg files for all\n# dart waterfalls. If you change anything here please also change it there.\nCHANNELS = [\n Channel('be', 'branches\/bleeding_edge', 0, '', 4),\n Channel('dev', 'trunk', 1, '-dev', 2),\n Channel('stable', 'branches\/1.9', 2, '-stable', 1),\n Channel('integration', 'branches\/dartium_integration', 3, '-integration', 3),\n]\n\nCHANNELS_BY_NAME = {}\nfor c in CHANNELS:\n CHANNELS_BY_NAME[c.name] = c\n","new_contents":"# Copyright 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nclass Channel(object):\n def __init__(self, name, branch, position, category_postfix, priority):\n self.branch = branch\n self.builder_postfix = '-' + name\n self.category_postfix = category_postfix\n self.name = name\n self.position = position\n self.priority = priority\n self.all_deps_path = '\/' + branch + '\/deps\/all.deps'\n self.standalone_deps_path = '\/' + branch + '\/deps\/standalone.deps'\n self.dartium_deps_path = '\/' + branch + '\/deps\/dartium.deps'\n\n# The channel names are replicated in the slave.cfg files for all\n# dart waterfalls. If you change anything here please also change it there.\nCHANNELS = [\n Channel('be', 'branches\/bleeding_edge', 0, '', 4),\n Channel('dev', 'trunk', 1, '-dev', 2),\n Channel('stable', 'branches\/1.10', 2, '-stable', 1),\n Channel('integration', 'branches\/dartium_integration', 3, '-integration', 3),\n]\n\nCHANNELS_BY_NAME = {}\nfor c in CHANNELS:\n CHANNELS_BY_NAME[c.name] = c\n","subject":"Make stable builder pull from 1.10","message":"Make stable builder pull from 1.10\n\nR=kasperl@google.com\nBUG=\n\nReview URL: https:\/\/codereview.chromium.org\/1107673002\n\ngit-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@294974 0039d316-1c4b-4281-b951-d872f2087c98\n","lang":"Python","license":"bsd-3-clause","repos":"eunchong\/build,eunchong\/build,eunchong\/build,eunchong\/build"} {"commit":"4e5ef4a04fd0b3b354b187ee6e8e8ef27337ad6f","old_file":"xclib\/dbmops.py","new_file":"xclib\/dbmops.py","old_contents":"import sys\nimport bsddb3\nfrom xclib.utf8 import utf8, unutf8\n\ndef perform(args):\n domain_db = bsddb3.hashopen(args.domain_db, 'c', 0o600)\n if args.get:\n print(unutf8(domain_db[utf8(args.get)]))\n elif args.put:\n domain_db[utf8(args.put[0])] = args.put[1]\n elif args.delete:\n del domain_db[utf8(args.delete)]\n elif args.unload:\n for k in list(domain_db.keys()):\n print('%s\\t%s' % (unutf8(k), unutf8(domain_db[k])))\n # Should work according to documentation, but doesn't\n # for k, v in DOMAIN_DB.iteritems():\n # print k, '\\t', v\n elif args.load:\n for line in sys.stdin:\n k, v = line.rstrip('\\r\\n').split('\\t', 1)\n domain_db[utf8(k)] = v\n domain_db.close()\n\n# vim: tabstop=8 softtabstop=0 expandtab shiftwidth=4\n","new_contents":"import sys\nimport bsddb3\nfrom xclib.utf8 import utf8, unutf8\n\ndef perform(args):\n domain_db = bsddb3.hashopen(args.domain_db, 'c', 0o600)\n if args.get:\n print(unutf8(domain_db[utf8(args.get)], 'illegal'))\n elif args.put:\n domain_db[utf8(args.put[0])] = args.put[1]\n elif args.delete:\n del domain_db[utf8(args.delete)]\n elif args.unload:\n for k in list(domain_db.keys()):\n print('%s\\t%s' % (unutf8(k, 'illegal'), unutf8(domain_db[k], 'illegal')))\n # Should work according to documentation, but doesn't\n # for k, v in DOMAIN_DB.iteritems():\n # print k, '\\t', v\n elif args.load:\n for line in sys.stdin:\n k, v = line.rstrip('\\r\\n').split('\\t', 1)\n domain_db[utf8(k)] = v\n domain_db.close()\n\n# vim: tabstop=8 softtabstop=0 expandtab shiftwidth=4\n","subject":"Allow dumping illegal utf-8 contents","message":"Allow dumping illegal utf-8 contents\n","lang":"Python","license":"mit","repos":"jsxc\/xmpp-cloud-auth,jsxc\/xmpp-cloud-auth,jsxc\/xmpp-cloud-auth,jsxc\/xmpp-cloud-auth"} {"commit":"53171f75a64a26dcec91facbdec95b2ed7f74338","old_file":"ironic\/drivers\/drac.py","new_file":"ironic\/drivers\/drac.py","old_contents":"#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\"\"\"\nDRAC Driver for remote system management using Dell Remote Access Card.\n\"\"\"\n\nfrom oslo.utils import importutils\n\nfrom ironic.common import exception\nfrom ironic.common.i18n import _\nfrom ironic.drivers import base\nfrom ironic.drivers.modules.drac import management\nfrom ironic.drivers.modules.drac import power\nfrom ironic.drivers.modules import pxe\n\n\nclass PXEDracDriver(base.BaseDriver):\n \"\"\"Drac driver using PXE for deploy.\"\"\"\n\n def __init__(self):\n if not importutils.try_import('pywsman'):\n raise exception.DriverLoadError(\n driver=self.__class__.__name__,\n reason=_('Unable to import pywsman library'))\n\n self.power = power.DracPower()\n self.deploy = pxe.PXEDeploy()\n self.management = management.DracManagement()\n","new_contents":"#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\"\"\"\nDRAC Driver for remote system management using Dell Remote Access Card.\n\"\"\"\n\nfrom oslo.utils import importutils\n\nfrom ironic.common import exception\nfrom ironic.common.i18n import _\nfrom ironic.drivers import base\nfrom ironic.drivers.modules.drac import management\nfrom ironic.drivers.modules.drac import power\nfrom ironic.drivers.modules import pxe\n\n\nclass PXEDracDriver(base.BaseDriver):\n \"\"\"Drac driver using PXE for deploy.\"\"\"\n\n def __init__(self):\n if not importutils.try_import('pywsman'):\n raise exception.DriverLoadError(\n driver=self.__class__.__name__,\n reason=_('Unable to import pywsman library'))\n\n self.power = power.DracPower()\n self.deploy = pxe.PXEDeploy()\n self.management = management.DracManagement()\n self.vendor = pxe.VendorPassthru()\n","subject":"Add the PXE VendorPassthru interface to PXEDracDriver","message":"Add the PXE VendorPassthru interface to PXEDracDriver\n\nWithout the PXE VendorPassthru interface to expose the \"pass_deploy_info\"\nmethod in the vendor_passthru endpoint of the API the DRAC it can't\ncontinue the deployment after the ramdisk is booted.\n\nCloses-Bug: #1379705\nChange-Id: I21042cbb95a486742abfcb430471d01cd73b3a4a\n(cherry picked from commit 78ec7d5336eb65ff845da7ea9f93d34b402f5a0f)\n","lang":"Python","license":"apache-2.0","repos":"SauloAislan\/ironic,dims\/ironic,NaohiroTamura\/ironic,hpproliant\/ironic,bacaldwell\/ironic,naterh\/ironic,redhat-openstack\/ironic,NaohiroTamura\/ironic,ionutbalutoiu\/ironic,SauloAislan\/ironic,dims\/ironic,openstack\/ironic,Tan0\/ironic,openstack\/ironic,pshchelo\/ironic,debayanray\/ironic_backup,ionutbalutoiu\/ironic,devananda\/ironic,pshchelo\/ironic,bacaldwell\/ironic"} {"commit":"84a2ece10b0e246564fd539eed119f46d44ca74d","old_file":"tests\/no_hadoop_bare_image_provider.py","new_file":"tests\/no_hadoop_bare_image_provider.py","old_contents":"# -*- coding: utf-8 -*-\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"\nProvides bare images for standalone clusters.\n\"\"\"\n\nfrom tests.bare_image_provider import TagBareImageProvider\nfrom tests.product.constants import BASE_IMAGES_TAG\nfrom tests.product.constants import BASE_IMAGE_NAME\n\n\nclass NoHadoopBareImageProvider(TagBareImageProvider):\n def __init__(self):\n super(NoHadoopBareImageProvider, self).__init__(\n BASE_IMAGE_NAME, BASE_IMAGE_NAME,\n BASE_IMAGES_TAG, 'nohadoop')\n","new_contents":"# -*- coding: utf-8 -*-\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"\nProvides bare images for standalone clusters.\n\"\"\"\n\nimport re\n\nfrom tests.bare_image_provider import TagBareImageProvider\nfrom tests.product.constants import BASE_IMAGES_TAG\nfrom tests.product.constants import BASE_IMAGE_NAME\n\n\nclass NoHadoopBareImageProvider(TagBareImageProvider):\n def __init__(self):\n # encode base image name into name of created test image, to prevent image name clash.\n decoration = 'nohadoop_' + re.sub(r\"[^A-Za-z0-9]\", \"_\", BASE_IMAGE_NAME)\n super(NoHadoopBareImageProvider, self).__init__(\n BASE_IMAGE_NAME, BASE_IMAGE_NAME,\n BASE_IMAGES_TAG, decoration)\n","subject":"Use bare image name dependent on base image name","message":"Use bare image name dependent on base image name\n","lang":"Python","license":"apache-2.0","repos":"prestodb\/presto-admin,prestodb\/presto-admin"} {"commit":"d927c0393709ac316a6054fb99444f4b14ebafb6","old_file":"aslack\/__init__.py","new_file":"aslack\/__init__.py","old_contents":"\"\"\"Asynchronous Slack API integration.\"\"\"\n\nimport logging\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n\n__author__ = 'Jonathan Sharpe'\n__version__ = '0.8.3'\n","new_contents":"\"\"\"Asynchronous Slack API integration.\"\"\"\n\nimport logging\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n\n__author__ = 'Jonathan Sharpe'\n__version__ = '0.8.4'\n","subject":"Bump version for Sphinx update","message":"Bump version for Sphinx update","lang":"Python","license":"isc","repos":"textbook\/aslack"} {"commit":"c671301e29e41b0ea7988bad0a5ff4793890ea64","old_file":"readKanjiFileTest.py","new_file":"readKanjiFileTest.py","old_contents":"from readKanjiFile import *\n\ndef main():\n testDict = readKanjiFile('kanjiTestFile.txt')\n\n print('There are {} entries in this dictionary.'.format(len(testDict)))\n for entry in list(testDict.items())[0:15]:\n print(entry)\n\nif __name__ == '__main__':\n main()\n","new_contents":"from readKanjiFile import *\n\ndef main():\n testDict = readKanjiFile('kanjiTestFile.txt')\n\n print('There are {} entries in this dictionary.'.format(len(testDict)))\n for entry in list(testDict.items()):\n print(entry)\n\nif __name__ == '__main__':\n main()\n","subject":"Test now prints out all kanji","message":"Test now prints out all kanji\n","lang":"Python","license":"mit","repos":"WilliamRayJohnson\/kanjiStudier"} {"commit":"d1d7684edb6d687206deea75d2ba13194046e376","old_file":"sixquiprend\/models\/chosen_card.py","new_file":"sixquiprend\/models\/chosen_card.py","old_contents":"from sixquiprend.sixquiprend import app, db\n\nclass ChosenCard(db.Model):\n id = db.Column(db.Integer, primary_key=True)\n user_id = db.Column(db.Integer, db.ForeignKey('user.id', ondelete=\"CASCADE\"))\n game_id = db.Column(db.Integer, db.ForeignKey('game.id', ondelete=\"CASCADE\"))\n card_id = db.Column(db.Integer, db.ForeignKey('card.id'))\n\n ################################################################################\n ## Serializer\n ################################################################################\n\n def serialize(self):\n from sixquiprend.models.card import Card\n return {\n 'id': self.id,\n 'user_id': self.user_id,\n 'game_id': self.game_id,\n 'card': Card.find(self.card_id)\n }\n","new_contents":"from sixquiprend.sixquiprend import app, db\nfrom sixquiprend.models.card import Card\n\nclass ChosenCard(db.Model):\n id = db.Column(db.Integer, primary_key=True)\n user_id = db.Column(db.Integer, db.ForeignKey('user.id', ondelete=\"CASCADE\"))\n game_id = db.Column(db.Integer, db.ForeignKey('game.id', ondelete=\"CASCADE\"))\n card_id = db.Column(db.Integer, db.ForeignKey('card.id'))\n\n ################################################################################\n ## Serializer\n ################################################################################\n\n def serialize(self):\n return {\n 'id': self.id,\n 'user_id': self.user_id,\n 'game_id': self.game_id,\n 'card': Card.find(self.card_id)\n }\n","subject":"Move an import to top","message":"Move an import to top\n","lang":"Python","license":"mit","repos":"nyddogghr\/SixQuiPrend,nyddogghr\/SixQuiPrend,nyddogghr\/SixQuiPrend,nyddogghr\/SixQuiPrend"} {"commit":"778f5b8f905195a54a09f1e9cc8768e5cabc4dcd","old_file":"navigator\/settings\/prod.py","new_file":"navigator\/settings\/prod.py","old_contents":"from .base import *\n\nDEBUG = False\nALLOWED_HOSTS = ['selling-online-overseas.export.great.gov.uk']\nADMINS = (('David Downes', 'david@downes.co.uk'),)\n\nMIDDLEWARE_CLASSES += [\n 'core.middleware.IpRestrictionMiddleware',\n]\n\nINSTALLED_APPS += [\n 'raven.contrib.django.raven_compat'\n]\n\nRAVEN_CONFIG = {\n 'dsn': os.environ.get('SENTRY_DSN'),\n}\n\nip_check = os.environ.get('RESTRICT_IPS', False)\nRESTRICT_IPS = ip_check == 'True' or ip_check == '1'\n\nALLOWED_IPS = []\nALLOWED_IP_RANGES = ['165.225.80.0\/22', '193.240.203.32\/29']\n\nSECURE_SSL_REDIRECT = True\n# XXX: This needs to be made longer once it is confirmed it works as desired\nSECURE_HSTS_SECONDS = 31536000\n","new_contents":"from .base import *\n\nDEBUG = False\nALLOWED_HOSTS = ['selling-online-overseas.export.great.gov.uk']\nADMINS = (('David Downes', 'david@downes.co.uk'),)\n\nMIDDLEWARE_CLASSES += [\n 'core.middleware.IpRestrictionMiddleware',\n]\n\nINSTALLED_APPS += [\n 'raven.contrib.django.raven_compat'\n]\n\nRAVEN_CONFIG = {\n 'dsn': os.environ.get('SENTRY_DSN'),\n}\n\nip_check = os.environ.get('RESTRICT_IPS', False)\nRESTRICT_IPS = ip_check == 'True' or ip_check == '1'\n\nALLOWED_IPS = []\nALLOWED_IP_RANGES = ['165.225.80.0\/22', '193.240.203.32\/29']\n\nSECURE_SSL_REDIRECT = True\nSECURE_HSTS_SECONDS = 31536000\nSESSION_COOKIE_SECURE = True\nCSRF_COOKIE_SECURE = True\n","subject":"Set the secure flag for both our cookies","message":"Set the secure flag for both our cookies\n","lang":"Python","license":"mit","repos":"uktrade\/navigator,dahfool\/navigator,dahfool\/navigator,uktrade\/navigator,uktrade\/navigator,dahfool\/navigator,dahfool\/navigator,uktrade\/navigator"} {"commit":"166c1a4dde981d5bd7d20a00c8329d7bbb4a3c00","old_file":"nipype\/interfaces\/setup.py","new_file":"nipype\/interfaces\/setup.py","old_contents":"def configuration(parent_package='',top_path=None):\n from numpy.distutils.misc_util import Configuration\n\n config = Configuration('interfaces', parent_package, top_path)\n\n config.add_data_dir('tests')\n config.add_data_dir('data')\n config.add_data_dir('script_templates')\n\n return config\n\nif __name__ == '__main__':\n from numpy.distutils.core import setup\n setup(**configuration(top_path='').todict())\n","new_contents":"def configuration(parent_package='',top_path=None):\n from numpy.distutils.misc_util import Configuration\n\n config = Configuration('interfaces', parent_package, top_path)\n\n config.add_data_dir('tests')\n config.add_data_dir('script_templates')\n\n return config\n\nif __name__ == '__main__':\n from numpy.distutils.core import setup\n setup(**configuration(top_path='').todict())\n","subject":"Remove reference to non-existing data directory.","message":"Remove reference to non-existing data directory.\n\ngit-svn-id: 24f545668198cdd163a527378499f2123e59bf9f@513 ead46cd0-7350-4e37-8683-fc4c6f79bf00\n","lang":"Python","license":"bsd-3-clause","repos":"arokem\/nipype,gerddie\/nipype,iglpdc\/nipype,pearsonlab\/nipype,carlohamalainen\/nipype,Leoniela\/nipype,glatard\/nipype,dgellis90\/nipype,satra\/NiPypeold,glatard\/nipype,mick-d\/nipype,carolFrohlich\/nipype,blakedewey\/nipype,dgellis90\/nipype,pearsonlab\/nipype,mick-d\/nipype_source,FCP-INDI\/nipype,dgellis90\/nipype,carolFrohlich\/nipype,FredLoney\/nipype,arokem\/nipype,rameshvs\/nipype,arokem\/nipype,gerddie\/nipype,carolFrohlich\/nipype,gerddie\/nipype,rameshvs\/nipype,mick-d\/nipype,sgiavasis\/nipype,carlohamalainen\/nipype,dmordom\/nipype,wanderine\/nipype,fprados\/nipype,grlee77\/nipype,carlohamalainen\/nipype,blakedewey\/nipype,mick-d\/nipype_source,Leoniela\/nipype,FredLoney\/nipype,dgellis90\/nipype,sgiavasis\/nipype,rameshvs\/nipype,blakedewey\/nipype,dmordom\/nipype,satra\/NiPypeold,dmordom\/nipype,sgiavasis\/nipype,christianbrodbeck\/nipype,FredLoney\/nipype,grlee77\/nipype,fprados\/nipype,iglpdc\/nipype,iglpdc\/nipype,JohnGriffiths\/nipype,sgiavasis\/nipype,gerddie\/nipype,grlee77\/nipype,christianbrodbeck\/nipype,wanderine\/nipype,fprados\/nipype,FCP-INDI\/nipype,arokem\/nipype,JohnGriffiths\/nipype,wanderine\/nipype,carolFrohlich\/nipype,FCP-INDI\/nipype,pearsonlab\/nipype,glatard\/nipype,JohnGriffiths\/nipype,blakedewey\/nipype,grlee77\/nipype,rameshvs\/nipype,glatard\/nipype,Leoniela\/nipype,pearsonlab\/nipype,mick-d\/nipype_source,FCP-INDI\/nipype,mick-d\/nipype,iglpdc\/nipype,wanderine\/nipype,JohnGriffiths\/nipype,mick-d\/nipype"} {"commit":"56ac100c8ca357a5600db7a16859cca1483ccb13","old_file":"blueprints\/multi_node_kubernetes_cluster\/teardown_kubernetes_cluster\/teardown_kubernetes_cluster.py","new_file":"blueprints\/multi_node_kubernetes_cluster\/teardown_kubernetes_cluster\/teardown_kubernetes_cluster.py","old_contents":"\"\"\"\nTeardown the CloudBolt resources (container_orchestrator, environment)\nassociated with this Kubernetes cluster.\n\"\"\"\nfrom common.methods import set_progress\nfrom containerorchestrators.kuberneteshandler.models import Kubernetes\n\n\ndef run(job, *args, **kwargs):\n resource = job.resource_set.first()\n \n container_orchestrator = Kubernetes.objects.get(id=resource.container_orchestrator_id)\n environment = container_orchestrator.environment_set.first()\n \n container_orchestrator.delete()\n environment.delete()","new_contents":"\"\"\"\nTeardown the CloudBolt resources (container_orchestrator, environment)\nassociated with this Kubernetes cluster.\n\"\"\"\nfrom common.methods import set_progress\nfrom containerorchestrators.kuberneteshandler.models import Kubernetes\nfrom utilities.run_command import execute_command\n\n\ndef run(job, *args, **kwargs):\n resource = job.resource_set.first()\n \n container_orchestrator = Kubernetes.objects.get(id=resource.container_orchestrator_id)\n environment = container_orchestrator.environment_set.first()\n \n container_orchestrator.delete()\n environment.delete()\n\n resource_dir = '\/var\/opt\/cloudbolt\/kubernetes\/resource-{}'.format(resource_id)\n execute_command('rm -rf {}'.format(RESOURCE_LOCATION))\n","subject":"Remove config files from filesystem on teardown","message":"Remove config files from filesystem on teardown\n\n[DEV-13843]\n","lang":"Python","license":"apache-2.0","repos":"CloudBoltSoftware\/cloudbolt-forge,CloudBoltSoftware\/cloudbolt-forge,CloudBoltSoftware\/cloudbolt-forge,CloudBoltSoftware\/cloudbolt-forge"} {"commit":"85405d242b41b9dd721131c6e3046c47fe86847d","old_file":"pic2map\/cli.py","new_file":"pic2map\/cli.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"Command Line Interface.\"\"\"\n\nimport argparse\nimport sys\n\ndef main(argv):\n \"\"\"Entry point for the pic2map.py script.\"\"\"\n args = parse_arguments(argv)\n print args\n\ndef parse_arguments(argv):\n \"\"\"Parse command line arguments.\n\n :returns: Parsed arguments\n :rtype: argparse.Namespace\n\n \"\"\"\n parser = argparse.ArgumentParser(\n description='Display pictures location in a map')\n args = parser.parse_args(argv)\n return args\n\nif __name__ == '__main__':\n main(sys.argv[1:])\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"Command Line Interface.\"\"\"\n\nimport argparse\nimport logging\nimport sys\n\n\ndef main(argv):\n \"\"\"Entry point for the pic2map.py script.\"\"\"\n args = parse_arguments(argv)\n print args\n\n\ndef configure_logging(log_level):\n \"\"\"Configure logging based on command line argument.\n\n :param log_level: Log level passed form the command line\n :type log_level: int\n\n \"\"\"\n root_logger = logging.getLogger()\n root_logger.setLevel(logging.DEBUG)\n\n # Log to sys.stderr using log level\n # passed through command line\n log_handler = logging.StreamHandler()\n formatter = logging.Formatter('%(levelname)s: %(message)s')\n log_handler.setFormatter(formatter)\n log_handler.setLevel(log_level)\n root_logger.addHandler(log_handler)\n\n\ndef parse_arguments(argv):\n \"\"\"Parse command line arguments.\n\n :returns: Parsed arguments\n :rtype: argparse.Namespace\n\n \"\"\"\n parser = argparse.ArgumentParser(\n description='Display pictures location in a map')\n log_levels = ['debug', 'info', 'warning', 'error', 'critical']\n parser.add_argument(\n '-l', '--log-level',\n dest='log_level',\n choices=log_levels,\n default='warning',\n help=('Log level. One of {0} or {1} '\n '(%(default)s by default)'\n .format(', '.join(log_levels[:-1]), log_levels[-1])))\n\n args = parser.parse_args(argv)\n args.log_level = getattr(logging, args.log_level.upper())\n return args\n\nif __name__ == '__main__':\n main(sys.argv[1:])\n","subject":"Add command line option to set log level","message":"Add command line option to set log level\n","lang":"Python","license":"mit","repos":"jcollado\/pic2map,jcollado\/pic2map,jcollado\/pic2map"} {"commit":"f1d2bb08dfde9402b7fc858d57130b43e3f1cd4d","old_file":"bootstrap\/hooks.py","new_file":"bootstrap\/hooks.py","old_contents":"# coding: utf-8\nfrom os.path import join, dirname, pardir, abspath\nfrom shutil import copy\nimport subprocess\n\n\nBOOTSTRAP = abspath(dirname(__file__))\nROOT = abspath(join(BOOTSTRAP, pardir))\n\n# Path where venv will be created. It's imported by bootstrapX.Y.py\nVIRTUALENV = abspath(join(BOOTSTRAP, pardir))\n\nACTIVATE = join(VIRTUALENV, 'bin', 'activate_this.py')\nWITH_VENV = join(BOOTSTRAP, 'with_venv.sh')\n\n\ndef with_venv(*args):\n \"\"\"\n Runs the given command inside virtualenv.\n \"\"\"\n cmd = list(args)\n cmd.insert(0, WITH_VENV)\n return subprocess.call(cmd)\n\n\ndef after_install(options, home_dir):\n copy(join(BOOTSTRAP, 'postactivate'), VIRTUALENV)\n with_venv('pip', 'install', '-r', join(ROOT, 'requirements.txt'))\n print \"Done! Activate your virtualenv: source bin\/activate\"\n\n","new_contents":"# coding: utf-8\nfrom os.path import join, dirname, pardir, abspath\nfrom shutil import copy\nimport subprocess\n\n\nBOOTSTRAP = abspath(dirname(__file__))\nROOT = abspath(join(BOOTSTRAP, pardir))\n\n# Path where venv will be created. It's imported by bootstrapX.Y.py\nVIRTUALENV = join(BOOTSTRAP, pardir)\nVIRTUALENV_BIN = join(VIRTUALENV, 'bin')\n\nACTIVATE = join(VIRTUALENV, 'bin', 'activate_this.py')\nWITH_VENV = join(BOOTSTRAP, 'with_venv.sh')\n\n\ndef with_venv(*args):\n \"\"\"\n Runs the given command inside virtualenv.\n \"\"\"\n cmd = list(args)\n cmd.insert(0, WITH_VENV)\n return subprocess.call(cmd)\n\n\ndef after_install(options, home_dir):\n copy(join(BOOTSTRAP, 'postactivate'), VIRTUALENV_BIN)\n with_venv('pip', 'install', '-r', join(ROOT, 'requirements.txt'))\n print \"Done! Activate your virtualenv: source bin\/activate\"\n\n","subject":"Fix wrong destination for postactivate file.","message":"Fix wrong destination for postactivate file.","lang":"Python","license":"mit","repos":"henriquebastos\/virtualenv-bootstrap,henriquebastos\/virtualenv-bootstrap"} {"commit":"8e72ef3fa525c961786e9b60c039c847bc2c710f","old_file":"caSandbox.py","new_file":"caSandbox.py","old_contents":"import map\nimport curses\n\n# Set up Curses screen\nscreen = curses.initscr()\n\ncurses.noecho()\nscreen.keypad(True)\ncurses.cbreak()\ncurses.halfdelay(5) # Wait for half a second for input before continuing\ncurses.start_color()\n\ncurses.init_pair(1, curses.COLOR_RED, curses.COLOR_BLACK)\ncurses.init_pair(2, curses.COLOR_WHITE, curses.COLOR_BLACK)\n\n# Initialize the map\nm = map.Map(screen, \"data\/test_data.txt\")\n\ni = 0\n\nwhile True:\n # Check for exit key\n char = screen.getch()\n if char == ord('q'):\n break\n\n # Advance the simulation\n m.print_cells(x=10, y=10)\n m.update_cells()\n\n# Clean up\ncurses.nocbreak()\nscreen.keypad(False)\ncurses.echo()\n\ncurses.endwin()\n","new_contents":"import map\nimport curses\n\n# Set up Curses screen\nscreen = curses.initscr()\n\ncurses.noecho()\nscreen.keypad(True)\ncurses.cbreak()\ncurses.halfdelay(5) # Wait for half a second for input before continuing\ncurses.start_color()\n\ncurses.init_pair(1, curses.COLOR_RED, curses.COLOR_BLACK)\ncurses.init_pair(2, curses.COLOR_WHITE, curses.COLOR_BLACK)\n\n# Initialize the map\nm = map.Map(screen, \"data\/test_data.txt\")\n\nwhile True:\n # Check for exit key\n char = screen.getch()\n if char != -1:\n break\n\n # Advance the simulation\n m.print_cells(x=10, y=10)\n m.update_cells()\n\n# Clean up\ncurses.nocbreak()\nscreen.keypad(False)\ncurses.echo()\n\ncurses.endwin()\n","subject":"Make program close on any keypress","message":"Make program close on any keypress\n","lang":"Python","license":"mit","repos":"cferwin\/CA-Sandbox"} {"commit":"bf38a26ea239ce70fd4fc3748912b243fb1f7d88","old_file":"tools\/perf\/benchmarks\/pica.py","new_file":"tools\/perf\/benchmarks\/pica.py","old_contents":"# Copyright 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nfrom telemetry import test\nfrom telemetry.page import page_measurement\n\nclass PicaMeasurement(page_measurement.PageMeasurement):\n def MeasurePage(self, _, tab, results):\n result = int(tab.EvaluateJavaScript('__pica_load_time'))\n results.Add('Total', 'ms', result)\n\n\nclass Pica(test.Test):\n test = PicaMeasurement\n page_set = 'page_sets\/pica.json'\n","new_contents":"# Copyright 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nfrom telemetry import test\nfrom telemetry.page import page_measurement\n\nclass PicaMeasurement(page_measurement.PageMeasurement):\n def CustomizeBrowserOptions(self, options):\n # Needed for native custom elements (document.register)\n options.AppendExtraBrowserArg('--enable-experimental-web-platform-features')\n\n def MeasurePage(self, _, tab, results):\n result = int(tab.EvaluateJavaScript('__pica_load_time'))\n results.Add('Total', 'ms', result)\n\n\nclass Pica(test.Test):\n test = PicaMeasurement\n page_set = 'page_sets\/pica.json'\n","subject":"Enable native custom elements for Pica benchmark","message":"Enable native custom elements for Pica benchmark\n\nR=tonyg@chromium.org\nBUG=245358\n\nReview URL: https:\/\/codereview.chromium.org\/22884003\n\ngit-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@217042 0039d316-1c4b-4281-b951-d872f2087c98\n","lang":"Python","license":"bsd-3-clause","repos":"krieger-od\/nwjs_chromium.src,ChromiumWebApps\/chromium,mohamed--abdel-maksoud\/chromium.src,krieger-od\/nwjs_chromium.src,patrickm\/chromium.src,hgl888\/chromium-crosswalk-efl,Just-D\/chromium-1,Chilledheart\/chromium,mogoweb\/chromium-crosswalk,Just-D\/chromium-1,bright-sparks\/chromium-spacewalk,mohamed--abdel-maksoud\/chromium.src,krieger-od\/nwjs_chromium.src,ChromiumWebApps\/chromium,hgl888\/chromium-crosswalk,M4sse\/chromium.src,markYoungH\/chromium.src,littlstar\/chromium.src,chuan9\/chromium-crosswalk,chuan9\/chromium-crosswalk,Just-D\/chromium-1,anirudhSK\/chromium,hgl888\/chromium-crosswalk,ltilve\/chromium,Fireblend\/chromium-crosswalk,ChromiumWebApps\/chromium,bright-sparks\/chromium-spacewalk,axinging\/chromium-crosswalk,mohamed--abdel-maksoud\/chromium.src,Just-D\/chromium-1,axinging\/chromium-crosswalk,Fireblend\/chromium-crosswalk,littlstar\/chromium.src,Chilledheart\/chromium,TheTypoMaster\/chromium-crosswalk,Fireblend\/chromium-crosswalk,fujunwei\/chromium-crosswalk,anirudhSK\/chromium,Pluto-tv\/chromium-crosswalk,Chilledheart\/chromium,littlstar\/chromium.src,ltilve\/chromium,Jonekee\/chromium.src,krieger-od\/nwjs_chromium.src,jaruba\/chromium.src,dednal\/chromium.src,bright-sparks\/chromium-spacewalk,chuan9\/chromium-crosswalk,dushu1203\/chromium.src,ltilve\/chromium,markYoungH\/chromium.src,anirudhSK\/chromium,PeterWangIntel\/chromium-crosswalk,mogoweb\/chromium-crosswalk,dednal\/chromium.src,crosswalk-project\/chromium-crosswalk-efl,dednal\/chromium.src,dushu1203\/chromium.src,hgl888\/chromium-crosswalk-efl,PeterWangIntel\/chromium-crosswalk,Fireblend\/chromium-crosswalk,ondra-novak\/chromium.src,Just-D\/chromium-1,TheTypoMaster\/chromium-crosswalk,markYoungH\/chromium.src,Chilledheart\/chromium,anirudhSK\/chromium,Jonekee\/chromium.src,axinging\/chromium-crosswalk,Pluto-tv\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,jaruba\/chromium.src,chuan9\/chromium-crosswalk,dushu1203\/chromium.src,Pluto-tv\/chromium-crosswalk,hgl888\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,dednal\/chromium.src,ltilve\/chromium,TheTypoMaster\/chromium-crosswalk,littlstar\/chromium.src,mogoweb\/chromium-crosswalk,M4sse\/chromium.src,PeterWangIntel\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,mohamed--abdel-maksoud\/chromium.src,ondra-novak\/chromium.src,mohamed--abdel-maksoud\/chromium.src,jaruba\/chromium.src,fujunwei\/chromium-crosswalk,hgl888\/chromium-crosswalk,chuan9\/chromium-crosswalk,dushu1203\/chromium.src,axinging\/chromium-crosswalk,TheTypoMaster\/chromium-crosswalk,anirudhSK\/chromium,TheTypoMaster\/chromium-crosswalk,krieger-od\/nwjs_chromium.src,hgl888\/chromium-crosswalk-efl,Just-D\/chromium-1,axinging\/chromium-crosswalk,ChromiumWebApps\/chromium,Chilledheart\/chromium,jaruba\/chromium.src,fujunwei\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,axinging\/chromium-crosswalk,Pluto-tv\/chromium-crosswalk,dednal\/chromium.src,Chilledheart\/chromium,TheTypoMaster\/chromium-crosswalk,anirudhSK\/chromium,Jonekee\/chromium.src,hgl888\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,krieger-od\/nwjs_chromium.src,hgl888\/chromium-crosswalk,littlstar\/chromium.src,mogoweb\/chromium-crosswalk,Pluto-tv\/chromium-crosswalk,hgl888\/chromium-crosswalk,ondra-novak\/chromium.src,Pluto-tv\/chromium-crosswalk,TheTypoMaster\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,ltilve\/chromium,chuan9\/chromium-crosswalk,jaruba\/chromium.src,markYoungH\/chromium.src,mohamed--abdel-maksoud\/chromium.src,hgl888\/chromium-crosswalk-efl,ChromiumWebApps\/chromium,mohamed--abdel-maksoud\/chromium.src,chuan9\/chromium-crosswalk,anirudhSK\/chromium,Pluto-tv\/chromium-crosswalk,jaruba\/chromium.src,PeterWangIntel\/chromium-crosswalk,PeterWangIntel\/chromium-crosswalk,anirudhSK\/chromium,crosswalk-project\/chromium-crosswalk-efl,fujunwei\/chromium-crosswalk,Fireblend\/chromium-crosswalk,TheTypoMaster\/chromium-crosswalk,ltilve\/chromium,Jonekee\/chromium.src,ondra-novak\/chromium.src,markYoungH\/chromium.src,ondra-novak\/chromium.src,dushu1203\/chromium.src,krieger-od\/nwjs_chromium.src,M4sse\/chromium.src,ondra-novak\/chromium.src,ondra-novak\/chromium.src,krieger-od\/nwjs_chromium.src,Just-D\/chromium-1,mohamed--abdel-maksoud\/chromium.src,Fireblend\/chromium-crosswalk,chuan9\/chromium-crosswalk,patrickm\/chromium.src,ondra-novak\/chromium.src,M4sse\/chromium.src,crosswalk-project\/chromium-crosswalk-efl,patrickm\/chromium.src,ChromiumWebApps\/chromium,axinging\/chromium-crosswalk,axinging\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,M4sse\/chromium.src,dushu1203\/chromium.src,Chilledheart\/chromium,patrickm\/chromium.src,Jonekee\/chromium.src,markYoungH\/chromium.src,markYoungH\/chromium.src,ChromiumWebApps\/chromium,TheTypoMaster\/chromium-crosswalk,hgl888\/chromium-crosswalk,patrickm\/chromium.src,dednal\/chromium.src,M4sse\/chromium.src,patrickm\/chromium.src,hgl888\/chromium-crosswalk,dushu1203\/chromium.src,mogoweb\/chromium-crosswalk,Fireblend\/chromium-crosswalk,jaruba\/chromium.src,dednal\/chromium.src,mogoweb\/chromium-crosswalk,Jonekee\/chromium.src,dednal\/chromium.src,mohamed--abdel-maksoud\/chromium.src,patrickm\/chromium.src,littlstar\/chromium.src,jaruba\/chromium.src,markYoungH\/chromium.src,bright-sparks\/chromium-spacewalk,mohamed--abdel-maksoud\/chromium.src,dushu1203\/chromium.src,jaruba\/chromium.src,crosswalk-project\/chromium-crosswalk-efl,Pluto-tv\/chromium-crosswalk,dednal\/chromium.src,ltilve\/chromium,dednal\/chromium.src,littlstar\/chromium.src,markYoungH\/chromium.src,axinging\/chromium-crosswalk,Jonekee\/chromium.src,markYoungH\/chromium.src,littlstar\/chromium.src,bright-sparks\/chromium-spacewalk,fujunwei\/chromium-crosswalk,Chilledheart\/chromium,Chilledheart\/chromium,hgl888\/chromium-crosswalk-efl,crosswalk-project\/chromium-crosswalk-efl,jaruba\/chromium.src,mogoweb\/chromium-crosswalk,Fireblend\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,chuan9\/chromium-crosswalk,patrickm\/chromium.src,Jonekee\/chromium.src,crosswalk-project\/chromium-crosswalk-efl,Pluto-tv\/chromium-crosswalk,fujunwei\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,Just-D\/chromium-1,dushu1203\/chromium.src,ChromiumWebApps\/chromium,anirudhSK\/chromium,Fireblend\/chromium-crosswalk,mohamed--abdel-maksoud\/chromium.src,M4sse\/chromium.src,ltilve\/chromium,fujunwei\/chromium-crosswalk,M4sse\/chromium.src,ChromiumWebApps\/chromium,PeterWangIntel\/chromium-crosswalk,axinging\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,axinging\/chromium-crosswalk,dushu1203\/chromium.src,Jonekee\/chromium.src,Just-D\/chromium-1,ltilve\/chromium,M4sse\/chromium.src,ChromiumWebApps\/chromium,patrickm\/chromium.src,ChromiumWebApps\/chromium,PeterWangIntel\/chromium-crosswalk,mogoweb\/chromium-crosswalk,krieger-od\/nwjs_chromium.src,Jonekee\/chromium.src,M4sse\/chromium.src,dednal\/chromium.src,anirudhSK\/chromium,jaruba\/chromium.src,anirudhSK\/chromium,ondra-novak\/chromium.src,PeterWangIntel\/chromium-crosswalk,fujunwei\/chromium-crosswalk,fujunwei\/chromium-crosswalk,markYoungH\/chromium.src,dushu1203\/chromium.src,mogoweb\/chromium-crosswalk,Jonekee\/chromium.src,krieger-od\/nwjs_chromium.src,M4sse\/chromium.src,krieger-od\/nwjs_chromium.src,crosswalk-project\/chromium-crosswalk-efl,ChromiumWebApps\/chromium,anirudhSK\/chromium,bright-sparks\/chromium-spacewalk,PeterWangIntel\/chromium-crosswalk,mogoweb\/chromium-crosswalk"} {"commit":"2f72f75da7ba03e331927c5ab0a5702c150b2f9a","old_file":"perfrunner\/celeryremote.py","new_file":"perfrunner\/celeryremote.py","old_contents":"BROKER_URL = 'amqp:\/\/couchbase:couchbase@ci.sc.couchbase.com:5672\/broker'\nCELERY_RESULT_BACKEND = 'amqp'\nCELERY_RESULT_EXCHANGE = 'perf_results'\nCELERY_RESULT_PERSISTENT = False\n","new_contents":"BROKER_URL = 'amqp:\/\/couchbase:couchbase@172.23.97.73:5672\/broker'\nCELERY_RESULT_BACKEND = 'amqp'\nCELERY_RESULT_EXCHANGE = 'perf_results'\nCELERY_RESULT_PERSISTENT = False\n","subject":"Use broker IP instead of domain name","message":"Use broker IP instead of domain name\n\nChange-Id: Ide27c97a00c18ac62c1a92e2ec51c74c5af4cf30\nReviewed-on: http:\/\/review.couchbase.org\/81029\nTested-by: Build Bot <80754af91bfb6d1073585b046fe0a474ce868509@couchbase.com>\nReviewed-by: Pavel Paulau \n","lang":"Python","license":"apache-2.0","repos":"couchbase\/perfrunner,couchbase\/perfrunner,pavel-paulau\/perfrunner,pavel-paulau\/perfrunner,pavel-paulau\/perfrunner,couchbase\/perfrunner,pavel-paulau\/perfrunner,couchbase\/perfrunner,couchbase\/perfrunner,couchbase\/perfrunner,pavel-paulau\/perfrunner"} {"commit":"a5bc36df3435258fad9700c150985998e9663ff9","old_file":"haas\/tests\/test_coverage.py","new_file":"haas\/tests\/test_coverage.py","old_contents":"# -*- coding: utf-8 -*-\n# Copyright (c) 2013-2014 Simon Jagoe\n# All rights reserved.\n#\n# This software may be modified and distributed under the terms\n# of the 3-clause BSD license. See the LICENSE.txt file for details.\nfrom __future__ import absolute_import, unicode_literals\n\ntry:\n import coverage\nexcept ImportError:\n coverage = None\n\nfrom mock import Mock, patch\n\nfrom ..coverage import Coverage\nfrom ..testing import unittest\n\n\n@unittest.skipIf(coverage is None, 'Coverage is not installed')\nclass TestCoverage(unittest.TestCase):\n\n @patch('coverage.coverage')\n def test_coverage(self, coverage_func):\n coverage_object = Mock()\n coverage_func.return_value = coverage_object\n coverage_object.start = Mock()\n coverage_object.stop = Mock()\n coverage_object.save = Mock()\n\n cov = Coverage()\n coverage_func.assert_called_once_with()\n cov.setup()\n coverage_object.start.assert_called_once_with()\n self.assertFalse(coverage_object.stop.called)\n self.assertFalse(coverage_object.save.called)\n cov.teardown()\n coverage_object.stop.assert_called_once_with()\n coverage_object.save.assert_called_once_with()\n","new_contents":"# -*- coding: utf-8 -*-\n# Copyright (c) 2013-2014 Simon Jagoe\n# All rights reserved.\n#\n# This software may be modified and distributed under the terms\n# of the 3-clause BSD license. See the LICENSE.txt file for details.\nfrom __future__ import absolute_import, unicode_literals\n\ntry:\n import coverage\n from ..coverage import Coverage\nexcept ImportError:\n coverage = None\n Coverage = None\n\nfrom mock import Mock, patch\n\nfrom ..testing import unittest\n\n\n@unittest.skipIf(coverage is None, 'Coverage is not installed')\nclass TestCoverage(unittest.TestCase):\n\n @patch('coverage.coverage')\n def test_coverage(self, coverage_func):\n coverage_object = Mock()\n coverage_func.return_value = coverage_object\n coverage_object.start = Mock()\n coverage_object.stop = Mock()\n coverage_object.save = Mock()\n\n cov = Coverage()\n coverage_func.assert_called_once_with()\n cov.setup()\n coverage_object.start.assert_called_once_with()\n self.assertFalse(coverage_object.stop.called)\n self.assertFalse(coverage_object.save.called)\n cov.teardown()\n coverage_object.stop.assert_called_once_with()\n coverage_object.save.assert_called_once_with()\n","subject":"Fix test error when coverage is not installed","message":"Fix test error when coverage is not installed\n","lang":"Python","license":"bsd-3-clause","repos":"itziakos\/haas,scalative\/haas,sjagoe\/haas,sjagoe\/haas,scalative\/haas,itziakos\/haas"} {"commit":"5b155b6e6b09874d7783135528d39ab7bbc61fdb","old_file":"config\/__init__.py","new_file":"config\/__init__.py","old_contents":"import logging\n\nimport yaml\n\n\nl = logging.getLogger(__name__)\n\n\ndef _replace_with_type(type_, replace_type, data):\n if isinstance(data, type_):\n return replace_type(data)\n return data\n\n\nclass Config(dict):\n\n def __init__(self, items=None):\n if items is not None:\n if hasattr(items, 'items'):\n items = list(items.items())\n for i, (k, v) in enumerate(items):\n items[i] = (k, _replace_with_type(dict, Config, v))\n super().__init__(items)\n else:\n super().__init__()\n\n def __getattr__(self, key):\n if key in self:\n return self[key]\n else:\n l.warn(\"AttrDict: did not find key '{}' in keys {}\", key, self.keys())\n\n if l.getEffectiveLevel() <= logging.INFO:\n import inspect\n stack = inspect.stack(1)[1:]\n l.info(\"-- AttrDict stack --\")\n for info in reversed(stack):\n l.info(' File \"{0[1]}\", line {0[2]}, in {0[3]} -- {1}',\n info, info[4][-1].strip())\n l.info(\"-- AttrDict stack -- end\")\n\n return Config() # return empty 'dict' as default\n\n\ndef read_file(filename):\n l.debug(\"reading config file: '{}'\", filename)\n with open(filename) as f:\n config = Config(yaml.safe_load(f))\n l.debug(\"config: {!s}\", config)\n return config\n","new_contents":"import logging\n\nimport yaml\n\n\nl = logging.getLogger(__name__)\n\n\ndef _replace_with_type(type_, replace_type, data):\n if isinstance(data, type_):\n return replace_type(data)\n return data\n\n\nclass Config(dict):\n\n def __init__(self, items=None):\n if items is not None:\n if hasattr(items, 'items'):\n items = list(items.items())\n for i, (k, v) in enumerate(items):\n items[i] = (k, _replace_with_type(dict, self.__class__, v))\n super().__init__(items)\n else:\n super().__init__()\n\n def __getattr__(self, key):\n if key in self:\n return self[key]\n else:\n l.warn(\"AttrDict: did not find key '{}' in {}\", key, self.keys())\n\n if l.getEffectiveLevel() <= logging.INFO:\n import inspect\n stack = inspect.stack(1)[1:]\n l.info(\"-- AttrDict stack --\")\n for info in reversed(stack):\n l.info(' File \"{0[1]}\", line {0[2]}, in {0[3]} -- {1}',\n info, info[4][-1].strip())\n l.info(\"-- AttrDict stack -- end\")\n\n return self.__class__() # return empty 'Config' as default\n\n\ndef read_file(filename):\n l.debug(\"reading config file: '{}'\", filename)\n with open(filename) as f:\n config = Config(yaml.safe_load(f))\n l.debug(\"config: {!s}\", config)\n return config\n","subject":"Use dynamic self.__class__ and not name directly","message":"Use dynamic self.__class__ and not name directly\n","lang":"Python","license":"mit","repos":"FichteFoll\/TelegramIRCImageProxy,codetalkio\/TelegramIRCImageProxy,FichteFoll\/CodetalkIRCBot"} {"commit":"49bf8bd8137928a1dc5165f38f8abfe423f5e7f0","old_file":"pi_director\/controllers\/user_controls.py","new_file":"pi_director\/controllers\/user_controls.py","old_contents":"from pyramid.response import Response\nfrom pi_director.models.models import (\n DBSession,\n MyModel,\n )\nfrom pi_director.models.UserModel import UserModel\n\n\ndef authorize_user(email):\n user=DBSession.query(UserModel).filter(UserModel.email==email).one()\n user.AccessLevel=2\n DBSession.flush()\n\ndef delete_user(email):\n DBSession.query(UserModel).filter(UserModel.email==email).delete()\n\ndef get_users():\n UserList=DBSession.query(UserModel).all()\n return UserList\n\n\ndef make_an_admin(request):\n email=request.matchdict['email']\n\n '''First, make sure there aren't already admins in the system'''\n res=DBSession.query(UserModel).filter(UserModel.AccessLevel==2).first()\n if res != None:\n msg=\"User already an admin: {user}\".format(user=res.email)\n return False\n user=DBSession.query(UserModel).filter(UserModel.email==email).first()\n user.AccessLevel=2\n DBSession.flush() \n return True\n\n\n","new_contents":"from pyramid.response import Response\nfrom pi_director.models.models import (\n DBSession,\n MyModel,\n )\nfrom pi_director.models.UserModel import UserModel\n\n\ndef authorize_user(email):\n user=DBSession.query(UserModel).filter(UserModel.email==email).one()\n user.AccessLevel=2\n DBSession.flush()\n\ndef delete_user(email):\n DBSession.query(UserModel).filter(UserModel.email==email).delete()\n\ndef get_users():\n UserList=DBSession.query(UserModel).all()\n return UserList\n\n\ndef make_an_admin(request):\n email=request.matchdict['email']\n\n '''First, make sure there aren't already admins in the system'''\n res=DBSession.query(UserModel).filter(UserModel.AccessLevel==2).first()\n if res != None:\n msg=\"User already an admin: {user}\".format(user=res.email)\n return False\n user=DBSession.query(UserModel).filter(UserModel.email==email).first()\n if user == None:\n user=UserModel()\n user.email=email\n DBSession.add(user)\n user.AccessLevel=2\n DBSession.flush() \n return True\n\n\n","subject":"Create the user if it isn't already in the database first, then make it an admin.","message":"Create the user if it isn't already in the database first, then make it an admin.\n","lang":"Python","license":"mit","repos":"selfcommit\/pi_director,PeterGrace\/pi_director,selfcommit\/pi_director,PeterGrace\/pi_director,PeterGrace\/pi_director,selfcommit\/pi_director"} {"commit":"6d450dccc7e89e4e90fd1f0f27cdf2aa67166859","old_file":"conanfile.py","new_file":"conanfile.py","old_contents":"from conans import ConanFile, CMake\n\n\nclass SocketwConan(ConanFile):\n name = \"SocketW\"\n version = \"3.10.36\"\n license = \"GNU Lesser General Public License v2.1\"\n url = \"https:\/\/github.com\/RigsOfRods\/socketw\/issues\"\n description = \"SocketW is a library which provides cross-platform socket abstraction\"\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n #options = {\"shared\": [True, False]}\n #default_options = \"shared=False\"\n generators = \"cmake\"\n exports_sources = \"src\/*\", \"CMakeLists.txt\", \"LICENSE\", \"README\"\n\n\n def requirements(self):\n self.requires.add('OpenSSL\/1.0.2@conan\/stable')\n\n def build(self):\n cmake = CMake(self)\n cmake.configure()\n cmake.build()\n\n def package(self):\n cmake = CMake(self)\n cmake.install()\n\n def package_info(self):\n self.cpp_info.libs = [\"SocketW\"]\n","new_contents":"from conans import ConanFile, CMake\n\n\nclass SocketwConan(ConanFile):\n name = \"SocketW\"\n version = \"3.10.36\"\n license = \"GNU Lesser General Public License v2.1\"\n url = \"https:\/\/github.com\/RigsOfRods\/socketw\/issues\"\n description = \"SocketW is a library which provides cross-platform socket abstraction\"\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n #options = {\"shared\": [True, False]}\n #default_options = \"shared=False\"\n generators = \"cmake\"\n exports_sources = \"src\/*\", \"CMakeLists.txt\", \"LICENSE\", \"README\"\n\n\n def requirements(self):\n self.requires.add('OpenSSL\/1.0.2@conan\/stable')\n\n def build(self):\n cmake = CMake(self)\n cmake.configure()\n cmake.build()\n\n def package(self):\n cmake = CMake(self)\n cmake.install()\n\n def package_info(self):\n self.cpp_info.libs = tools.collect_libs(self)\n","subject":"Use collect_libs for finding libs","message":"Use collect_libs for finding libs","lang":"Python","license":"lgpl-2.1","repos":"Hiradur\/mysocketw,Hiradur\/mysocketw"} {"commit":"445db7dedb7c60edc03caa1969dca4253e2b9048","old_file":"example_project\/example_project\/web\/templatetags\/testags.py","new_file":"example_project\/example_project\/web\/templatetags\/testags.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom django_jinja.base import Library\nimport jinja2\n\nregister = Library()\n\n@register.filter\n@jinja2.contextfilter\ndef datetimeformat(ctx, value, format='%H:%M \/ %d-%m-%Y'):\n return value.strftime(format)\n\n@register.global_function\ndef hello(name):\n return \"Hello\" + name\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom django_jinja.library import Library\nimport jinja2\n\nregister = Library()\n\n@register.filter\n@jinja2.contextfilter\ndef datetimeformat(ctx, value, format='%H:%M \/ %d-%m-%Y'):\n return value.strftime(format)\n\n@register.global_function\ndef hello(name):\n return \"Hello\" + name\n","subject":"Adjust the example_projects to work","message":"Adjust the example_projects to work\n","lang":"Python","license":"bsd-3-clause","repos":"glogiotatidis\/django-jinja,akx\/django-jinja,akx\/django-jinja,niwinz\/django-jinja,niwinz\/django-jinja,akx\/django-jinja,niwinz\/django-jinja,akx\/django-jinja,glogiotatidis\/django-jinja,glogiotatidis\/django-jinja,glogiotatidis\/django-jinja"} {"commit":"0a02b896c7f8499504a855652de22bab10824c69","old_file":"database_setup.py","new_file":"database_setup.py","old_contents":"import sys\nfrom sqlalchemy import Column, ForeignKey, Integer, String\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import relationship\nfrom sqlalchemy import create_engine\n\nBase = declarative_base()\n\n\nclass Restaurant(Base):\n\n __tablename__ = 'restaurant'\n\n name = Column(String(80), nullable = False)\n id = Column(Integer, primary_key = True)\n\n @property\n def serialize(self):\n return {\n 'name': self.name,\n 'id': self.id,\n }\n\nclass MenuItem(Base):\n\n __tablename__ = 'menu_item'\n\n name = Column(String(80), nullable = False)\n id = Column(Integer,primary_key = True)\n course = Column(String(250))\n description = Column(String(250))\n price = Column(String(8))\n restaurant_id = Column(Integer, ForeignKey('restaurant.id'))\n restaurant = relationship(Restaurant)\n\n @property\n def serialize(self):\n return {\n 'name': self.name,\n 'description': self.description,\n 'id': self.id,\n 'price': self.price,\n 'course': self.course,\n }\n\n\nengine = create_engine('sqlite:\/\/\/restaurantmenu.db')\nBase.metadata.create_all(engine)\n","new_contents":"import sys\nfrom sqlalchemy import Column, ForeignKey, Integer, String\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.orm import relationship\nfrom sqlalchemy import create_engine\n\nBase = declarative_base()\n\n\nclass Restaurant(Base):\n\n __tablename__ = 'restaurant'\n\n name = Column(String(80), nullable = False)\n description = Column(String(250))\n id = Column(Integer, primary_key = True)\n\n @property\n def serialize(self):\n return {\n 'name': self.name,\n 'description': self.description,\n 'id': self.id,\n }\n\nclass MenuItem(Base):\n\n __tablename__ = 'menu_item'\n\n name = Column(String(80), nullable = False)\n id = Column(Integer,primary_key = True)\n course = Column(String(250))\n description = Column(String(250))\n price = Column(String(8))\n restaurant_id = Column(Integer, ForeignKey('restaurant.id'))\n restaurant = relationship(Restaurant)\n\n @property\n def serialize(self):\n return {\n 'name': self.name,\n 'description': self.description,\n 'id': self.id,\n 'price': self.price,\n 'course': self.course,\n }\n\n\nengine = create_engine('sqlite:\/\/\/restaurantmenu.db')\nBase.metadata.create_all(engine)\n","subject":"Add description column to Restaurant","message":"feat: Add description column to Restaurant\n","lang":"Python","license":"mit","repos":"gsbullmer\/restaurant-menu-directory,gsbullmer\/restaurant-menu-directory"} {"commit":"3e02a38a9ae52603f620a7969ce532b61de531d7","old_file":"libgreader\/__init__.py","new_file":"libgreader\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n\n# libgreader\n# Copyright (C) 2012 Matt Behrens \n# Python library for the Google Reader API\n\n__author__ = \"Matt Behrens \"\n__version__ = \"0.8.0\"\n__copyright__ = \"Copyright (C) 2012 Matt Behrens\"\n\nfrom .googlereader import GoogleReader\nfrom .auth import AuthenticationMethod, ClientAuthMethod, OAuthMethod, OAuth2Method\nfrom .items import *\nfrom .url import ReaderUrl\n","new_contents":"# -*- coding: utf-8 -*-\n\n# libgreader\n# Copyright (C) 2012 Matt Behrens \n# Python library for the Google Reader API\n\n__author__ = \"Matt Behrens \"\n__version__ = \"0.8.0\"\n__copyright__ = \"Copyright (C) 2012 Matt Behrens\"\n\ntry:\n import requests\nexcept ImportError:\n # Will occur during setup.py install\n pass\nelse:\n from .googlereader import GoogleReader\n from .auth import AuthenticationMethod, ClientAuthMethod, OAuthMethod, OAuth2Method\n from .items import *\n from .url import ReaderUrl\n","subject":"Fix import error during setup.py install","message":"Fix import error during setup.py install\n","lang":"Python","license":"mit","repos":"smurfix\/librssreader,askedrelic\/libgreader"} {"commit":"3d8d82be3528cc0150dac0c8ade1f6c306b412e4","old_file":"channels\/apps.py","new_file":"channels\/apps.py","old_contents":"from django.apps import AppConfig\n\n#from .binding.base import BindingMetaclass\nfrom .package_checks import check_all\n\n\nclass ChannelsConfig(AppConfig):\n\n name = \"channels\"\n verbose_name = \"Channels\"\n\n def ready(self):\n # Check versions\n check_all()\n # Do django monkeypatches\n from .hacks import monkeypatch_django\n monkeypatch_django()\n # Instantiate bindings\n #BindingMetaclass.register_all()\n","new_contents":"from django.apps import AppConfig\n\n# We import this here to ensure the reactor is installed very early on\n# in case other packages accidentally import twisted.internet.reactor\n# (e.g. raven does this).\nimport daphne.server # noqa\n\n#from .binding.base import BindingMetaclass\nfrom .package_checks import check_all\n\n\nclass ChannelsConfig(AppConfig):\n\n name = \"channels\"\n verbose_name = \"Channels\"\n\n def ready(self):\n # Check versions\n check_all()\n # Do django monkeypatches\n from .hacks import monkeypatch_django\n monkeypatch_django()\n # Instantiate bindings\n #BindingMetaclass.register_all()\n","subject":"Add early import to fix problems with other packages and Twisted.","message":"Add early import to fix problems with other packages and Twisted.\n","lang":"Python","license":"bsd-3-clause","repos":"andrewgodwin\/channels,andrewgodwin\/django-channels,django\/channels"} {"commit":"b94f849fe28918a343a142da57b6055064d5b194","old_file":"tests\/test_abort_generate_on_hook_error.py","new_file":"tests\/test_abort_generate_on_hook_error.py","old_contents":"# -*- coding: utf-8 -*-\n\nimport pytest\n\nfrom cookiecutter import generate\nfrom cookiecutter import exceptions\n\n\n@pytest.mark.usefixtures('clean_system')\ndef test_pre_gen_hook(tmpdir):\n context = {\n 'cookiecutter': {\n \"repo_dir\": \"foobar\",\n \"abort_pre_gen\": \"yes\",\n \"abort_post_gen\": \"no\"\n }\n }\n\n with pytest.raises(exceptions.FailedHookException):\n generate.generate_files(\n repo_dir='tests\/hooks-abort-render',\n context=context,\n output_dir=str(tmpdir)\n )\n\n assert not tmpdir.join('foobar').isdir()\n\n","new_contents":"# -*- coding: utf-8 -*-\n\nimport pytest\n\nfrom cookiecutter import generate\nfrom cookiecutter import exceptions\n\n\n@pytest.mark.usefixtures('clean_system')\ndef test_pre_gen_hook(tmpdir):\n context = {\n 'cookiecutter': {\n \"repo_dir\": \"foobar\",\n \"abort_pre_gen\": \"yes\",\n \"abort_post_gen\": \"no\"\n }\n }\n\n with pytest.raises(exceptions.FailedHookException):\n generate.generate_files(\n repo_dir='tests\/hooks-abort-render',\n context=context,\n output_dir=str(tmpdir)\n )\n\n assert not tmpdir.join('foobar').isdir()\n\n\n@pytest.mark.usefixtures('clean_system')\ndef test_post_gen_hook(tmpdir):\n context = {\n 'cookiecutter': {\n \"repo_dir\": \"foobar\",\n \"abort_pre_gen\": \"no\",\n \"abort_post_gen\": \"yes\"\n }\n }\n\n with pytest.raises(exceptions.FailedHookException):\n generate.generate_files(\n repo_dir='tests\/hooks-abort-render',\n context=context,\n output_dir=str(tmpdir)\n )\n\n assert not tmpdir.join('foobar').isdir()\n","subject":"Test that an error in post_gen_project aborts generation","message":"Test that an error in post_gen_project aborts generation\n","lang":"Python","license":"bsd-3-clause","repos":"dajose\/cookiecutter,pjbull\/cookiecutter,willingc\/cookiecutter,audreyr\/cookiecutter,michaeljoseph\/cookiecutter,terryjbates\/cookiecutter,hackebrot\/cookiecutter,Springerle\/cookiecutter,dajose\/cookiecutter,audreyr\/cookiecutter,stevepiercy\/cookiecutter,terryjbates\/cookiecutter,hackebrot\/cookiecutter,pjbull\/cookiecutter,willingc\/cookiecutter,luzfcb\/cookiecutter,michaeljoseph\/cookiecutter,stevepiercy\/cookiecutter,Springerle\/cookiecutter,luzfcb\/cookiecutter"} {"commit":"1f66670b94d2eca70ecf8e26b21f8b28986154b9","old_file":"test-mm.py","new_file":"test-mm.py","old_contents":"from psautohint import autohint\nfrom psautohint import psautohint\n\nbaseDir = \"tests\/data\/source-code-pro\"\nmasters = (\"Black\", \"Bold\", \"ExtraLight\", \"Light\", \"Medium\", \"Regular\", \"Semibold\")\n\nglyphList = None\n\nfonts = []\nfor master in masters:\n print(\"Hinting %s\" % master)\n\n path = \"%s\/%s\/font.otf\" % (baseDir, master)\n font = autohint.openOpenTypeFile(path, \"font.otf\", None)\n names = font.getGlyphList()\n info = font.getFontInfo(font.getPSName(), path, False, False, [], [])\n info = info.getFontInfo()\n\n if glyphList is None:\n glyphList = names\n else:\n assert glyphList == names\n\n glyphs = []\n for name in names:\n glyph = font.convertToBez(name, False)\n glyphs.append(glyph[0])\n fonts.append(psautohint.autohint(info, glyphs, False, False, False))\n\nglyphs = []\nfor i in range(len(glyphList)):\n glyphs.append([f[i] for f in fonts])\n\nprint(\"MM Hinting\")\nglyphs = psautohint.autohintmm(info, glyphs, masters, True)\n","new_contents":"from psautohint import autohint\nfrom psautohint import psautohint\n\nbaseDir = \"tests\/data\/source-code-pro\"\nmasters = (\"Black\", \"Bold\", \"ExtraLight\", \"Light\", \"Medium\", \"Regular\", \"Semibold\")\n\nglyphList = None\n\nfonts = []\nfor master in masters:\n print(\"Hinting %s\" % master)\n\n options = autohint.ACOptions()\n options.quiet = True\n\n path = \"%s\/%s\/font.ufo\" % (baseDir, master)\n font = autohint.openUFOFile(path, \"font.ufo\", False, options)\n font.useProcessedLayer = False\n names = font.getGlyphList()\n info = font.getFontInfo(font.getPSName(), path, False, False, [], [])\n info = info.getFontInfo()\n\n if glyphList is None:\n glyphList = names\n else:\n assert glyphList == names\n\n glyphs = []\n for name in glyphList:\n glyph = font.convertToBez(name, False, True)[0]\n if not glyph:\n glyph = \"%%%s\\n\" % name\n glyphs.append(glyph)\n fonts.append(psautohint.autohint(info, glyphs, False, False, False))\n\nglyphs = []\nfor i in range(len(glyphList)):\n glyphs.append([f[i] for f in fonts])\n\nprint(\"MM Hinting\")\nglyphs = psautohint.autohintmm(info, glyphs, masters, True)\n","subject":"Use the UFOs not the OTFs","message":"Use the UFOs not the OTFs\n\nOops, the OTF are not interpolation compatible due to overlap removal, I\nshould have use the UFOs all along. Now the script passes without\nerrors, still need to verify the output.\n","lang":"Python","license":"apache-2.0","repos":"khaledhosny\/psautohint,khaledhosny\/psautohint"} {"commit":"7f6da4dee6464e48a0e6b491f3f740a750e86ed2","old_file":"dataactcore\/scripts\/resetAlembicVersion.py","new_file":"dataactcore\/scripts\/resetAlembicVersion.py","old_contents":"import argparse\nfrom dataactcore.models.errorInterface import ErrorInterface\nfrom dataactcore.models.jobTrackerInterface import JobTrackerInterface\nfrom dataactcore.models.userInterface import UserInterface\nfrom dataactcore.models.validationInterface import ValidationInterface\nfrom sqlalchemy import MetaData, Table\nfrom sqlalchemy.orm import sessionmaker\nfrom sqlalchemy.sql import update\n\n\ndef reset_alembic(alembic_version):\n\n engine_list = [\n ErrorInterface().engine,\n JobTrackerInterface().engine,\n UserInterface().engine,\n ValidationInterface().engine,\n ]\n\n for e in engine_list:\n Session = sessionmaker(bind=e)\n session = Session()\n metadata = MetaData(bind=e)\n alembic_table = Table('alembic_version', metadata, autoload=True)\n u = update(alembic_table)\n u = u.values({\"version_num\": alembic_version})\n session.execute(u)\n session.commit()\n\nparser = argparse.ArgumentParser\\\n (description=\"Reset alembic version tables across broker databases.\")\nparser.add_argument(\n 'version', help=\"Version to set the Alembic migration tables to.\")\nv = vars(parser.parse_args())['version']\nreset_alembic(v)\n\n","new_contents":"import argparse\n\nfrom sqlalchemy import MetaData, Table\nfrom sqlalchemy.sql import update\n\nfrom dataactcore.interfaces.db import GlobalDB\nfrom dataactvalidator.app import createApp\n\n\ndef reset_alembic(alembic_version):\n\n with createApp().app_context():\n db = GlobalDB.db()\n\n engine = db.engine\n sess = db.session\n metadata = MetaData(bind=engine)\n alembic_table = Table('alembic_version', metadata, autoload=True)\n u = update(alembic_table)\n u = u.values({\"version_num\": alembic_version})\n sess.execute(u)\n sess.commit()\n\nparser = argparse.ArgumentParser\\\n (description=\"Reset alembic version table.\")\nparser.add_argument(\n 'version', help=\"Version to set the Alembic migration table to.\")\nv = vars(parser.parse_args())['version']\nreset_alembic(v)\n\n","subject":"Remove db interfaces from the alembic version reset helper script.","message":"Remove db interfaces from the alembic version reset helper script.\n\nArguably, this script is no longer especially useful now that we\nonly have a single database for the broker. That said, removed the\ninterfaces in case folks are still using it.\n","lang":"Python","license":"cc0-1.0","repos":"fedspendingtransparency\/data-act-broker-backend,chambers-brian\/SIG_Digital-Strategy_SI_ODP_Backend,chambers-brian\/SIG_Digital-Strategy_SI_ODP_Backend,fedspendingtransparency\/data-act-broker-backend"} {"commit":"14329daf571400812594c0388eac87538cd10079","old_file":"denim\/api.py","new_file":"denim\/api.py","old_contents":"from fabric import api as __api\n\n# Setup some default values.\n__api.env.deploy_user = 'webapps'\n\nfrom denim.paths import (cd_deploy, cd_package, deploy_path, package_path)\nfrom denim import (scm, service, system, virtualenv, webserver)\nfrom denim.decorators import deploy_env\n\n\n@__api.task(name=\"help\")\ndef show_help():\n \"\"\"\n Help on common operations.\n \"\"\"\n from denim.environment import get_environments\n import denim\n\n print \"\"\"\nCommon operations with Denim (%(version)s).\n\nProvision server:\n> fab {%(environments)s} init\n\nDeploy (require a source control revision to be supplied. i.e. master):\n> fab {%(environments)s} deploy:{revision}\n\nStatus of service:\n> fab {%(environments)s} service.status\n\n\"\"\" % {\n 'environments': '|'.join(get_environments()),\n 'version': denim.__version__,\n }\n\n\n@__api.task\ndef environment():\n \"\"\"\n Environments defined in fabfile.\n \"\"\"\n from denim.environment import get_environments\n print 'Environments defined in fab file:'\n print ', '.join(get_environments())\n","new_contents":"from fabric import api as _api\n\n# Setup some default values.\n_api.env.deploy_user = 'webapps'\n\nfrom denim.paths import (cd_deploy, cd_application, deploy_path, application_path)\nfrom denim import (scm, service, system, virtualenv, webserver)\nfrom denim.decorators import deploy_env\n\n# Pending deprecation\nfrom denim.paths import (cd_package, package_path)\n\n@_api.task(name=\"help\")\ndef show_help():\n \"\"\"\n Help on common operations.\n \"\"\"\n from denim.environment import get_environments\n import denim\n\n print \"\"\"\nCommon operations with Denim (%(version)s).\n\nProvision server:\n> fab {%(environments)s} init\n\nDeploy (require a source control revision to be supplied. i.e. master):\n> fab {%(environments)s} deploy:{revision}\n\nStatus of service:\n> fab {%(environments)s} service.status\n\n\"\"\" % {\n 'environments': '|'.join(get_environments()),\n 'version': denim.__version__,\n }\n\n\n@_api.task\ndef environments():\n \"\"\"\n Environments defined in fabfile.\n \"\"\"\n from denim.environment import get_environments\n print 'Environments defined in fab file:'\n print ', '.join(get_environments())\n","subject":"Break out items pending deprecation, remove double underscores","message":"Break out items pending deprecation, remove double underscores\n","lang":"Python","license":"bsd-2-clause","repos":"timsavage\/denim"} {"commit":"770f9dd75a223fb31a18af2fcb089398663f2065","old_file":"concentration.py","new_file":"concentration.py","old_contents":"from major import Major\n\nclass Concentration(Major):\n\tdef __init__(self, dept=\"NONE\"):\n\t\tsuper().__init__(dept, path=\"concentrations\/\")\n\nif __name__ == '__main__':\n\ttmp = [\n\t\tConcentration(dept=\"Asian\")\n\t]\n\tfor i in tmp:\n\t\tprint(i)\n","new_contents":"from major import Major\n\nclass Concentration(Major):\n\tdef __init__(self, dept=\"NONE\"):\n\t\tsuper().__init__(dept, path=\"concentrations\/\")\n\n\tdef getConcentrationRequirement(self, string):\n\t\treturn self.requirements[string]\n\nif __name__ == '__main__':\n\ttmp = [\n\t\tConcentration(dept=\"Asian\")\n\t]\n\tfor i in tmp:\n\t\tprint(i)\n","subject":"Add a getConcentrationRequirement to corrospond to getMajorRequirement","message":"Add a getConcentrationRequirement to corrospond to getMajorRequirement\n","lang":"Python","license":"agpl-3.0","repos":"hawkrives\/gobbldygook,hawkrives\/gobbldygook,hawkrives\/gobbldygook"} {"commit":"1f61ece6f6b0950706ebef159665eafbfeeaa1fd","old_file":"app\/api\/utils\/containerMapper.py","new_file":"app\/api\/utils\/containerMapper.py","old_contents":"\n\ndef getContainerDetails(container):\n ip = 'N\/A'\n if container.state().network != None and container.state().network.get('eth0') != None:\n if len(container.state().network.get('eth0')['addresses']) > 0:\n ip = container.state().network['eth0']['addresses'][0].get('address', 'N\/A')\n\n return {\n 'name': container.name,\n 'status': container.status,\n 'ip': ip,\n 'ephemeral': container.ephemeral,\n 'image': ''.join(container.config.get('image.os') + ' ' + container.config.get('image.release') + ' ' + container.config.get('image.architecture')),\n 'created_at': container.created_at\n }","new_contents":"\n\ndef getContainerDetails(container):\n ip = 'N\/A'\n if container.state().network != None and container.state().network.get('eth0') != None:\n if len(container.state().network.get('eth0')['addresses']) > 0:\n ip = container.state().network['eth0']['addresses'][0].get('address', 'N\/A')\n\n image = 'N\/A'\n if container.config.get('image.os') != None and container.config.get('image.release') != None and container.config.get('image.architecture') != None:\n image = ''.join(container.config.get('image.os') + ' ' + container.config.get('image.release') + ' ' + container.config.get('image.architecture'))\n\n return {\n 'name': container.name,\n 'status': container.status,\n 'ip': ip,\n 'ephemeral': container.ephemeral,\n 'image': image,\n 'created_at': container.created_at\n }","subject":"Fix container list bug when missing image","message":"Fix container list bug when missing image\n","lang":"Python","license":"apache-2.0","repos":"AdaptiveScale\/lxdui,AdaptiveScale\/lxdui,AdaptiveScale\/lxdui,AdaptiveScale\/lxdui"} {"commit":"26538b01b5865c4b5f68fc0ea0f22268d4768376","old_file":"scrapy\/contrib\/memdebug.py","new_file":"scrapy\/contrib\/memdebug.py","old_contents":"\"\"\"\nMemoryDebugger extension\n\nSee documentation in docs\/topics\/extensions.rst\n\"\"\"\n\nimport gc\n\nfrom scrapy import signals\nfrom scrapy.exceptions import NotConfigured\nfrom scrapy.utils.trackref import live_refs\n\n\nclass MemoryDebugger(object):\n\n def __init__(self, stats):\n self.stats = stats\n\n @classmethod\n def from_crawler(cls, crawler):\n if not crawler.settings.getbool('MEMDEBUG_ENABLED'):\n raise NotConfigured\n o = cls(crawler.stats)\n crawler.signals.connect(o.engine_stopped, signals.engine_stopped)\n return o\n\n def engine_stopped(self):\n gc.collect()\n self.stats.set_value('memdebug\/gc_garbage_count', len(gc.garbage))\n for cls, wdict in live_refs.iteritems():\n if not wdict:\n continue\n self.stats.set_value('memdebug\/live_refs\/%s' % cls.__name__, len(wdict))\n","new_contents":"\"\"\"\nMemoryDebugger extension\n\nSee documentation in docs\/topics\/extensions.rst\n\"\"\"\n\nimport gc\n\nfrom scrapy import signals\nfrom scrapy.exceptions import NotConfigured\nfrom scrapy.utils.trackref import live_refs\n\n\nclass MemoryDebugger(object):\n\n def __init__(self, stats):\n self.stats = stats\n\n @classmethod\n def from_crawler(cls, crawler):\n if not crawler.settings.getbool('MEMDEBUG_ENABLED'):\n raise NotConfigured\n o = cls(crawler.stats)\n crawler.signals.connect(o.spider_closed, signal=signals.spider_closed)\n return o\n\n def spider_closed(self, spider, reason):\n gc.collect()\n self.stats.set_value('memdebug\/gc_garbage_count', len(gc.garbage), spider=spider)\n for cls, wdict in live_refs.iteritems():\n if not wdict:\n continue\n self.stats.set_value('memdebug\/live_refs\/%s' % cls.__name__, len(wdict), spider=spider)\n","subject":"Fix logging of stats collected by MemoryDebugger extension.","message":"Fix logging of stats collected by MemoryDebugger extension.\n\nStats are printed on spider_closed event;\nengine_stopped signal is called after spider_closed signal,\nso stats for MemoryDebugger extension were not printed to user.\n","lang":"Python","license":"bsd-3-clause","repos":"olorz\/scrapy,OpenWhere\/scrapy,cursesun\/scrapy,pablohoffman\/scrapy,avtoritet\/scrapy,nikgr95\/scrapy,eLRuLL\/scrapy,rolando-contrib\/scrapy,amboxer21\/scrapy,aivarsk\/scrapy,zjuwangg\/scrapy,dacjames\/scrapy,cursesun\/scrapy,TarasRudnyk\/scrapy,Digenis\/scrapy,bmess\/scrapy,agreen\/scrapy,CENDARI\/scrapy,redapple\/scrapy,shaform\/scrapy,lacrazyboy\/scrapy,fontenele\/scrapy,scorphus\/scrapy,JacobStevenR\/scrapy,CENDARI\/scrapy,beni55\/scrapy,zackslash\/scrapy,liyy7\/scrapy,rdowinton\/scrapy,fontenele\/scrapy,bmess\/scrapy,smaty1\/scrapy,haiiiiiyun\/scrapy,huoxudong125\/scrapy,rolando-contrib\/scrapy,mlyundin\/scrapy,Timeship\/scrapy,amboxer21\/scrapy,fqul\/scrapy,stenskjaer\/scrapy,carlosp420\/scrapy,jdemaeyer\/scrapy,legendtkl\/scrapy,kimimj\/scrapy,ashishnerkar1\/scrapy,profjrr\/scrapy,gbirke\/scrapy,kashyap32\/scrapy,xiao26\/scrapy,foromer4\/scrapy,zjuwangg\/scrapy,jc0n\/scrapy,jiezhu2007\/scrapy,darkrho\/scrapy-scrapy,Allianzcortex\/scrapy,Digenis\/scrapy,zhangtao11\/scrapy,elacuesta\/scrapy,devGregA\/scrapy,TarasRudnyk\/scrapy,Lucifer-Kim\/scrapy,YeelerG\/scrapy,eLRuLL\/scrapy,KublaikhanGeek\/scrapy,cleydson\/scrapy,huoxudong125\/scrapy,raphaelfruneaux\/scrapy,nikgr95\/scrapy,GregoryVigoTorres\/scrapy,carlosp420\/scrapy,wzyuliyang\/scrapy,nguyenhongson03\/scrapy,IvanGavran\/scrapy,jdemaeyer\/scrapy,Bourneer\/scrapy,gbirke\/scrapy,xiao26\/scrapy,haiiiiiyun\/scrapy,elijah513\/scrapy,songfj\/scrapy,olafdietsche\/scrapy,ssteo\/scrapy,jiezhu2007\/scrapy,nguyenhongson03\/scrapy,KublaikhanGeek\/scrapy,redapple\/scrapy,jorik041\/scrapy,hyrole\/scrapy,umrashrf\/scrapy,Ryezhang\/scrapy,dgillis\/scrapy,pfctdayelise\/scrapy,sardok\/scrapy,rklabs\/scrapy,nguyenhongson03\/scrapy,crasker\/scrapy,arush0311\/scrapy,tagatac\/scrapy,w495\/scrapy,scrapy\/scrapy,Djlavoy\/scrapy,Allianzcortex\/scrapy,tliber\/scrapy,zhangtao11\/scrapy,yarikoptic\/scrapy,jeffreyjinfeng\/scrapy,jamesblunt\/scrapy,redapple\/scrapy,URXtech\/scrapy,dracony\/scrapy,fafaman\/scrapy,livepy\/scrapy,kalessin\/scrapy,cursesun\/scrapy,beni55\/scrapy,Lucifer-Kim\/scrapy,Chenmxs\/scrapy,hbwzhsh\/scrapy,zhangtao11\/scrapy,avtoritet\/scrapy,pranjalpatil\/scrapy,rolando\/scrapy,pawelmhm\/scrapy,ArturGaspar\/scrapy,aivarsk\/scrapy,fafaman\/scrapy,Preetwinder\/scrapy,Parlin-Galanodel\/scrapy,elijah513\/scrapy,yidongliu\/scrapy,kimimj\/scrapy,1yvT0s\/scrapy,barraponto\/scrapy,tliber\/scrapy,jc0n\/scrapy,aivarsk\/scrapy,rahulsharma1991\/scrapy,dhenyjarasandy\/scrapy,zorojean\/scrapy,kmike\/scrapy,Parlin-Galanodel\/scrapy,amboxer21\/scrapy,jiezhu2007\/scrapy,Geeglee\/scrapy,legendtkl\/scrapy,Slater-Victoroff\/scrapy,moraesnicol\/scrapy,hyrole\/scrapy,dgillis\/scrapy,starrify\/scrapy,sigma-random\/scrapy,chekunkov\/scrapy,pawelmhm\/scrapy,wzyuliyang\/scrapy,Preetwinder\/scrapy,legendtkl\/scrapy,fpy171\/scrapy,taito\/scrapy,Digenis\/scrapy,farhan0581\/scrapy,Chenmxs\/scrapy,nowopen\/scrapy,yusofm\/scrapy,scrapy\/scrapy,starrify\/scrapy,mgedmin\/scrapy,ashishnerkar1\/scrapy,scorphus\/scrapy,WilliamKinaan\/scrapy,Ryezhang\/scrapy,Preetwinder\/scrapy,ssteo\/scrapy,eliasdorneles\/scrapy,agusc\/scrapy,Ryezhang\/scrapy,zorojean\/scrapy,jc0n\/scrapy,godfreyy\/scrapy,wenyu1001\/scrapy,raphaelfruneaux\/scrapy,pfctdayelise\/scrapy,Cnfc19932\/scrapy,crasker\/scrapy,agusc\/scrapy,Lucifer-Kim\/scrapy,Geeglee\/scrapy,avtoritet\/scrapy,rolando-contrib\/scrapy,hwsyy\/scrapy,famorted\/scrapy,hwsyy\/scrapy,beni55\/scrapy,kmike\/scrapy,eLRuLL\/scrapy,GregoryVigoTorres\/scrapy,elijah513\/scrapy,rootAvish\/scrapy,cyrixhero\/scrapy,heamon7\/scrapy,github-account-because-they-want-it\/scrapy,rolando\/scrapy,yarikoptic\/scrapy,wangjun\/scrapy,Djlavoy\/scrapy,ssh-odoo\/scrapy,jorik041\/scrapy,zjuwangg\/scrapy,1yvT0s\/scrapy,Geeglee\/scrapy,olorz\/scrapy,agusc\/scrapy,coderabhishek\/scrapy,webmakin\/scrapy,agreen\/scrapy,CodeJuan\/scrapy,yarikoptic\/scrapy,hectoruelo\/scrapy,pranjalpatil\/scrapy,nikgr95\/scrapy,AaronTao1990\/scrapy,pombredanne\/scrapy,pablohoffman\/scrapy,rdowinton\/scrapy,Slater-Victoroff\/scrapy,ylcolala\/scrapy,cyberplant\/scrapy,lacrazyboy\/scrapy,Partoo\/scrapy,irwinlove\/scrapy,dgillis\/scrapy,pombredanne\/scrapy,Bourneer\/scrapy,taito\/scrapy,dracony\/scrapy,jeffreyjinfeng\/scrapy,chekunkov\/scrapy,Cnfc19932\/scrapy,csalazar\/scrapy,fafaman\/scrapy,rdowinton\/scrapy,tagatac\/scrapy,crasker\/scrapy,kazitanvirahsan\/scrapy,Parlin-Galanodel\/scrapy,hyrole\/scrapy,TarasRudnyk\/scrapy,mgedmin\/scrapy,wenyu1001\/scrapy,hansenDise\/scrapy,snowdream1314\/scrapy,tntC4stl3\/scrapy,liyy7\/scrapy,ssh-odoo\/scrapy,irwinlove\/scrapy,ramiro\/scrapy,webmakin\/scrapy,profjrr\/scrapy,huoxudong125\/scrapy,URXtech\/scrapy,raphaelfruneaux\/scrapy,dangra\/scrapy,IvanGavran\/scrapy,z-fork\/scrapy,foromer4\/scrapy,stenskjaer\/scrapy,wujuguang\/scrapy,curita\/scrapy,joshlk\/scrapy,OpenWhere\/scrapy,wangjun\/scrapy,jdemaeyer\/scrapy,ylcolala\/scrapy,sardok\/scrapy,rklabs\/scrapy,livepy\/scrapy,finfish\/scrapy,dangra\/scrapy,famorted\/scrapy,songfj\/scrapy,pranjalpatil\/scrapy,zackslash\/scrapy,kalessin\/scrapy,hectoruelo\/scrapy,olorz\/scrapy,Timeship\/scrapy,tagatac\/scrapy,smaty1\/scrapy,farhan0581\/scrapy,mgedmin\/scrapy,kazitanvirahsan\/scrapy,rootAvish\/scrapy,finfish\/scrapy,jeffreyjinfeng\/scrapy,snowdream1314\/scrapy,CENDARI\/scrapy,elacuesta\/scrapy,codebhendi\/scrapy,hwsyy\/scrapy,ramiro\/scrapy,fpy171\/scrapy,tliber\/scrapy,darkrho\/scrapy-scrapy,umrashrf\/scrapy,foromer4\/scrapy,fpy171\/scrapy,darkrho\/scrapy-scrapy,dacjames\/scrapy,mlyundin\/scrapy,wzyuliyang\/scrapy,johnardavies\/scrapy,mlyundin\/scrapy,xiao26\/scrapy,AaronTao1990\/scrapy,1yvT0s\/scrapy,AaronTao1990\/scrapy,bmess\/scrapy,snowdream1314\/scrapy,johnardavies\/scrapy,dracony\/scrapy,shaform\/scrapy,stenskjaer\/scrapy,Djlavoy\/scrapy,pawelmhm\/scrapy,cleydson\/scrapy,webmakin\/scrapy,kashyap32\/scrapy,moraesnicol\/scrapy,olafdietsche\/scrapy,hbwzhsh\/scrapy,GregoryVigoTorres\/scrapy,Zephor5\/scrapy,cyrixhero\/scrapy,barraponto\/scrapy,scrapy\/scrapy,Bourneer\/scrapy,URXtech\/scrapy,livepy\/scrapy,arush0311\/scrapy,IvanGavran\/scrapy,elacuesta\/scrapy,ArturGaspar\/scrapy,profjrr\/scrapy,zorojean\/scrapy,OpenWhere\/scrapy,coderabhishek\/scrapy,shaform\/scrapy,famorted\/scrapy,finfish\/scrapy,umrashrf\/scrapy,farhan0581\/scrapy,Timeship\/scrapy,ENjOyAbLE1991\/scrapy,carlosp420\/scrapy,Zephor5\/scrapy,eliasdorneles\/scrapy,devGregA\/scrapy,jamesblunt\/scrapy,z-fork\/scrapy,rahulsharma1991\/scrapy,ENjOyAbLE1991\/scrapy,pombredanne\/scrapy,kazitanvirahsan\/scrapy,wujuguang\/scrapy,Adai0808\/scrapy-1,chekunkov\/scrapy,github-account-because-they-want-it\/scrapy,Chenmxs\/scrapy,godfreyy\/scrapy,WilliamKinaan\/scrapy,moraesnicol\/scrapy,JacobStevenR\/scrapy,dhenyjarasandy\/scrapy,ssteo\/scrapy,nfunato\/scrapy,heamon7\/scrapy,curita\/scrapy,zackslash\/scrapy,z-fork\/scrapy,kimimj\/scrapy,hbwzhsh\/scrapy,sigma-random\/scrapy,ylcolala\/scrapy,dhenyjarasandy\/scrapy,tntC4stl3\/scrapy,w495\/scrapy,devGregA\/scrapy,Adai0808\/scrapy-1,rklabs\/scrapy,agreen\/scrapy,codebhendi\/scrapy,olafdietsche\/scrapy,hectoruelo\/scrapy,csalazar\/scrapy,codebhendi\/scrapy,Partoo\/scrapy,pablohoffman\/scrapy,dacjames\/scrapy,barraponto\/scrapy,WilliamKinaan\/scrapy,dangra\/scrapy,irwinlove\/scrapy,godfreyy\/scrapy,arush0311\/scrapy,wenyu1001\/scrapy,wujuguang\/scrapy,kmike\/scrapy,wangjun\/scrapy,Cnfc19932\/scrapy,Partoo\/scrapy,yusofm\/scrapy,curita\/scrapy,github-account-because-they-want-it\/scrapy,JacobStevenR\/scrapy,rootAvish\/scrapy,fqul\/scrapy,Zephor5\/scrapy,fontenele\/scrapy,Slater-Victoroff\/scrapy,csalazar\/scrapy,YeelerG\/scrapy,kashyap32\/scrapy,ArturGaspar\/scrapy,scorphus\/scrapy,CodeJuan\/scrapy,rolando\/scrapy,heamon7\/scrapy,cyberplant\/scrapy,KublaikhanGeek\/scrapy,Adai0808\/scrapy-1,tntC4stl3\/scrapy,nowopen\/scrapy,rahulsharma1991\/scrapy,lacrazyboy\/scrapy,yidongliu\/scrapy,ENjOyAbLE1991\/scrapy,YeelerG\/scrapy,haiiiiiyun\/scrapy,nfunato\/scrapy,joshlk\/scrapy,nowopen\/scrapy,starrify\/scrapy,cleydson\/scrapy,cyrixhero\/scrapy,ssh-odoo\/scrapy,yusofm\/scrapy,w495\/scrapy,hansenDise\/scrapy,cyberplant\/scrapy,songfj\/scrapy,coderabhishek\/scrapy,hansenDise\/scrapy,CodeJuan\/scrapy,kalessin\/scrapy,liyy7\/scrapy,Allianzcortex\/scrapy,jorik041\/scrapy,johnardavies\/scrapy,yidongliu\/scrapy,ramiro\/scrapy,eliasdorneles\/scrapy,fqul\/scrapy,taito\/scrapy,joshlk\/scrapy,nfunato\/scrapy,smaty1\/scrapy,pfctdayelise\/scrapy"} {"commit":"9d7eb3a939a7189ed2d45fe13c1d629e850d8c16","old_file":"julia\/release.py","new_file":"julia\/release.py","old_contents":"# This file is executed via setup.py and imported via __init__.py\n\n__version__ = \"0.2.0.dev\"\n# For Python versioning scheme, see:\n# https:\/\/www.python.org\/dev\/peps\/pep-0440\/#version-scheme\n","new_contents":"# This file is executed via setup.py and imported via __init__.py\n\n__version__ = \"0.2.0\"\n# For Python versioning scheme, see:\n# https:\/\/www.python.org\/dev\/peps\/pep-0440\/#version-scheme\n","subject":"Bump version number to 0.2.0","message":"Bump version number to 0.2.0\n","lang":"Python","license":"mit","repos":"JuliaPy\/pyjulia,JuliaLang\/pyjulia,JuliaPy\/pyjulia"} {"commit":"7fc576f3dd4d8d7dbe64dbecfc6dcc9ac9ad6b12","old_file":"conman\/routes\/utils.py","new_file":"conman\/routes\/utils.py","old_contents":"import os\n\n\ndef split_path(path):\n \"\"\"\n Split a url path into its sub-paths.\n\n A url's sub-paths consist of all substrings ending in \/ and starting at\n the start of the url.\n \"\"\"\n paths = ['\/']\n path = path.rstrip('\/')\n\n while path:\n paths.insert(1, path + '\/')\n path = os.path.split(path)[0]\n if path == '\/':\n break\n return paths\n","new_contents":"from collections import deque\n\n\ndef split_path(path):\n \"\"\"\n Split a url path into its sub-paths.\n\n A url's sub-paths consist of all substrings ending in \/ and starting at\n the start of the url.\n\n eg: \/path\/containing\/subpaths\/ becomes:\n\n \/\n \/path\/\n \/path\/containing\/\n \/path\/containing\/subpaths\/\n \"\"\"\n paths = deque()\n path = path or '\/'\n while path:\n path = path.rpartition('\/')[0]\n paths.appendleft(path + '\/')\n return list(paths)\n","subject":"Refactor split_path code for brevity and clarity","message":"Refactor split_path code for brevity and clarity\n","lang":"Python","license":"bsd-2-clause","repos":"meshy\/django-conman,meshy\/django-conman"} {"commit":"15c773250b52a03196a023e286f4f3a2405ba94e","old_file":"backend\/uclapi\/dashboard\/app_helpers.py","new_file":"backend\/uclapi\/dashboard\/app_helpers.py","old_contents":"from binascii import hexlify\nimport os\n\n\ndef generate_api_token():\n key = hexlify(os.urandom(30)).decode()\n dashes_key = \"\"\n for idx, char in enumerate(key):\n if idx % 15 == 0 and idx != len(key)-1:\n dashes_key += \"-\"\n else:\n dashes_key += char\n\n final = \"uclapi\" + dashes_key\n\n return final\n\n\ndef generate_app_id():\n key = hexlify(os.urandom(5)).decode()\n final = \"A\" + key\n\n return final\n","new_contents":"from binascii import hexlify\nfrom random import choice\n\nimport os\nimport string\n\n\ndef generate_api_token():\n key = hexlify(os.urandom(30)).decode()\n dashes_key = \"\"\n for idx, char in enumerate(key):\n if idx % 15 == 0 and idx != len(key)-1:\n dashes_key += \"-\"\n else:\n dashes_key += char\n\n final = \"uclapi\" + dashes_key\n\n return final\n\n\ndef generate_app_id():\n key = hexlify(os.urandom(5)).decode()\n final = \"A\" + key\n\n return final\n\ndef generate_app_client_id():\n client_id = ''.join(random.choice(string.digits, k=16))\n client_id += \".\"\n client_id += ''.join(random.choice(string.digits, k=16))\n\n return client_id\n\ndef generate_app_client_secret():\n client_secret = ''.join(random.choice(string.ascii_lowercase + string.digits, k=64))\n\n return client_secret","subject":"Add helpers to the dashboard code to generate OAuth keys","message":"Add helpers to the dashboard code to generate OAuth keys\n","lang":"Python","license":"mit","repos":"uclapi\/uclapi,uclapi\/uclapi,uclapi\/uclapi,uclapi\/uclapi"} {"commit":"5bb4c61e9950de4c8c000a4ab02b0c901e0b06ff","old_file":"version.py","new_file":"version.py","old_contents":"\"\"\"\nautomatically maintains the latest git tag + revision info in a python file\n\n\"\"\"\n\nimport imp\nimport os\nimport subprocess\n\n\ndef get_project_version(version_file):\n version_file = os.path.join(os.path.dirname(os.path.realpath(__file__)), version_file)\n try:\n module = imp.load_source(\"verfile\", version_file)\n file_ver = module.__version__\n except: # pylint: disable=bare-except\n file_ver = None\n\n try:\n proc = subprocess.Popen([\"git\", \"describe\", \"--always\"],\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n stdout, _ = proc.communicate()\n if stdout:\n git_ver = stdout.splitlines()[0].strip().decode(\"utf-8\")\n if git_ver and ((git_ver != file_ver) or not file_ver):\n open(version_file, \"w\").write(\"__version__ = '%s'\\n\" % git_ver)\n return git_ver\n except OSError:\n pass\n\n if not file_ver:\n raise Exception(\"version not available from git or from file %r\"\n % version_file)\n\n return file_ver\n\n\nif __name__ == \"__main__\":\n import sys\n get_project_version(sys.argv[1])\n","new_contents":"\"\"\"\nautomatically maintains the latest git tag + revision info in a python file\n\n\"\"\"\n\nimport importlib\nimport os\nimport subprocess\n\n\ndef get_project_version(version_file):\n version_file = os.path.join(os.path.dirname(os.path.realpath(__file__)), version_file)\n try:\n module = importlib.load_module(version_file)\n file_ver = module.__version__\n except: # pylint: disable=bare-except\n file_ver = None\n\n try:\n proc = subprocess.Popen([\"git\", \"describe\", \"--always\"],\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n stdout, _ = proc.communicate()\n if stdout:\n git_ver = stdout.splitlines()[0].strip().decode(\"utf-8\")\n if git_ver and ((git_ver != file_ver) or not file_ver):\n open(version_file, \"w\").write(\"__version__ = '%s'\\n\" % git_ver)\n return git_ver\n except OSError:\n pass\n\n if not file_ver:\n raise Exception(\"version not available from git or from file %r\"\n % version_file)\n\n return file_ver\n\n\nif __name__ == \"__main__\":\n import sys\n get_project_version(sys.argv[1])\n","subject":"Migrate from deprecated imp to importlib","message":"Migrate from deprecated imp to importlib\n","lang":"Python","license":"apache-2.0","repos":"aiven\/aiven-client"} {"commit":"e799c94932aa3a9284d99739502645abc47aa8b4","old_file":"docs\/conf.py","new_file":"docs\/conf.py","old_contents":"import os\n\nimport sdv\n\nproject = u'stix-validator'\ncopyright = u'2015, The MITRE Corporation'\nversion = sdv.__version__\nrelease = version\n\nextensions = [\n 'sphinx.ext.autodoc',\n 'sphinxcontrib.napoleon',\n]\n\ntemplates_path = ['_templates']\nsource_suffix = '.rst'\nmaster_doc = 'index'\n\nexclude_patterns = ['_build']\npygments_style = 'sphinx'\n\non_rtd = os.environ.get('READTHEDOCS', None) == 'True'\nif not on_rtd: # only import and set the theme if we're building docs locally\n import sphinx_rtd_theme\n html_theme = 'sphinx_rtd_theme'\n html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]\nelse:\n html_theme = 'default'\n\nhtml_show_sourcelink = False\nhtmlhelp_basename = 'stix-validatordoc'\n\nlatex_elements = {}\nlatex_documents = [\n ('index', 'stix-validator.tex', u'stix-validator Documentation',\n u'The MITRE Corporation', 'manual'),\n]\n","new_contents":"import os\n\nimport sdv\n\nproject = u'stix-validator'\ncopyright = u'2015, The MITRE Corporation'\nversion = sdv.__version__\nrelease = version\n\nextensions = [\n 'sphinx.ext.autodoc',\n 'sphinxcontrib.napoleon',\n]\n\ntemplates_path = ['_templates']\nsource_suffix = '.rst'\nmaster_doc = 'index'\n\nrst_prolog = \"\"\"\n**Version**: {}\n\"\"\".format(release)\n\nexclude_patterns = ['_build']\npygments_style = 'sphinx'\n\non_rtd = os.environ.get('READTHEDOCS', None) == 'True'\nif not on_rtd: # only import and set the theme if we're building docs locally\n import sphinx_rtd_theme\n html_theme = 'sphinx_rtd_theme'\n html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]\nelse:\n html_theme = 'default'\n\nhtml_show_sourcelink = False\nhtmlhelp_basename = 'stix-validatordoc'\n\nlatex_elements = {}\nlatex_documents = [\n ('index', 'stix-validator.tex', u'stix-validator Documentation',\n u'The MITRE Corporation', 'manual'),\n]\n","subject":"Add package version to rst_prolog.","message":"Add package version to rst_prolog.\n","lang":"Python","license":"bsd-3-clause","repos":"STIXProject\/stix-validator,pombredanne\/stix-validator"} {"commit":"80002186ecec3a50d13d04e8467d8d7456d189b5","old_file":"webdriver\/webdriver-w3c-tests\/base_test.py","new_file":"webdriver\/webdriver-w3c-tests\/base_test.py","old_contents":"\nimport ConfigParser\nimport json\nimport os\nimport unittest\n\nfrom webserver import Httpd\nfrom network import get_lan_ip\n\nfrom selenium import webdriver\n\nclass WebDriverBaseTest(unittest.TestCase):\n\n @classmethod\n def setUpClass(cls):\n cls.driver = create_driver()\n\n cls.webserver = Httpd(host=get_lan_ip())\n cls.webserver.start()\n\n @classmethod\n def tearDownClass(cls):\n cls.webserver.stop()\n if cls.driver:\n cls.driver.quit()\n\n\ndef create_driver():\n capabilities = {\n 'xwalkOptions': {\n 'androidPackage': 'org.xwalk.xwalkdrivertest',\n 'androidActivity': '.XwalkDriverTestActivity',\n }\n }\n return webdriver.Remote('http:\/\/localhost:9515', capabilities)\n","new_contents":"\nimport ConfigParser\nimport json\nimport os\nimport unittest\n\nfrom webserver import Httpd\nfrom network import get_lan_ip\n\nfrom selenium import webdriver\n\nclass WebDriverBaseTest(unittest.TestCase):\n\n @classmethod\n def setUpClass(cls):\n cls.driver = create_driver()\n\n cls.webserver = Httpd(host=get_lan_ip())\n cls.webserver.start()\n\n @classmethod\n def tearDownClass(cls):\n cls.webserver.stop()\n if cls.driver:\n cls.driver.quit()\n\n\ndef create_driver():\n capabilities = {\n 'xwalkOptions': {\n 'androidPackage': 'org.xwalk.xwalkdrivertest',\n 'androidActivity': '.XwalkdrivertestActivity',\n }\n }\n return webdriver.Remote('http:\/\/localhost:9515', capabilities)\n","subject":"Update activity name for crosswalk android package rule changed","message":"[xwalkdriver] Update activity name for crosswalk android package rule changed\n","lang":"Python","license":"bsd-3-clause","repos":"YongseopKim\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,jacky-young\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,xiaojunwu\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,Honry\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,jacky-young\/crosswalk-test-suite,xiaojunwu\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,Honry\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,YongseopKim\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,jacky-young\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,Honry\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,jacky-young\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,yugang\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,Honry\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,YongseopKim\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,Honry\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,YongseopKim\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,yugang\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,xiaojunwu\/crosswalk-test-suite,yugang\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,xiaojunwu\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,Honry\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,Honry\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,yugang\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,jacky-young\/crosswalk-test-suite,ibelem\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,yugang\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite,yugang\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,xiaojunwu\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,YongseopKim\/crosswalk-test-suite,haoyunfeix\/crosswalk-test-suite,yugang\/crosswalk-test-suite,YongseopKim\/crosswalk-test-suite,crosswalk-project\/crosswalk-test-suite,xiaojunwu\/crosswalk-test-suite,zhuyongyong\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,jacky-young\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,YongseopKim\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,jiajiax\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,wanghongjuan\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,Honry\/crosswalk-test-suite,Shao-Feng\/crosswalk-test-suite,chunywang\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite,kangxu\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,XiaosongWei\/crosswalk-test-suite,yhe39\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,kaixinjxq\/crosswalk-test-suite,zqzhang\/crosswalk-test-suite,haoxli\/crosswalk-test-suite,BruceDai\/crosswalk-test-suite,JianfengXu\/crosswalk-test-suite,yunxliu\/crosswalk-test-suite,qiuzhong\/crosswalk-test-suite,pk-sam\/crosswalk-test-suite"} {"commit":"d641d7d843899258d88da0d1dffaa762c1378712","old_file":"opps\/fields\/widgets.py","new_file":"opps\/fields\/widgets.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*\nimport json\nfrom django import forms\nfrom django.template.loader import render_to_string\n\nfrom .models import Field, Option, FieldOption\n\n\nclass JSONField(forms.TextInput):\n model = Field\n def render(self, name, value, attrs=None):\n elements = []\n values = json.loads(value)\n objs = self.model.objects.all()\n for obj in objs:\n o = {}\n o['name'] = obj.name\n o['slug'] = obj.slug\n\n element_attr = {}\n element_attr['name'] = obj.name\n element_attr['slug'] = obj.slug\n \"\"\"\n element_attr['value'] = '1'\n element_attr['obj_value'] = values.get(obj.slug, '')\n \"\"\"\n\n if obj.type in [\"checkbox\", \"radiobox\"]:\n obj_value = []\n fo = FieldOption.objects.filter(field=obj)\n for i in fo:\n key = \"{}_{}\".format(obj.slug, i.option.slug)\n obj_value.append(values.get(key, ''))\n element_attr['list'] = zip(fo, obj_value)\n\n o['element'] = render_to_string(\n \"admin\/opps\/fields\/json_{}.html\".format(obj.type),\n dictionary=element_attr\n )\n elements.append(o)\n\n return render_to_string(\n \"admin\/opps\/fields\/json.html\",\n {\"elements\": elements,\n \"name\": name,\n \"value\": value})\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*\nimport json\nfrom django import forms\nfrom django.template.loader import render_to_string\n\nfrom .models import Field, FieldOption\n\n\nclass JSONField(forms.TextInput):\n model = Field\n def render(self, name, value, attrs=None):\n elements = []\n try:\n values = json.loads(value)\n except TypeError:\n values = {}\n objs = self.model.objects.filter(\n application__contains=self.attrs.get('_model', None))\n for obj in objs:\n o = {}\n o['name'] = obj.name\n o['slug'] = obj.slug\n\n element_attr = {}\n element_attr['name'] = obj.name\n element_attr['slug'] = obj.slug\n \"\"\"\n element_attr['value'] = '1'\n element_attr['obj_value'] = values.get(obj.slug, '')\n \"\"\"\n\n if obj.type in [\"checkbox\", \"radiobox\"]:\n obj_value = []\n fo = FieldOption.objects.filter(field=obj)\n for i in fo:\n key = \"{}_{}\".format(obj.slug, i.option.slug)\n obj_value.append(values.get(key, ''))\n element_attr['list'] = zip(fo, obj_value)\n\n o['element'] = render_to_string(\n \"admin\/opps\/fields\/json_{}.html\".format(obj.type),\n dictionary=element_attr\n )\n elements.append(o)\n\n return render_to_string(\n \"admin\/opps\/fields\/json.html\",\n {\"elements\": elements,\n \"name\": name,\n \"value\": value})\n","subject":"Fix bug TypeError, not exist values (json) is dict None","message":"Fix bug TypeError, not exist values (json) is dict None\n","lang":"Python","license":"mit","repos":"williamroot\/opps,jeanmask\/opps,williamroot\/opps,jeanmask\/opps,jeanmask\/opps,opps\/opps,williamroot\/opps,YACOWS\/opps,williamroot\/opps,jeanmask\/opps,YACOWS\/opps,YACOWS\/opps,opps\/opps,opps\/opps,YACOWS\/opps,opps\/opps"} {"commit":"ea57d89c1acc82a473a648f1c53430fadc27f7b2","old_file":"opps\/polls\/__init__.py","new_file":"opps\/polls\/__init__.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nVERSION = (0, 1, 4)\n\n__version__ = \".\".join(map(str, VERSION))\n__status__ = \"Development\"\n__description__ = u\"Poll App for Opps CMS\"\n\n__author__ = u\"Bruno Cezar Rocha\"\n__credits__ = []\n__email__ = u\"rochacbruno@gmail.com\"\n__license__ = u\"MIT License\"\n__copyright__ = u\"Copyright 2013, YACOWS\"\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nVERSION = (0, 1, 4)\n\n__version__ = \".\".join(map(str, VERSION))\n__status__ = \"Development\"\n__description__ = u\"Poll App for Opps CMS\"\n\n__author__ = u\"Bruno Cezar Rocha\"\n__credits__ = []\n__email__ = u\"rochacbruno@gmail.com\"\n__license__ = u\"MIT License\"\n__copyright__ = u\"Copyright 2013, Opps Projects\"\n","subject":"Fix Copyright application, YACOWS to Opps Projects","message":"Fix Copyright application, YACOWS to Opps Projects\n","lang":"Python","license":"mit","repos":"opps\/opps-polls,opps\/opps-polls"} {"commit":"e2282e8832bcf1f2621dc56488fec5eb9f1b1acc","old_file":"src\/sentry\/api\/endpoints\/group_events.py","new_file":"src\/sentry\/api\/endpoints\/group_events.py","old_contents":"from __future__ import absolute_import\n\nfrom sentry.api.base import DocSection\nfrom sentry.api.bases import GroupEndpoint\nfrom sentry.api.serializers import serialize\nfrom sentry.api.paginator import DateTimePaginator\nfrom sentry.models import Event, Group\nfrom sentry.utils.apidocs import scenario, attach_scenarios\n\n\n@scenario('ListAvailableSamples')\ndef list_available_samples_scenario(runner):\n group = Group.objects.filter(project=runner.default_project).first()\n runner.request(\n method='GET',\n path='\/issues\/%s\/events\/' % group.id\n )\n\n\nclass GroupEventsEndpoint(GroupEndpoint):\n doc_section = DocSection.EVENTS\n\n @attach_scenarios([list_available_samples_scenario])\n def get(self, request, group):\n \"\"\"\n List an Issue's Events\n ``````````````````````\n\n This endpoint lists an issue's events.\n\n :pparam string issue_id: the ID of the issue to retrieve.\n :auth: required\n \"\"\"\n\n events = Event.objects.filter(\n group=group\n )\n\n query = request.GET.get('query')\n if query:\n events = events.filter(\n message__iexact=query,\n )\n\n return self.paginate(\n request=request,\n queryset=events,\n order_by='-datetime',\n on_results=lambda x: serialize(x, request.user),\n paginator_cls=DateTimePaginator,\n )\n","new_contents":"from __future__ import absolute_import\n\nfrom sentry.api.base import DocSection\nfrom sentry.api.bases import GroupEndpoint\nfrom sentry.api.serializers import serialize\nfrom sentry.api.paginator import DateTimePaginator\nfrom sentry.models import Event, Group\nfrom sentry.utils.apidocs import scenario, attach_scenarios\n\n\n@scenario('ListAvailableSamples')\ndef list_available_samples_scenario(runner):\n group = Group.objects.filter(project=runner.default_project).first()\n runner.request(\n method='GET',\n path='\/issues\/%s\/events\/' % group.id\n )\n\n\nclass GroupEventsEndpoint(GroupEndpoint):\n doc_section = DocSection.EVENTS\n\n @attach_scenarios([list_available_samples_scenario])\n def get(self, request, group):\n \"\"\"\n List an Issue's Events\n ``````````````````````\n\n This endpoint lists an issue's events.\n\n :pparam string issue_id: the ID of the issue to retrieve.\n :auth: required\n \"\"\"\n\n events = Event.objects.filter(\n group=group\n )\n\n query = request.GET.get('query')\n if query:\n events = events.filter(\n message__icontains=query,\n )\n\n return self.paginate(\n request=request,\n queryset=events,\n order_by='-datetime',\n on_results=lambda x: serialize(x, request.user),\n paginator_cls=DateTimePaginator,\n )\n","subject":"Make event search actually useful","message":"Make event search actually useful\n","lang":"Python","license":"bsd-3-clause","repos":"looker\/sentry,zenefits\/sentry,zenefits\/sentry,zenefits\/sentry,jean\/sentry,mvaled\/sentry,ifduyue\/sentry,beeftornado\/sentry,beeftornado\/sentry,JamesMura\/sentry,fotinakis\/sentry,looker\/sentry,alexm92\/sentry,beeftornado\/sentry,looker\/sentry,alexm92\/sentry,JamesMura\/sentry,JackDanger\/sentry,mvaled\/sentry,jean\/sentry,JackDanger\/sentry,gencer\/sentry,daevaorn\/sentry,jean\/sentry,ifduyue\/sentry,JamesMura\/sentry,nicholasserra\/sentry,zenefits\/sentry,BuildingLink\/sentry,ifduyue\/sentry,nicholasserra\/sentry,mvaled\/sentry,mvaled\/sentry,mitsuhiko\/sentry,looker\/sentry,gencer\/sentry,daevaorn\/sentry,BuildingLink\/sentry,JamesMura\/sentry,gencer\/sentry,ifduyue\/sentry,nicholasserra\/sentry,jean\/sentry,JackDanger\/sentry,gencer\/sentry,mitsuhiko\/sentry,fotinakis\/sentry,gencer\/sentry,alexm92\/sentry,fotinakis\/sentry,zenefits\/sentry,jean\/sentry,fotinakis\/sentry,mvaled\/sentry,looker\/sentry,daevaorn\/sentry,BuildingLink\/sentry,ifduyue\/sentry,mvaled\/sentry,BuildingLink\/sentry,daevaorn\/sentry,JamesMura\/sentry,BuildingLink\/sentry"} {"commit":"ebfeba2704dc73c136fa2ed217ef4337265b92dd","old_file":"addie\/utilities\/__init__.py","new_file":"addie\/utilities\/__init__.py","old_contents":"import os\nfrom qtpy.uic import loadUi\n\ndef load_ui(ui_filename, baseinstance):\n ui_filename = os.path.split(ui_filename)[-1]\n\n # directory containing this file\n filename = __file__\n if not os.path.isdir(filename):\n filename = os.path.split(filename)[0]\n # get the location of the designer directory\n # this function assumes that all ui files are there\n filename = os.path.join(filename, '..', '..', 'designer')\n\n # put together the full path to the ui file\n filename = os.path.join(filename, ui_filename)\n return loadUi(filename, baseinstance=baseinstance)\n","new_contents":"import os\nfrom qtpy.uic import loadUi\n\ndef load_ui(ui_filename, baseinstance):\n cwd = os.getcwd()\n ui_filename = os.path.split(ui_filename)[-1]\n\n # get the location of the designer directory\n # this function assumes that all ui files are there\n filename = os.path.join(cwd, 'designer', ui_filename)\n\n return loadUi(filename, baseinstance=baseinstance)\n","subject":"Fix path for designer directory with ui files","message":"Fix path for designer directory with ui files\n","lang":"Python","license":"mit","repos":"neutrons\/FastGR,neutrons\/FastGR,neutrons\/FastGR"} {"commit":"e2bee28d9b9f5034fe1c92b94e08fd73d2050b30","old_file":"PythonClient\/index.py","new_file":"PythonClient\/index.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom socketIO_client import SocketIO, LoggingNamespace\n\n\ndef on_bbb_response(*args):\n print('on_bbb_response', args)\n\n\ndef on_chat_message_response(*args):\n print(args)\n\nwith SocketIO('http:\/\/192.168.0.110', 8080, LoggingNamespace) as socketIO:\n socketIO.emit('hello', {'xxx': 'yyy'}, on_bbb_response)\n # socketIO.wait_for_callbacks(seconds=1)\n\n socketIO.on('chat message', on_chat_message_response)\n\nprint('hello')","new_contents":"# -*- coding: utf-8 -*-\n\nfrom socketIO_client import SocketIO, LoggingNamespace\n\n\ndef on_bbb_response(*args):\n print('on_bbb_response', args)\n\n\ndef on_chat_message_response(*args):\n print(args)\n\nwith SocketIO('http:\/\/192.168.0.110', 8080, LoggingNamespace) as socketIO:\n socketIO.emit('hello', {'xxx': 'yyy'}, on_bbb_response)\n # socketIO.wait_for_callbacks(seconds=1)\n\n socketIO.on('chat_message', on_chat_message_response)\n\nprint('hello')","subject":"Correct chat message event name","message":"Correct chat message event name\n","lang":"Python","license":"mit","repos":"yakovenkodenis\/websockets_secure_chat,yakovenkodenis\/websockets_secure_chat,yakovenkodenis\/websockets_secure_chat,yakovenkodenis\/websockets_secure_chat"} {"commit":"631665a8aeee54d5094480ddf4140a61dce4a960","old_file":"ostinato\/blog\/apps.py","new_file":"ostinato\/blog\/apps.py","old_contents":"from django.apps import AppConfig\n\n\nclass OstinatoBlogConfig(AppConfig):\n name = 'ostinato.blog'\n label = 'ost_blog'\n verbose_name = 'Ostinato Blog Engine'\n\n","new_contents":"from django.apps import AppConfig\n\n\nclass OstinatoBlogConfig(AppConfig):\n name = 'ostinato.blog'\n label = 'ostinato_blog'\n verbose_name = 'Ostinato Blog Engine'\n\n","subject":"Correct app label of ostinato_blog","message":"Correct app label of ostinato_blog\n","lang":"Python","license":"mit","repos":"andrewebdev\/django-ostinato,andrewebdev\/django-ostinato,andrewebdev\/django-ostinato"} {"commit":"7643635278fc1c92289e8fdd456614ce85a2c2f3","old_file":"addons\/osfstorage\/models.py","new_file":"addons\/osfstorage\/models.py","old_contents":"import logging\n\nfrom addons.base.models import BaseNodeSettings, BaseStorageAddon\n\nlogger = logging.getLogger(__name__)\n\nclass NodeSettings(BaseStorageAddon, BaseNodeSettings):\n pass\n","new_contents":"import logging\n\nfrom addons.base.models import BaseNodeSettings, BaseStorageAddon\n\nlogger = logging.getLogger(__name__)\n\nclass NodeSettings(BaseStorageAddon, BaseNodeSettings):\n # Required overrides\n complete = True\n has_auth = True\n","subject":"Add required overrides to osfstorage.NodeSettings","message":"Add required overrides to osfstorage.NodeSettings\n","lang":"Python","license":"apache-2.0","repos":"felliott\/osf.io,CenterForOpenScience\/osf.io,laurenrevere\/osf.io,cslzchen\/osf.io,monikagrabowska\/osf.io,caneruguz\/osf.io,Nesiehr\/osf.io,alexschiller\/osf.io,aaxelb\/osf.io,caseyrollins\/osf.io,Johnetordoff\/osf.io,caneruguz\/osf.io,mluo613\/osf.io,chrisseto\/osf.io,monikagrabowska\/osf.io,cslzchen\/osf.io,monikagrabowska\/osf.io,monikagrabowska\/osf.io,saradbowman\/osf.io,leb2dg\/osf.io,cslzchen\/osf.io,alexschiller\/osf.io,hmoco\/osf.io,mfraezz\/osf.io,saradbowman\/osf.io,CenterForOpenScience\/osf.io,mluo613\/osf.io,brianjgeiger\/osf.io,acshi\/osf.io,acshi\/osf.io,leb2dg\/osf.io,Nesiehr\/osf.io,erinspace\/osf.io,erinspace\/osf.io,chrisseto\/osf.io,alexschiller\/osf.io,binoculars\/osf.io,icereval\/osf.io,brianjgeiger\/osf.io,alexschiller\/osf.io,Nesiehr\/osf.io,crcresearch\/osf.io,adlius\/osf.io,adlius\/osf.io,chrisseto\/osf.io,sloria\/osf.io,sloria\/osf.io,chennan47\/osf.io,laurenrevere\/osf.io,HalcyonChimera\/osf.io,laurenrevere\/osf.io,chennan47\/osf.io,leb2dg\/osf.io,icereval\/osf.io,felliott\/osf.io,mattclark\/osf.io,HalcyonChimera\/osf.io,acshi\/osf.io,erinspace\/osf.io,baylee-d\/osf.io,TomBaxter\/osf.io,cwisecarver\/osf.io,TomBaxter\/osf.io,mfraezz\/osf.io,leb2dg\/osf.io,baylee-d\/osf.io,acshi\/osf.io,aaxelb\/osf.io,icereval\/osf.io,monikagrabowska\/osf.io,Johnetordoff\/osf.io,caneruguz\/osf.io,pattisdr\/osf.io,hmoco\/osf.io,caseyrollins\/osf.io,chennan47\/osf.io,TomBaxter\/osf.io,caseyrollins\/osf.io,felliott\/osf.io,Johnetordoff\/osf.io,mluo613\/osf.io,binoculars\/osf.io,mluo613\/osf.io,acshi\/osf.io,adlius\/osf.io,HalcyonChimera\/osf.io,baylee-d\/osf.io,Johnetordoff\/osf.io,cslzchen\/osf.io,aaxelb\/osf.io,sloria\/osf.io,cwisecarver\/osf.io,mluo613\/osf.io,adlius\/osf.io,alexschiller\/osf.io,cwisecarver\/osf.io,CenterForOpenScience\/osf.io,cwisecarver\/osf.io,crcresearch\/osf.io,mfraezz\/osf.io,caneruguz\/osf.io,hmoco\/osf.io,brianjgeiger\/osf.io,brianjgeiger\/osf.io,mfraezz\/osf.io,mattclark\/osf.io,HalcyonChimera\/osf.io,CenterForOpenScience\/osf.io,Nesiehr\/osf.io,mattclark\/osf.io,chrisseto\/osf.io,crcresearch\/osf.io,pattisdr\/osf.io,binoculars\/osf.io,pattisdr\/osf.io,felliott\/osf.io,aaxelb\/osf.io,hmoco\/osf.io"} {"commit":"3875b14e6c94c4a6a7ad47a3eb55cae62096d0e4","old_file":"agateremote\/table_remote.py","new_file":"agateremote\/table_remote.py","old_contents":"#!\/usr\/bin\/env python\n\n\"\"\"\nThis module contains the Remote extension to :class:`Table `.\n\"\"\"\n\nimport agate\nimport requests\nimport six\n\ndef from_url(cls, url, callback=agate.Table.from_csv, binary=False, **kwargs):\n \"\"\"\n Download a remote file and pass it to a :class:`.Table` parser.\n\n :param url:\n URL to a file to load.\n :param callback:\n The method to invoke to create the table. Typically either\n :meth:`agate.Table.from_csv` or :meth:`agate.Table.from_json`, but\n it could also be a method provided by an extension.\n :param binary:\n If :code:`False` the downloaded data will be processed as a string,\n otherwise it will be treated as binary data. (e.g. for Excel files)\n \"\"\"\n r = requests.get(url)\n\n if binary:\n content = six.BytesIO(r.content)\n else:\n if six.PY2:\n content = six.StringIO(r.content.decode('utf-8'))\n else:\n content = six.StringIO(r.text)\n\n return callback(content, **kwargs)\n\nagate.Table.from_url = classmethod(from_url)\n","new_contents":"#!\/usr\/bin\/env python\n\n\"\"\"\nThis module contains the Remote extension to :class:`Table `.\n\"\"\"\n\nimport agate\nimport requests\nimport six\n\ndef from_url(cls, url, callback=agate.Table.from_csv, requests_encoding=None, binary=False, **kwargs):\n \"\"\"\n Download a remote file and pass it to a :class:`.Table` parser.\n\n :param url:\n URL to a file to load.\n :param callback:\n The method to invoke to create the table. Typically either\n :meth:`agate.Table.from_csv` or :meth:`agate.Table.from_json`, but\n it could also be a method provided by an extension.\n :param requests_encoding:\n An encoding to pass to requests for use when decoding the response\n content. (e.g. force use of 'utf-8-sig' when CSV has a BOM).\n :param binary:\n If :code:`False` the downloaded data will be processed as a string,\n otherwise it will be treated as binary data. (e.g. for Excel files)\n \"\"\"\n r = requests.get(url)\n\n if requests_encoding:\n r.encoding = requests_encoding\n\n if binary:\n content = six.BytesIO(r.content)\n else:\n if six.PY2:\n content = six.StringIO(r.content.decode('utf-8'))\n else:\n content = six.StringIO(r.text)\n\n return callback(content, **kwargs)\n\nagate.Table.from_url = classmethod(from_url)\n","subject":"Add 'requests_encoding' parameter Allows user to override Requests' 'educated guess' about encoding of a response. Useful when loading a remote CSV that has a BOM that has been served with a 'text\/csv' content-type, which Requests guesses needs a 'ISO-8859-1' encoding.","message":"Add 'requests_encoding' parameter\nAllows user to override Requests' 'educated guess' about encoding of a response. Useful when loading a remote CSV that has a BOM that has been served with a 'text\/csv' content-type, which Requests guesses needs a 'ISO-8859-1' encoding.\n","lang":"Python","license":"mit","repos":"wireservice\/agate-remote"} {"commit":"9262dad14237d57a3817a199f9a8b04371de9607","old_file":"mis_bot\/scraper\/database.py","new_file":"mis_bot\/scraper\/database.py","old_contents":"from sqlalchemy import create_engine\nfrom sqlalchemy.orm import scoped_session, sessionmaker\nfrom sqlalchemy.ext.declarative import declarative_base\n\n# Database\nengine = create_engine('sqlite:\/\/\/files\/chats.db', convert_unicode=True)\ndb_session = scoped_session(sessionmaker(autocommit=False,\n autoflush=False,\n bind=engine))\nBase = declarative_base()\nBase.query = db_session.query_property()\n\ndef init_db():\n # import all modules here that might define models so that\n # they will be registered properly on the metadata. Otherwise\n # you will have to import them first before calling init_db()\n import scraper.models\n Base.metadata.create_all(bind=engine)","new_contents":"from sqlalchemy import create_engine\nfrom sqlalchemy.orm import scoped_session, sessionmaker\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.pool import StaticPool\n\n# Database\nengine = create_engine('sqlite:\/\/\/files\/chats.db', convert_unicode=True, \n connect_args= {'check_same_thread': False},\n poolclass=StaticPool)\ndb_session = scoped_session(sessionmaker(autocommit=False,\n autoflush=False,\n bind=engine))\nBase = declarative_base()\nBase.query = db_session.query_property()\n\ndef init_db():\n # import all modules here that might define models so that\n # they will be registered properly on the metadata. Otherwise\n # you will have to import them first before calling init_db()\n import scraper.models\n Base.metadata.create_all(bind=engine)","subject":"Allow sharing db connection across threads","message":"Allow sharing db connection across threads\n\n","lang":"Python","license":"mit","repos":"ArionMiles\/MIS-Bot"} {"commit":"a04116d32931c5e85de417b5da048c91d495261b","old_file":"pyeventstore\/client.py","new_file":"pyeventstore\/client.py","old_contents":"import asyncio\nimport uuid\nimport json\n\nimport requests\nfrom requests.exceptions import HTTPError\n\nfrom pyeventstore.events import (get_all_events,\n start_subscription,\n publish_events)\nfrom pyeventstore.stream_page import StreamPage\n\n\nclass Client:\n\n def __init__(self, host, secure=False, port=2113):\n proto = \"https\" if secure else \"http\"\n self.uri_base = '{}:\/\/{}:{}'.format(proto, host, port)\n\n @asyncio.coroutine\n def publish_events(self, stream_name, events):\n uri = self.stream_head_uri(stream_name)\n yield from publish_events(uri, events)\n\n def stream_head_uri(self, stream_name):\n return '{}\/streams\/{}'.format(self.uri_base, stream_name)\n\n @asyncio.coroutine\n def get_all_events(self, stream_name):\n head_uri = self.stream_head_uri(stream_name)\n return (yield from get_all_events(head_uri))\n\n @asyncio.coroutine\n def subscribe(self, stream_name, interval_seconds=1):\n head_uri = self.stream_head_uri(stream_name)\n return (yield from start_subscription(head_uri, interval_seconds))\n\n def get_projection(self, projection_name):\n uri = self.uri_base + '\/projection\/{}'.format(projection_name)\n headers = {'Accept': 'application\/json'}\n response = requests.get(uri, headers=headers)\n return response.json()\n\n def get_projection_state(self, projection_name, partition=None):\n uri = self.uri_base + '\/projection\/{}\/state'.format(projection_name)\n headers = {'Accept': 'application\/json'}\n params = {}\n if partition:\n params['partition'] = partition\n response = requests.get(uri, headers=headers, params=params)\n return response.json()\n","new_contents":"import asyncio\nimport uuid\nimport json\n\nfrom pyeventstore.events import (get_all_events,\n start_subscription,\n publish_events)\n\nclass Client:\n\n def __init__(self, host, secure=False, port=2113):\n proto = \"https\" if secure else \"http\"\n self.uri_base = '{}:\/\/{}:{}'.format(proto, host, port)\n\n @asyncio.coroutine\n def publish_events(self, stream_name, events):\n uri = self.stream_head_uri(stream_name)\n yield from publish_events(uri, events)\n\n def stream_head_uri(self, stream_name):\n return '{}\/streams\/{}'.format(self.uri_base, stream_name)\n\n @asyncio.coroutine\n def get_all_events(self, stream_name):\n head_uri = self.stream_head_uri(stream_name)\n return (yield from get_all_events(head_uri))\n\n @asyncio.coroutine\n def subscribe(self, stream_name, interval_seconds=1):\n head_uri = self.stream_head_uri(stream_name)\n return (yield from start_subscription(head_uri, interval_seconds))\n","subject":"Remove projections methods for now","message":"Remove projections methods for now\n","lang":"Python","license":"mit","repos":"cjlarose\/pyeventstore"} {"commit":"5a3ffb93131c83f81eb123c2969714dcc80513ca","old_file":"django\/crashreport\/processor\/signals.py","new_file":"django\/crashreport\/processor\/signals.py","old_contents":"# -*- Mode: python; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*- *\/\n#\n# This Source Code Form is subject to the terms of the Mozilla Public\n# License, v. 2.0. If a copy of the MPL was not distributed with this\n# file, You can obtain one at http:\/\/mozilla.org\/MPL\/2.0\/.\n#\n\nfrom .processor import MinidumpProcessor\n\nfrom uwsgidecoratorsfallback import spool\n\nimport logging\n\nlogger = logging.getLogger(__name__)\n\n@spool\ndef do_process_uploaded_crash(*args, **kwargs):\n minproc = MinidumpProcessor()\n minproc.process(kwargs['crash_id'])\n logger.info('processed: %s' % (kwargs['crash_id']))\n\n# vim:set shiftwidth=4 softtabstop=4 expandtab: *\/\n","new_contents":"# -*- Mode: python; tab-width: 4; indent-tabs-mode: nil; c-basic-offset: 4 -*- *\/\n#\n# This Source Code Form is subject to the terms of the Mozilla Public\n# License, v. 2.0. If a copy of the MPL was not distributed with this\n# file, You can obtain one at http:\/\/mozilla.org\/MPL\/2.0\/.\n#\n\nfrom .processor import MinidumpProcessor\n\nfrom uwsgidecoratorsfallback import spool\n\nimport logging\n\nlogger = logging.getLogger(__name__)\n\n@spool\ndef do_process_uploaded_crash(env):\n minproc = MinidumpProcessor()\n minproc.process(env['crash_id'])\n logger.info('processed: %s' % (env['crash_id']))\n\n# vim:set shiftwidth=4 softtabstop=4 expandtab: *\/\n","subject":"Revert \"make the uwsgi spooler code also work with the fallback\"","message":"Revert \"make the uwsgi spooler code also work with the fallback\"\n\nThis reverts commit 84ab847b444fbd41b9cc17e5c79a609efdcdf6cf.\n","lang":"Python","license":"mpl-2.0","repos":"mmohrhard\/crash,mmohrhard\/crash,Liongold\/crash,Liongold\/crash,mmohrhard\/crash,Liongold\/crash"} {"commit":"98dafbc7578209b9768e6ca6ccfa7854f70deb16","old_file":"runTwircBot.py","new_file":"runTwircBot.py","old_contents":"from TwircBot import TwircBot as tw\n\nbot = tw(\"config\/sampleConfig.sample\")\nbot.print_config()\nbot.connect()\n","new_contents":"from TwircBot import TwircBot as tw\nimport sys\n\nbot = tw(sys.argv[1])\nbot.print_config()\nbot.connect()\n","subject":"Modify runTwirc.py to accept system arguments","message":"Modify runTwirc.py to accept system arguments\n","lang":"Python","license":"mit","repos":"johnmarcampbell\/twircBot"} {"commit":"f868b126b3bd81ec900f378ff1fa8bd29ab8ea4c","old_file":"transformations\/Transformations.py","new_file":"transformations\/Transformations.py","old_contents":"from transformations.BackTranslation import BackTranslation\nfrom transformations.ButterFingersPerturbation import ButterFingersPerturbation\nfrom transformations.ChangeNamedEntities import ChangeNamedEntities\nfrom transformations.SentenceTransformation import SentenceTransformation\nfrom transformations.WithoutPunctuation import WithoutPunctuation\n\n\nclass TransformationsList(SentenceTransformation):\n\n def __init__(self):\n transformations = []\n transformations.append(ButterFingersPerturbation())\n transformations.append(WithoutPunctuation())\n transformations.append(ChangeNamedEntities())\n transformations.append(BackTranslation())\n self.transformations = transformations\n\n def generate(self, sentence: str):\n print(f\"Original Input : {sentence}\")\n generations = {\"Original\": sentence}\n for transformation in self.transformations:\n generations[transformation.name()] = transformation.generate(sentence)\n return generations\n","new_contents":"\nfrom transformations.BackTranslation import BackTranslation\nfrom transformations.ButterFingersPerturbation import ButterFingersPerturbation\nfrom transformations.ChangeNamedEntities import ChangeNamedEntities\nfrom transformations.SentenceTransformation import SentenceTransformation\nfrom transformations.WithoutPunctuation import WithoutPunctuation\n\n\nclass TransformationsList(SentenceTransformation):\n\n def __init__(self):\n transformations = [ButterFingersPerturbation(), WithoutPunctuation(), ChangeNamedEntities(), BackTranslation()]\n self.transformations = transformations\n\n def generate(self, sentence: str):\n print(f\"Original Input : {sentence}\")\n generations = {\"Original\": sentence}\n for transformation in self.transformations:\n generations[transformation.name()] = transformation.generate(sentence)\n return generations\n","subject":"Add interface for source+label pertubation","message":"Add interface for source+label pertubation\n","lang":"Python","license":"mit","repos":"GEM-benchmark\/NL-Augmenter"} {"commit":"44226dabb65cb06522c128539660e407e53ca602","old_file":"parseHTML.py","new_file":"parseHTML.py","old_contents":"from bs4 import BeautifulSoup\n\n#uses beautiful soup to parse html file\n#finds the correct span tag\n#Gets the percentage of ink left in the printer\nsoup = BeautifulSoup(open(\"test\/test.html\"))\nres = soup.find('span',{'class':'hpConsumableBlockHeaderText'}).text\nnum = res[24] + res[25] \n\n","new_contents":"from bs4 import BeautifulSoup\n\n#uses beautiful soup to parse html file\n#finds the correct span tag\n#Gets the percentage of ink left in the printer\nsoup = BeautifulSoup(open(\"test\/test.html\"))\nres = soup.find('span',{'class':'hpConsumableBlockHeaderText'}).text\nnum = res[24] + res[25]\n\nfile = open('test\/data.csv', 'w+') \nfile.write(num + \",\")\n\n","subject":"Write information from test.html to csv file","message":"Write information from test.html to csv file\n","lang":"Python","license":"mit","repos":"tfahl\/printfo,tfahl\/printfo"} {"commit":"f820ef6cef8037942d18dcc912fb6de093ecc8de","old_file":"txircd\/modules\/rfc\/cmd_userhost.py","new_file":"txircd\/modules\/rfc\/cmd_userhost.py","old_contents":"from twisted.plugin import IPlugin\nfrom twisted.words.protocols import irc\nfrom txircd.module_interface import Command, ICommand, IModuleData, ModuleData\nfrom zope.interface import implements\n\nclass UserhostCommand(ModuleData, Command):\n\timplements(IPlugin, IModuleData, ICommand)\n\t\n\tname = \"UserhostCommand\"\n\tcore = True\n\t\n\tdef userCommands(self):\n\t\treturn [ (\"USERHOST\", 1, self) ]\n\t\n\tdef parseParams(self, user, params, prefix, tags):\n\t\tif not params:\n\t\t\tuser.sendSingleError(\"UserhostParams\", irc.ERR_NEEDMOREPARAMS, \"USERHOST\", \"Not enough parameters\")\n\t\t\treturn None\n\t\treturn {\n\t\t\t\"nicks\": params[:5]\n\t\t}\n\t\n\tdef execute(self, user, data):\n\t\tuserHosts = []\n\t\tfor nick in data[\"nicks\"]:\n\t\t\tif nick not in self.ircd.userNicks:\n\t\t\t\tcontinue\n\t\t\ttargetUser = self.ircd.users[self.ircd.userNicks[nick]]\n\t\t\toutput = targetUser.nick\n\t\t\tif self.ircd.runActionUntilValue(\"userhasoperpermission\", targetUser, \"\", users=[targetUser]):\n\t\t\t\toutput += \"*\"\n\t\t\toutput += \"=\"\n\t\t\tif user.metadataKeyExists(\"away\"):\n\t\t\t\toutput += \"-\"\n\t\t\telse:\n\t\t\t\toutput += \"+\"\n\t\t\toutput += \"{}@{}\".format(targetUser.ident, targetUser.host())\n\t\t\tuserHosts.append(output)\n\t\tuser.sendMessage(irc.RPL_USERHOST, \" \".join(userHosts))\n\t\treturn True\n\nuserhostCmd = UserhostCommand()","new_contents":"from twisted.plugin import IPlugin\nfrom twisted.words.protocols import irc\nfrom txircd.module_interface import Command, ICommand, IModuleData, ModuleData\nfrom zope.interface import implements\n\nclass UserhostCommand(ModuleData, Command):\n\timplements(IPlugin, IModuleData, ICommand)\n\t\n\tname = \"UserhostCommand\"\n\tcore = True\n\t\n\tdef userCommands(self):\n\t\treturn [ (\"USERHOST\", 1, self) ]\n\t\n\tdef parseParams(self, user, params, prefix, tags):\n\t\tif not params:\n\t\t\tuser.sendSingleError(\"UserhostParams\", irc.ERR_NEEDMOREPARAMS, \"USERHOST\", \"Not enough parameters\")\n\t\t\treturn None\n\t\treturn {\n\t\t\t\"nicks\": params[:5]\n\t\t}\n\t\n\tdef execute(self, user, data):\n\t\tuserHosts = []\n\t\tfor nick in data[\"nicks\"]:\n\t\t\tif nick not in self.ircd.userNicks:\n\t\t\t\tcontinue\n\t\t\ttargetUser = self.ircd.users[self.ircd.userNicks[nick]]\n\t\t\toutput = targetUser.nick\n\t\t\tif self.ircd.runActionUntilValue(\"userhasoperpermission\", targetUser, \"\", users=[targetUser]):\n\t\t\t\toutput += \"*\"\n\t\t\toutput += \"=\"\n\t\t\tif targetUser.metadataKeyExists(\"away\"):\n\t\t\t\toutput += \"-\"\n\t\t\telse:\n\t\t\t\toutput += \"+\"\n\t\t\toutput += \"{}@{}\".format(targetUser.ident, targetUser.host())\n\t\t\tuserHosts.append(output)\n\t\tuser.sendMessage(irc.RPL_USERHOST, \" \".join(userHosts))\n\t\treturn True\n\nuserhostCmd = UserhostCommand()","subject":"Check away status of the target, not user, of USERHOST","message":"Check away status of the target, not user, of USERHOST\n","lang":"Python","license":"bsd-3-clause","repos":"ElementalAlchemist\/txircd,Heufneutje\/txircd"} {"commit":"36fe90ca170525cabc4f2a496a12a28c86b7e82d","old_file":"uchicagohvz\/production_settings.py","new_file":"uchicagohvz\/production_settings.py","old_contents":"from local_settings import *\n\nALLOWED_HOSTS = ['uchicagohvz.org']\n\n# Database\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.postgresql_psycopg2', # Add 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'.\n 'NAME': 'uchicagohvz', # Or path to database file if using sqlite3.\n 'USER': 'user', # Not used with sqlite3.\n 'PASSWORD': '', # Not used with sqlite3.\n 'HOST': '', # Set to empty string for localhost. Not used with sqlite3.\n 'PORT': '', # Set to empty string for default. Not used with sqlite3.\n }\n}\n\n# REST framework settings\nREST_FRAMEWORK = {\n\t'DEFAULT_RENDERER_CLASSES': (\n\t\t'rest_framework.renderers.JSONRenderer',\n\t)\n}\n\n# Mandrill email settings\nEMAIL_HOST = 'smtp.mandrillapp.com'\nfrom secrets import EMAIL_HOST_USER, EMAIL_HOST_PASSWORD\nEMAIL_PORT = '587'\nEMAIL_USE_TLS = True","new_contents":"from local_settings import *\n\nsettings.DEBUG = False\n\nALLOWED_HOSTS = ['uchicagohvz.org']\n\n# Database\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.postgresql_psycopg2', # Add 'postgresql_psycopg2', 'postgresql', 'mysql', 'sqlite3' or 'oracle'.\n 'NAME': 'uchicagohvz', # Or path to database file if using sqlite3.\n 'USER': 'user', # Not used with sqlite3.\n 'PASSWORD': '', # Not used with sqlite3.\n 'HOST': '', # Set to empty string for localhost. Not used with sqlite3.\n 'PORT': '', # Set to empty string for default. Not used with sqlite3.\n }\n}\n\n# REST framework settings\nREST_FRAMEWORK = {\n\t'DEFAULT_RENDERER_CLASSES': (\n\t\t'rest_framework.renderers.JSONRenderer',\n\t)\n}\n\n# Mandrill email settings\nEMAIL_HOST = 'smtp.mandrillapp.com'\nfrom secrets import EMAIL_HOST_USER, EMAIL_HOST_PASSWORD\nEMAIL_PORT = '587'\nEMAIL_USE_TLS = True","subject":"Set DEBUG = False in production","message":"Set DEBUG = False in production\n","lang":"Python","license":"mit","repos":"kz26\/uchicago-hvz,kz26\/uchicago-hvz,kz26\/uchicago-hvz"} {"commit":"2b603ebe92e308aa78928772e8681f3cc46775cb","old_file":"numba\/cloudpickle\/compat.py","new_file":"numba\/cloudpickle\/compat.py","old_contents":"import sys\n\n\nif sys.version_info < (3, 8):\n try:\n import pickle5 as pickle # noqa: F401\n from pickle5 import Pickler # noqa: F401\n except ImportError:\n import pickle # noqa: F401\n from pickle import _Pickler as Pickler # noqa: F401\nelse:\n import pickle # noqa: F401\n from _pickle import Pickler # noqa: F401\n","new_contents":"import sys\n\n\nif sys.version_info < (3, 8):\n # NOTE: pickle5 is disabled due to problems in testing.\n # try:\n # import pickle5 as pickle # noqa: F401\n # from pickle5 import Pickler # noqa: F401\n # except ImportError:\n import pickle # noqa: F401\n from pickle import _Pickler as Pickler # noqa: F401\nelse:\n import pickle # noqa: F401\n from _pickle import Pickler # noqa: F401\n","subject":"Disable pickle5 use in cloudpickle","message":"Disable pickle5 use in cloudpickle\n","lang":"Python","license":"bsd-2-clause","repos":"numba\/numba,IntelLabs\/numba,stuartarchibald\/numba,stonebig\/numba,seibert\/numba,cpcloud\/numba,stonebig\/numba,cpcloud\/numba,numba\/numba,IntelLabs\/numba,stuartarchibald\/numba,seibert\/numba,cpcloud\/numba,numba\/numba,IntelLabs\/numba,cpcloud\/numba,seibert\/numba,cpcloud\/numba,stonebig\/numba,stonebig\/numba,stuartarchibald\/numba,seibert\/numba,stuartarchibald\/numba,stuartarchibald\/numba,stonebig\/numba,seibert\/numba,IntelLabs\/numba,numba\/numba,numba\/numba,IntelLabs\/numba"} {"commit":"7e166ba983bd1470f6eac6776107a64539c38581","old_file":"animal_spharm\/test\/test_animal_spharm.py","new_file":"animal_spharm\/test\/test_animal_spharm.py","old_contents":"","new_contents":"import numpy as np\nimport pytest\nimport xray\n\nfrom aospy_user import SpharmInterface\n\n\n@pytest.fixture\ndef compute_vrtdiv(u, v):\n sphint = SpharmInterface(u, v)\n sphint.make_vectorwind()\n sphint.make_spharmt()\n\n vort, divg = sphint.vectorwind.vrtdiv()\n return sphint.to_xray(vort), sphint.to_xray(divg)\n\n\ndef test_vrtdiv():\n path = ('\/archive\/Spencer.Hill\/am2\/am2clim_reyoi\/gfdl.ncrc2-default-prod\/'\n 'pp\/atmos_level\/ts\/monthly\/1yr\/atmos_level.198301-198312.')\n\n # Vertically defined, sigma levels.\n u_arr = xray.open_dataset(path + 'ucomp.nc').ucomp\n v_arr = xray.open_dataset(path + 'vcomp.nc').vcomp\n vort, divg = compute_vrtdiv(u_arr, v_arr)\n assert vort.shape == u_arr.shape\n assert divg.shape == u_arr.shape\n np.testing.assert_array_equal(u_arr.lat, vort.lat)\n np.testing.assert_array_equal(u_arr.lon, vort.lon)\n np.testing.assert_array_equal(u_arr.time, vort.time)\n np.testing.assert_array_equal(u_arr.pfull, vort.pfull)\n\n # Not vertically defined.\n u0 = u_arr[:,0]\n v0 = v_arr[:,0]\n vort0, divg0 = compute_vrtdiv(u0, v0)\n assert vort0.shape == u0.shape\n assert divg0.shape == u0.shape\n\n # Dummy case: zeros everywhere\n u_arr_zeros = xray.DataArray(np.zeros_like(u_arr.values), dims=u_arr.dims,\n coords=u_arr.coords)\n v_arr_zeros = u_arr_zeros.copy()\n vort_zeros, divg_zeros = compute_vrtdiv(u_arr_zeros, v_arr_zeros)\n assert not vort_zeros.any()\n assert not divg_zeros.any()\n","subject":"Copy over tests from aospy-obj-lib","message":"Copy over tests from aospy-obj-lib\n","lang":"Python","license":"apache-2.0","repos":"spencerahill\/animal-spharm"} {"commit":"bbd3b1939712d9784fe61884d9b06faa95c36006","old_file":"tests\/test_project\/test_app\/models.py","new_file":"tests\/test_project\/test_app\/models.py","old_contents":"from django.db import models\n\n\nclass TestModel(models.Model):\n\n name = models.CharField(max_length=63, unique=True, verbose_name='Name')\n image = models.ImageField(verbose_name='Image')\n","new_contents":"from django.db import models\n\n\nclass TestModel(models.Model):\n\n name = models.CharField(max_length=63, unique=True, verbose_name='Name')\n image = models.ImageField(verbose_name='Image', upload_to='uploads\/')\n","subject":"Test compatibility with older Django versions.","message":"Test compatibility with older Django versions.\n","lang":"Python","license":"mit","repos":"dessibelle\/sorl-thumbnail-serializer-field"} {"commit":"59ce3ca9c1572dcf71aa5de5cdb354def594a36c","old_file":"downloads\/urls.py","new_file":"downloads\/urls.py","old_contents":"from django.conf.urls import patterns, url\nfrom functools import partial\n\nfrom problems.models import UserSolution\n\nfrom .views import download_protected_file\n\nurlpatterns = patterns('',\n url(r'solutions\/(?P.*)$', partial(download_protected_file,\n path_prefix='solutions\/',\n model_class=UserSolution),\n name='download_solution'),\n)\n","new_contents":"from django.conf.urls import patterns, url\nfrom functools import partial\n\nfrom problems.models import UserSolution\n\nfrom .views import download_protected_file\n\nurlpatterns = patterns('',\n url(r'solutions\/(?P.*)$', download_protected_file,\n dict(path_prefix='solutions\/', model_class=UserSolution),\n name='download_solution'),\n)\n","subject":"Remove unnecessary usage of functools.partial","message":"downloads: Remove unnecessary usage of functools.partial\n","lang":"Python","license":"mit","repos":"matus-stehlik\/roots,rtrembecky\/roots,matus-stehlik\/roots,tbabej\/roots,rtrembecky\/roots,tbabej\/roots,tbabej\/roots,rtrembecky\/roots,matus-stehlik\/roots"} {"commit":"8cd2c9a4b9cf69a97e758d1e6416d2efbbf7028a","old_file":"misc\/python\/botan\/__init__.py","new_file":"misc\/python\/botan\/__init__.py","old_contents":"from _botan import *\n\ninit = LibraryInitializer()\n\ndef Filter(name, key = None, iv = None, dir = None):\n if key != None and iv != None and dir != None:\n return make_filter(name, key, iv, dir)\n elif key != None and dir != None:\n return make_filter(name, key, dir)\n elif key != None:\n return make_filter(name, key)\n else:\n return make_filter(name)\n\ndef Pipe(*filters):\n pipe = PipeObj();\n for filter in filters:\n if filter:\n pipe.append(filter)\n return pipe\n\n#def Filter(name, key):\n# return make_filter(name, key)\n","new_contents":"from _botan import *\n\ninit = LibraryInitializer()\n\ndef Filter(name, key = None, iv = None, dir = None):\n if key != None and iv != None and dir != None:\n return make_filter(name, key, iv, dir)\n elif key != None and dir != None:\n return make_filter(name, key, dir)\n elif key != None:\n return make_filter(name, key)\n else:\n return make_filter(name)\n\ndef Pipe(*filters):\n pipe = PipeObj()\n for filter in filters:\n if filter:\n pipe.append(filter)\n return pipe\n\n#def Filter(name, key):\n# return make_filter(name, key)\n","subject":"Remove a semicolon that didn't need to be there","message":"Remove a semicolon that didn't need to be there\n","lang":"Python","license":"bsd-2-clause","repos":"randombit\/botan,randombit\/botan,Rohde-Schwarz-Cybersecurity\/botan,randombit\/botan,Rohde-Schwarz-Cybersecurity\/botan,Rohde-Schwarz-Cybersecurity\/botan,Rohde-Schwarz-Cybersecurity\/botan,webmaster128\/botan,Rohde-Schwarz-Cybersecurity\/botan,webmaster128\/botan,randombit\/botan,webmaster128\/botan,webmaster128\/botan,webmaster128\/botan,Rohde-Schwarz-Cybersecurity\/botan,randombit\/botan"} {"commit":"6cb9008ee2ed49d9630735378bd84727aef3caef","old_file":"dipy\/core\/tests\/test_qball.py","new_file":"dipy\/core\/tests\/test_qball.py","old_contents":"\"\"\" Testing qball \n\n\"\"\"\n\nimport numpy as np\n\nimport dipy.core.qball as qball\n\nfrom nose.tools import assert_true, assert_false, \\\n assert_equal, assert_raises\n\nfrom numpy.testing import assert_array_equal, assert_array_almost_equal\n\nfrom dipy.testing import parametric\n\n\n@parametric\ndef test_real_sph_harm():\n real_sh = qball.real_sph_harm(0, 0, 0, 0)\n yield assert_true(True)\n yield assert_false(True)\n\n\n","new_contents":"\"\"\" Testing qball \n\n\"\"\"\n\nimport numpy as np\n\nimport dipy.core.qball as qball\n\nfrom nose.tools import assert_true, assert_false, \\\n assert_equal, assert_raises\n\nfrom numpy.testing import assert_array_equal, assert_array_almost_equal\n\nfrom dipy.testing import parametric\n\n@parametric\ndef test_sph_harm_ind_list():\n m_list, n_list = qball.sph_harm_ind_list(8)\n yield assert_equal(m_list.shape, n_list.shape)\n yield assert_equal(m_list.ndim, 2)\n yield assert_equal(m_list.shape, (45,1))\n yield assert_true(np.all(np.abs(m_list) <= n_list))\n yield assert_array_equal(n_list % 2, 0)\n yield assert_raises(ValueError, qball.sph_harm_ind_list, 1)\n\n@parametric\ndef test_real_sph_harm():\n real_sh = qball.real_sph_harm(0, 0, 0, 0)\n yield assert_true(True)\n\n\n","subject":"TEST - some real_sph_harm tests","message":"TEST - some real_sph_harm tests\n","lang":"Python","license":"bsd-3-clause","repos":"villalonreina\/dipy,samuelstjean\/dipy,jyeatman\/dipy,sinkpoint\/dipy,mdesco\/dipy,Messaoud-Boudjada\/dipy,maurozucchelli\/dipy,villalonreina\/dipy,nilgoyyou\/dipy,beni55\/dipy,demianw\/dipy,FrancoisRheaultUS\/dipy,rfdougherty\/dipy,JohnGriffiths\/dipy,mdesco\/dipy,Messaoud-Boudjada\/dipy,JohnGriffiths\/dipy,samuelstjean\/dipy,samuelstjean\/dipy,StongeEtienne\/dipy,oesteban\/dipy,maurozucchelli\/dipy,matthieudumont\/dipy,matthieudumont\/dipy,nilgoyyou\/dipy,oesteban\/dipy,rfdougherty\/dipy,beni55\/dipy,FrancoisRheaultUS\/dipy,jyeatman\/dipy,maurozucchelli\/dipy,sinkpoint\/dipy,StongeEtienne\/dipy,demianw\/dipy"} {"commit":"46ab8d71824f80ba5d02349a9f89328e5c47f434","old_file":"app\/views.py","new_file":"app\/views.py","old_contents":"from app import app, \\\n cors_header\nfrom flask import request, \\\n make_response, \\\n send_from_directory\nimport json\nimport os\n\n\n@app.route('\/', methods=['GET'])\n@cors_header\ndef index():\n if 'X-Forwarded-For' in request.headers:\n ipAddress = request.headers['X-Forwarded-For']\n else:\n ipAddress = request.remote_addr\n\n response = json.dumps({\n \"address\": ipAddress\n })\n\n return app.response_class(response=response, mimetype='application\/json')\n","new_contents":"from app import app, \\\n cors_header\nfrom flask import request, \\\n make_response, \\\n send_from_directory\nimport json\nimport re\nimport os\n\n\n@app.route('\/', methods=['GET'])\n@cors_header\ndef index():\n if 'X-Forwarded-For' in request.headers:\n ipAddress = request.headers['X-Forwarded-For']\n else:\n ipAddress = request.remote_addr\n\n response = json.dumps({\n \"address\": re.sub(r\",.+$\", \"\", ipAddress)\n })\n\n return app.response_class(response=response, mimetype='application\/json')\n","subject":"Handle instance where we get multiple IP addresess in the response","message":"Handle instance where we get multiple IP addresess in the response\n","lang":"Python","license":"mit","repos":"taeram\/gipsy"} {"commit":"3439eb09916212cd71650aecc49ae1c22f650274","old_file":"apps\/package\/templatetags\/package_tags.py","new_file":"apps\/package\/templatetags\/package_tags.py","old_contents":"from datetime import timedelta\nfrom datetime import datetime\n\nfrom django import template\n\nfrom github2.client import Github\n\nfrom package.models import Package, Commit\n\nregister = template.Library()\n\ngithub = Github()\n\n@register.filter\ndef commits_over_52(package):\n\n current = datetime.now()\n weeks = []\n commits = Commit.objects.filter(package=package).values_list('commit_date', flat=True)\n for week in range(52):\n weeks.append(len([x for x in commits if x < current and x > (current - timedelta(7))]))\n current -= timedelta(7) \n\n weeks.reverse()\n weeks = map(str, weeks)\n return ','.join(weeks)\n","new_contents":"from datetime import datetime, timedelta\n\nfrom django import template\n\nfrom package.models import Commit\n\nregister = template.Library()\n\n@register.filter\ndef commits_over_52(package):\n\n current = datetime.now()\n weeks = []\n commits = Commit.objects.filter(package=package).values_list('commit_date', flat=True)\n for week in range(52):\n weeks.append(len([x for x in commits if x < current and x > (current - timedelta(7))]))\n current -= timedelta(7) \n\n weeks.reverse()\n weeks = map(str, weeks)\n return ','.join(weeks)\n","subject":"Clean up some imports in the package app's template_tags.py file.","message":"Clean up some imports in the package app's template_tags.py file.\n","lang":"Python","license":"mit","repos":"nanuxbe\/djangopackages,QLGu\/djangopackages,cartwheelweb\/packaginator,nanuxbe\/djangopackages,nanuxbe\/djangopackages,miketheman\/opencomparison,QLGu\/djangopackages,benracine\/opencomparison,cartwheelweb\/packaginator,audreyr\/opencomparison,pydanny\/djangopackages,miketheman\/opencomparison,QLGu\/djangopackages,benracine\/opencomparison,cartwheelweb\/packaginator,pydanny\/djangopackages,pydanny\/djangopackages,audreyr\/opencomparison"} {"commit":"b034eeda25fcf55e7da018f3c91a23a5e252ae2f","old_file":"bm\/app\/models.py","new_file":"bm\/app\/models.py","old_contents":"from django.db import models\nfrom django.conf import settings\n\nclass Category(models.Model):\n\tuser = models.ForeignKey(settings.AUTH_USER_MODEL)\n\tname = models.CharField(max_length=21)\n\trow_number = models.IntegerField(default=0)\n\tcolumn_number = models.IntegerField(default=0)\n\tprogress_bar_color = models.CharField(max_length=6, default=\"335544\")\n\t# hidden = models.BooleanField(default=False)\n\t# trash = models.BooleanField(default=False)\n\n\tdef __str__(self):\n\t\treturn str(self.user) + ' ' + str(self.name)\n\nclass Bookmark(models.Model):\n\tcategory = models.ForeignKey(Category)\n\tname = models.CharField(max_length=50)\n\tlink = models.TextField()\n\trow_number = models.IntegerField(default=0)\n\tglyphicon = models.CharField(max_length=30, default=\"asterisk\")\n\n\tdef __str__(self):\n\t\treturn str(self.category) + ' ' + str(self.name)\n\nclass Trash(models.Model):\n\tcategory = models.ForeignKey(Category)\n\tname = models.CharField(max_length=50)\n\tlink = models.TextField()\n\tglyphicon = models.CharField(max_length=30)\n\n\tdef __str__(self):\n\t\treturn str(self.category) + ' ' + str(self.name)","new_contents":"from django.db import models\nfrom django.conf import settings\n\nclass Category(models.Model):\n\tuser = models.ForeignKey(settings.AUTH_USER_MODEL)\n\tname = models.CharField(max_length=21)\n\trow_number = models.IntegerField(default=0)\n\tcolumn_number = models.IntegerField(default=0)\n\tprogress_bar_color = models.CharField(max_length=6, default=\"335544\")\n\t# hidden = models.BooleanField(default=False)\n\t# trash = models.BooleanField(default=False)\n\n\tdef __str__(self):\n\t\treturn str(self.name)\n\nclass Bookmark(models.Model):\n\tcategory = models.ForeignKey(Category)\n\tname = models.CharField(max_length=50)\n\tlink = models.TextField()\n\trow_number = models.IntegerField(default=0)\n\tglyphicon = models.CharField(max_length=30, default=\"asterisk\")\n\n\tdef __str__(self):\n\t\treturn str(self.category) + ' ' + str(self.name)\n\nclass Trash(models.Model):\n\tcategory = models.ForeignKey(Category)\n\tname = models.CharField(max_length=50)\n\tlink = models.TextField()\n\tglyphicon = models.CharField(max_length=30)\n\n\tdef __str__(self):\n\t\treturn str(self.category) + ' ' + str(self.name)","subject":"Change str() of Category for easier form handling","message":"Change str() of Category for easier form handling\n","lang":"Python","license":"mit","repos":"GSC-RNSIT\/bookmark-manager,rohithpr\/bookmark-manager,rohithpr\/bookmark-manager,GSC-RNSIT\/bookmark-manager"} {"commit":"f0b705f1fde5ac33a58ccd56ed67bfcbecf0daf3","old_file":"djstripe\/management\/commands\/djstripe_sync_customers.py","new_file":"djstripe\/management\/commands\/djstripe_sync_customers.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.core.management.base import BaseCommand\n\nfrom ...settings import get_user_model\nfrom ...sync import sync_customer\n\nUser = get_user_model()\n\n\nclass Command(BaseCommand):\n\n help = \"Sync customer data with stripe\"\n\n def handle(self, *args, **options):\n qs = User.objects.exclude(customer__isnull=True)\n count = 0\n total = qs.count()\n for user in qs:\n count += 1\n perc = int(round(100 * (float(count) \/ float(total))))\n print(\n \"[{0}\/{1} {2}%] Syncing {3} [{4}]\".format(\n count, total, perc, user.username, user.pk\n )\n )\n sync_customer(user)\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.core.management.base import BaseCommand\n\nfrom ...settings import get_user_model\nfrom ...sync import sync_customer\n\nUser = get_user_model()\n\n\nclass Command(BaseCommand):\n\n help = \"Sync customer data with stripe\"\n\n def handle(self, *args, **options):\n qs = User.objects.exclude(customer__isnull=True)\n count = 0\n total = qs.count()\n for user in qs:\n count += 1\n perc = int(round(100 * (float(count) \/ float(total))))\n print(\n \"[{0}\/{1} {2}%] Syncing {3} [{4}]\".format(\n count, total, perc, user.get_username(), user.pk\n )\n )\n sync_customer(user)\n","subject":"Use user.get_username() instead of user.username","message":"Use user.get_username() instead of user.username\n\nTo support custom User models\n","lang":"Python","license":"bsd-3-clause","repos":"andrewyoung1991\/dj-stripe,areski\/dj-stripe,benmurden\/dj-stripe,koobs\/dj-stripe,areski\/dj-stripe,tkwon\/dj-stripe,kavdev\/dj-stripe,aliev\/dj-stripe,ctrengove\/dj-stripe,cjrh\/dj-stripe,StErMi\/dj-stripe,mthornhill\/dj-stripe,ctrengove\/dj-stripe,cjrh\/dj-stripe,iddqd1\/dj-stripe,dj-stripe\/dj-stripe,maxmalynowsky\/django-stripe-rest,photocrowd\/dj-stripe,davidgillies\/dj-stripe,doctorwidget\/dj-stripe,davidgillies\/dj-stripe,andrewyoung1991\/dj-stripe,mwarkentin\/dj-stripe,jpadilla\/dj-stripe,aliev\/dj-stripe,benmurden\/dj-stripe,jameshiew\/dj-stripe,mwarkentin\/dj-stripe,mthornhill\/dj-stripe,LaunchlabAU\/dj-stripe,tkwon\/dj-stripe,jameshiew\/dj-stripe,kavdev\/dj-stripe,jleclanche\/dj-stripe,doctorwidget\/dj-stripe,iddqd1\/dj-stripe,photocrowd\/dj-stripe,LaunchlabAU\/dj-stripe,koobs\/dj-stripe,jleclanche\/dj-stripe,jpadilla\/dj-stripe,StErMi\/dj-stripe,pydanny\/dj-stripe,dj-stripe\/dj-stripe,pydanny\/dj-stripe"} {"commit":"bf17a86bccf25ead90d11dd15a900cb784d9cb9f","old_file":"raco\/myrial\/myrial_test.py","new_file":"raco\/myrial\/myrial_test.py","old_contents":"\nimport collections\nimport math\nimport unittest\n\nimport raco.fakedb\nimport raco.myrial.interpreter as interpreter\nimport raco.myrial.parser as parser\nfrom raco.myrialang import compile_to_json\n\nclass MyrialTestCase(unittest.TestCase):\n\n def setUp(self):\n self.db = raco.fakedb.FakeDatabase()\n self.parser = parser.Parser()\n self.processor = interpreter.StatementProcessor(self.db)\n\n def execute_query(self, query, test_logical=False):\n '''Run a test query against the fake database'''\n statements = self.parser.parse(query)\n self.processor.evaluate(statements)\n\n if test_logical:\n plan = self.processor.get_logical_plan()\n else:\n plan = self.processor.get_physical_plan()\n json = compile_to_json(query, '', [('A', plan)])\n\n self.db.evaluate(plan)\n\n\n return self.db.get_temp_table('__OUTPUT0__')\n\n def run_test(self, query, expected, test_logical=False):\n '''Execute a test query with an expected output'''\n actual = self.execute_query(query, test_logical)\n self.assertEquals(actual, expected)\n\n","new_contents":"\nimport collections\nimport math\nimport unittest\n\nimport raco.fakedb\nimport raco.myrial.interpreter as interpreter\nimport raco.myrial.parser as parser\n\nclass MyrialTestCase(unittest.TestCase):\n\n def setUp(self):\n self.db = raco.fakedb.FakeDatabase()\n self.parser = parser.Parser()\n self.processor = interpreter.StatementProcessor(self.db)\n\n def execute_query(self, query, test_logical=False):\n '''Run a test query against the fake database'''\n statements = self.parser.parse(query)\n self.processor.evaluate(statements)\n\n if test_logical:\n plan = self.processor.get_logical_plan()\n else:\n plan = self.processor.get_physical_plan()\n self.db.evaluate(plan)\n\n return self.db.get_temp_table('__OUTPUT0__')\n\n def run_test(self, query, expected, test_logical=False):\n '''Execute a test query with an expected output'''\n actual = self.execute_query(query, test_logical)\n self.assertEquals(actual, expected)\n\n","subject":"Revert \"Add compile_to_json invocation in Myrial test fixture\"","message":"Revert \"Add compile_to_json invocation in Myrial test fixture\"\n\nThis reverts commit ceb848021d5323b5bad8518ac7ed850a51fc89ca.\n","lang":"Python","license":"bsd-3-clause","repos":"uwescience\/raco,uwescience\/raco,uwescience\/raco,uwescience\/raco,uwescience\/raco"} {"commit":"2bdc5c33b1e9eb394eb62533f4ae4df081ea1452","old_file":"numpy\/setup.py","new_file":"numpy\/setup.py","old_contents":"#!\/usr\/bin\/env python3\n\ndef configuration(parent_package='',top_path=None):\n from numpy.distutils.misc_util import Configuration\n config = Configuration('numpy', parent_package, top_path)\n\n config.add_subpackage('compat')\n config.add_subpackage('core')\n config.add_subpackage('distutils')\n config.add_subpackage('doc')\n config.add_subpackage('f2py')\n config.add_subpackage('fft')\n config.add_subpackage('lib')\n config.add_subpackage('linalg')\n config.add_subpackage('ma')\n config.add_subpackage('matrixlib')\n config.add_subpackage('polynomial')\n config.add_subpackage('random')\n config.add_subpackage('testing')\n config.add_subpackage('typing')\n config.add_data_dir('doc')\n config.add_data_files('py.typed')\n config.add_data_files('*.pyi')\n config.add_subpackage('tests')\n config.make_config_py() # installs __config__.py\n return config\n\nif __name__ == '__main__':\n print('This is the wrong setup.py file to run')\n","new_contents":"#!\/usr\/bin\/env python3\n\ndef configuration(parent_package='',top_path=None):\n from numpy.distutils.misc_util import Configuration\n config = Configuration('numpy', parent_package, top_path)\n\n config.add_subpackage('_array_api')\n config.add_subpackage('compat')\n config.add_subpackage('core')\n config.add_subpackage('distutils')\n config.add_subpackage('doc')\n config.add_subpackage('f2py')\n config.add_subpackage('fft')\n config.add_subpackage('lib')\n config.add_subpackage('linalg')\n config.add_subpackage('ma')\n config.add_subpackage('matrixlib')\n config.add_subpackage('polynomial')\n config.add_subpackage('random')\n config.add_subpackage('testing')\n config.add_subpackage('typing')\n config.add_data_dir('doc')\n config.add_data_files('py.typed')\n config.add_data_files('*.pyi')\n config.add_subpackage('tests')\n config.make_config_py() # installs __config__.py\n return config\n\nif __name__ == '__main__':\n print('This is the wrong setup.py file to run')\n","subject":"Make the _array_api submodule install correctly","message":"Make the _array_api submodule install correctly\n","lang":"Python","license":"bsd-3-clause","repos":"seberg\/numpy,mhvk\/numpy,mattip\/numpy,endolith\/numpy,seberg\/numpy,mattip\/numpy,numpy\/numpy,simongibbons\/numpy,simongibbons\/numpy,charris\/numpy,endolith\/numpy,charris\/numpy,rgommers\/numpy,rgommers\/numpy,simongibbons\/numpy,anntzer\/numpy,pdebuyl\/numpy,jakirkham\/numpy,endolith\/numpy,mattip\/numpy,anntzer\/numpy,jakirkham\/numpy,seberg\/numpy,seberg\/numpy,jakirkham\/numpy,charris\/numpy,mhvk\/numpy,pdebuyl\/numpy,simongibbons\/numpy,numpy\/numpy,numpy\/numpy,mattip\/numpy,mhvk\/numpy,numpy\/numpy,anntzer\/numpy,endolith\/numpy,rgommers\/numpy,rgommers\/numpy,simongibbons\/numpy,anntzer\/numpy,pdebuyl\/numpy,pdebuyl\/numpy,charris\/numpy,jakirkham\/numpy,jakirkham\/numpy,mhvk\/numpy,mhvk\/numpy"} {"commit":"29cc95bbdb12e50d09e8079bfae5841a7e734743","old_file":"plinth\/modules\/help\/urls.py","new_file":"plinth\/modules\/help\/urls.py","old_contents":"#\n# This file is part of Plinth.\n#\n# This program is free software: you can redistribute it and\/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n\n\"\"\"\nURLs for the Help module\n\"\"\"\n\nfrom django.conf.urls import url\n\nfrom . import help as views\n\n\nurlpatterns = [\n # having two urls for one page is a hack to help the current url\/menu\n # system highlight the correct menu item. Every submenu-item with the same\n # url prefix as the main-menu is highlighted automatically.\n url(r'^help\/$', views.index, name='index'),\n url(r'^help\/index\/$', views.index, name='index_explicit'),\n url(r'^help\/about\/$', views.about, name='about'),\n url(r'^help\/manual\/$', views.manual, name='manual'),\n url(r'^help\/status-log\/$', views.status_log, name='status-log'),\n]\n","new_contents":"#\n# This file is part of Plinth.\n#\n# This program is free software: you can redistribute it and\/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n\n\"\"\"\nURLs for the Help module\n\"\"\"\n\nfrom django.conf.urls import url\n\nfrom plinth.utils import non_admin_view\n\nfrom . import help as views\n\n\nurlpatterns = [\n # having two urls for one page is a hack to help the current url\/menu\n # system highlight the correct menu item. Every submenu-item with the same\n # url prefix as the main-menu is highlighted automatically.\n url(r'^help\/$', non_admin_view(views.index), name='index'),\n url(r'^help\/index\/$', non_admin_view(views.index), name='index_explicit'),\n url(r'^help\/about\/$', non_admin_view(views.about), name='about'),\n url(r'^help\/manual\/$', non_admin_view(views.manual), name='manual'),\n url(r'^help\/status-log\/$', non_admin_view(views.status_log), name='status-log'),\n]\n","subject":"Make help accessible for logged-in non-admin users","message":"Make help accessible for logged-in non-admin users\n\nSigned-off-by: Hemanth Kumar Veeranki \n\nReviewed-by: Johannes Keyser <187051b70230423a457adbc3e507f9e4fff08d4b@posteo.de>\n","lang":"Python","license":"agpl-3.0","repos":"vignanl\/Plinth,vignanl\/Plinth,kkampardi\/Plinth,vignanl\/Plinth,harry-7\/Plinth,harry-7\/Plinth,kkampardi\/Plinth,kkampardi\/Plinth,vignanl\/Plinth,kkampardi\/Plinth,harry-7\/Plinth,kkampardi\/Plinth,vignanl\/Plinth,harry-7\/Plinth,harry-7\/Plinth"} {"commit":"5edddcc85b0e21bb576b71db63d082c8ace5cf70","old_file":"examples\/boilerplates\/samples\/google_test.py","new_file":"examples\/boilerplates\/samples\/google_test.py","old_contents":"'''\nGoogle.com testing example\n'''\n\nfrom seleniumbase import BaseCase\nfrom google_objects import HomePage, ResultsPage\n\n\nclass GoogleTests(BaseCase):\n\n def test_google_dot_com(self):\n self.open('http:\/\/www.google.com')\n self.assert_element(HomePage.search_button)\n self.assert_element(HomePage.feeling_lucky_button)\n self.update_text(HomePage.search_box, \"github\\n\")\n self.assert_text(\"github.com\", ResultsPage.search_results)\n self.assert_element(ResultsPage.google_logo)\n self.click_link_text(\"Images\")\n self.assert_element('img[alt=\"Image result for github\"]')\n","new_contents":"'''\nGoogle.com testing example\n'''\n\nfrom seleniumbase import BaseCase\nfrom google_objects import HomePage, ResultsPage\n\n\nclass GoogleTests(BaseCase):\n\n def test_google_dot_com(self):\n self.open('http:\/\/www.google.com')\n self.assert_element(HomePage.search_button)\n self.assert_element(HomePage.feeling_lucky_button)\n self.update_text(HomePage.search_box, \"github\\n\")\n self.assert_text(\"github.com\", ResultsPage.search_results)\n self.click_link_text(\"Images\")\n self.assert_element('img[alt=\"Image result for github\"]')\n","subject":"Update Google boilerplate test. (Logo frequently changes)","message":"Update Google boilerplate test. (Logo frequently changes)\n","lang":"Python","license":"mit","repos":"seleniumbase\/SeleniumBase,mdmintz\/seleniumspot,mdmintz\/SeleniumBase,mdmintz\/seleniumspot,mdmintz\/SeleniumBase,mdmintz\/SeleniumBase,mdmintz\/SeleniumBase,seleniumbase\/SeleniumBase,seleniumbase\/SeleniumBase,seleniumbase\/SeleniumBase"} {"commit":"1f5c592595a18a521f08213c5bcb1e6b604e0eec","old_file":"coil\/__init__.py","new_file":"coil\/__init__.py","old_contents":"# Copyright (c) 2005-2006 Itamar Shtull-Trauring.\n# Copyright (c) 2008-2009 ITA Software, Inc.\n# See LICENSE.txt for details.\n\n\"\"\"Coil: A Configuration Library.\"\"\"\n\n__version_info__ = (0,3,16)\n__version__ = \".\".join([str(x) for x in __version_info__])\n__all__ = ['struct', 'parser', 'tokenizer', 'errors']\n\nfrom coil.parser import Parser\n\ndef parse_file(file_name, **kwargs):\n \"\"\"Open and parse a coil file.\n\n See :class:`Parser ` for possible keyword arguments.\n\n :param file_name: Name of file to parse.\n :type file_name: str\n\n :return: The root object.\n :rtype: :class:`Struct `\n \"\"\"\n coil = open(file_name)\n return Parser(coil, file_name, **kwargs).root()\n\ndef parse(string, **kwargs):\n \"\"\"Parse a coil string.\n\n See :class:`Parser ` for possible keyword arguments.\n\n :param file_name: String containing data to parse.\n :type file_name: str\n\n :return: The root object.\n :rtype: :class:`Struct `\n \"\"\"\n return Parser(string.splitlines(), **kwargs).root()\n","new_contents":"# Copyright (c) 2005-2006 Itamar Shtull-Trauring.\n# Copyright (c) 2008-2009 ITA Software, Inc.\n# See LICENSE.txt for details.\n\n\"\"\"Coil: A Configuration Library.\"\"\"\n\n__version_info__ = (0,3,99)\n__version__ = \".\".join([str(x) for x in __version_info__])\n__all__ = ['struct', 'parser', 'tokenizer', 'errors']\n\nfrom coil.parser import Parser\n\ndef parse_file(file_name, **kwargs):\n \"\"\"Open and parse a coil file.\n\n See :class:`Parser ` for possible keyword arguments.\n\n :param file_name: Name of file to parse.\n :type file_name: str\n\n :return: The root object.\n :rtype: :class:`Struct `\n \"\"\"\n coil = open(file_name)\n return Parser(coil, file_name, **kwargs).root()\n\ndef parse(string, **kwargs):\n \"\"\"Parse a coil string.\n\n See :class:`Parser ` for possible keyword arguments.\n\n :param file_name: String containing data to parse.\n :type file_name: str\n\n :return: The root object.\n :rtype: :class:`Struct `\n \"\"\"\n return Parser(string.splitlines(), **kwargs).root()\n","subject":"Bump coil version to 0.3.99 to denote that this is 0.4-dev.","message":"Bump coil version to 0.3.99 to denote that this is 0.4-dev.","lang":"Python","license":"mit","repos":"marineam\/coil,kovacsbalu\/coil,tectronics\/coil,marineam\/coil,kovacsbalu\/coil,tectronics\/coil"} {"commit":"4b065401e3679d479163bd825c0a3526f0ba1bdf","old_file":"conf-template.py","new_file":"conf-template.py","old_contents":"\nflickrKey = \"\"\nflickrSecret = \"\"\n\nrecaptchaEnabled = False\nrecaptchaKey = \"\"\nrecaptchaSecret = \"\"\n\n","new_contents":"\nflickrKey = \"\"\nflickrSecret = \"\"\n\nrecaptchaEnabled = False\nrecaptchaKey = \"\"\nrecaptchaSecret = \"\"\n\ndebug = False\n\n","subject":"Add default for debug in conf template","message":"Add default for debug in conf template\n","lang":"Python","license":"bsd-3-clause","repos":"TimSC\/auxgis"} {"commit":"2c87cc9a864cbcd0ff0668a348fda1ba82b8a74c","old_file":"gensysinfo.py","new_file":"gensysinfo.py","old_contents":"#!\/usr\/bin\/env python\nimport psutil\nimport os\nimport time\n\ndef create_bar(filled):\n low = '.'\n high = '|'\n if filled > 1:\n low = str(int(filled))\n high = str(int(filled + 1))\n filled = filled - int(filled)\n filled = int(filled * 10)\n if filled < 5:\n color = \"green\"\n elif filled < 8:\n color = \"yellow\"\n else:\n color = \"red\"\n bar = '#[fg=' + color + ']['\n bar += high * filled\n bar += low * (10 - filled)\n bar += ']#[fg=default]'\n return bar\n\n\nwhile True:\n meminfo = psutil.virtual_memory()\n numcpus = psutil.cpu_count()\n\n with open(os.path.expanduser(\"~\/.memblock\"), \"w\") as memblock:\n memblock.write(create_bar((meminfo.total - meminfo.available) \/ meminfo.total))\n with open(os.path.expanduser(\"~\/.cpuutilblock\"), \"w\") as cpuutilblock:\n cpuutilblock.write(create_bar(psutil.cpu_percent()))\n time.sleep(20)\n","new_contents":"#!\/usr\/bin\/env python3\nimport psutil\nimport os\nimport time\n\ndef create_bar(filled):\n low = '.'\n high = '|'\n if filled > 1:\n low = str(int(filled))\n high = str(int(filled + 1))\n filled = filled - int(filled)\n filled = int(filled * 10)\n if filled < 5:\n color = \"green\"\n elif filled < 8:\n color = \"yellow\"\n else:\n color = \"red\"\n bar = '#[fg=' + color + ']['\n bar += high * filled\n bar += low * (10 - filled)\n bar += ']#[fg=default]'\n return bar\n\n\nwhile True:\n meminfo = psutil.virtual_memory()\n numcpus = psutil.cpu_count()\n\n with open(os.path.expanduser(\"~\/.memblock\"), \"w\") as memblock:\n memblock.write(create_bar((meminfo.total - meminfo.available) \/ meminfo.total))\n with open(os.path.expanduser(\"~\/.cpuutilblock\"), \"w\") as cpuutilblock:\n cpuutilblock.write(create_bar(psutil.cpu_percent() \/ 100))\n time.sleep(20)\n","subject":"Convert percentage to a value between 0 and 1","message":"Convert percentage to a value between 0 and 1\n","lang":"Python","license":"mit","repos":"wilfriedvanasten\/miscvar,wilfriedvanasten\/miscvar,wilfriedvanasten\/miscvar"} {"commit":"50fc79e1231eeb59f66a79c29abe2b843e6fd893","old_file":"atlas\/api\/__init__.py","new_file":"atlas\/api\/__init__.py","old_contents":"#!\/usr\/bin\/env python\n\nimport logging\n\nfrom flask import Blueprint, current_app, request\n\napi_v1_blueprint = Blueprint(\"api_v1\", __name__, url_prefix='\/api\/v1')\n\nlog = logging.getLogger('api')\n\n\n@api_v1_blueprint.after_request\ndef log_response(response):\n \"\"\"Log any requests\/responses with an error code\"\"\"\n if current_app.debug: # pragma: no cover, debugging only\n log.debug('%7s: %s - %i', request.method, request.url,\n response.status_code)\n if response.status_code >= 400:\n log.debug('Response data: \\n%s', response.data)\n log.debug('Request data: \\n%s', request.data)\n\n return response\n\n\n# Import the resources to add the routes to the blueprint before the app is\n# initialized\nfrom . import webhook\n","new_contents":"#!\/usr\/bin\/env python\n\nimport logging\n\nfrom flask import Blueprint, current_app, request\n\napi_v1_blueprint = Blueprint(\"api_v1\", __name__, url_prefix='\/api\/v1')\n\nlog = logging.getLogger('api')\n\n\n@api_v1_blueprint.after_request\ndef log_response(response):\n \"\"\"Log any requests\/responses with an error code\"\"\"\n if log.getEffectiveLevel() == logging.DEBUG: # pragma: no cover, debugging only\n log.debug('%7s: %s - %i', request.method, request.url,\n response.status_code)\n if response.status_code >= 400:\n log.debug('Response data: \\n%s', response.data)\n log.debug('Request data: \\n%s', request.data)\n\n return response\n\n\n# Import the resources to add the routes to the blueprint before the app is\n# initialized\nfrom . import webhook\n","subject":"Use LOG_LEVEL to enable\/disable request logging","message":"Use LOG_LEVEL to enable\/disable request logging\n","lang":"Python","license":"mit","repos":"joshfriend\/atlas,joshfriend\/atlas"} {"commit":"92c58061d4460addbe75081090371814e1a5f580","old_file":"working_waterfronts\/working_waterfronts_api\/views\/pointsofinterest.py","new_file":"working_waterfronts\/working_waterfronts_api\/views\/pointsofinterest.py","old_contents":"from django.http import (HttpResponse,\n HttpResponseNotFound)\nfrom django.contrib.gis.measure import D\nfrom working_waterfronts.working_waterfronts_api.models import PointOfInterest\nfrom working_waterfronts.working_waterfronts_api.functions import get_lat_long_prox\n\nimport json\nfrom .serializer import ObjectSerializer\n\n\ndef poi_list(request):\n \"\"\"\n *\/pointsofinterest\/*\n\n List all pointsofinterest in the database. There is no order to this list,\n only whatever is returned by the database.\n \"\"\"\n error = {\n 'status': False,\n 'name': None,\n 'text': None,\n 'level': None,\n 'debug': None\n }\n data = {}\n\n point, proximity, limit, error = get_lat_long_prox(request, error)\n\n if point:\n poi_list = PointOfInterest.objects.filter(\n location__distance_lte=(point, D(mi=proximity)))[:limit]\n else:\n poi_list = PointOfInterest.objects.all()[:limit]\n\n if not poi_list:\n error = {\n \"status\": True,\n \"name\": \"No PointsOfInterest\",\n \"text\": \"No PointsOfInterest found\",\n \"level\": \"Information\",\n \"debug\": \"\"\n }\n\n serializer = ObjectSerializer()\n\n data = {\n \"pointsofinterest\": json.loads(\n serializer.serialize(\n poi_list,\n use_natural_foreign_keys=True\n )\n ),\n \"error\": error\n }\n\n return HttpResponse(json.dumps(data), content_type=\"application\/json\")\n","new_contents":"from django.http import (HttpResponse,\n HttpResponseNotFound)\nfrom django.contrib.gis.measure import D\nfrom working_waterfronts.working_waterfronts_api.models import PointOfInterest\nfrom working_waterfronts.working_waterfronts_api.functions import get_lat_long_prox\n\nimport json\nfrom .serializer import ObjectSerializer\n\n\ndef poi_list(request):\n \"\"\"\n *\/pois\/*\n\n List all pointsofinterest in the database. There is no order to this list,\n only whatever is returned by the database.\n \"\"\"\n error = {\n 'status': False,\n 'name': None,\n 'text': None,\n 'level': None,\n 'debug': None\n }\n data = {}\n\n point, proximity, limit, error = get_lat_long_prox(request, error)\n\n if point:\n poi_list = PointOfInterest.objects.filter(\n location__distance_lte=(point, D(mi=proximity)))[:limit]\n else:\n poi_list = PointOfInterest.objects.all()[:limit]\n\n if not poi_list:\n error = {\n \"status\": True,\n \"name\": \"No PointsOfInterest\",\n \"text\": \"No PointsOfInterest found\",\n \"level\": \"Information\",\n \"debug\": \"\"\n }\n\n serializer = ObjectSerializer()\n\n data = {\n \"pointsofinterest\": json.loads(\n serializer.serialize(\n poi_list,\n use_natural_foreign_keys=True\n )\n ),\n \"error\": error\n }\n\n return HttpResponse(json.dumps(data), content_type=\"application\/json\")\n","subject":"Change docstrong to reflect actual endpoint","message":"Change docstrong to reflect actual endpoint\n","lang":"Python","license":"apache-2.0","repos":"osu-cass\/working-waterfronts-api,osu-cass\/working-waterfronts-api,osu-cass\/working-waterfronts-api"} {"commit":"052e5b1a2fb5f926028bc6c1dfac657ef58a31b7","old_file":"get_input_data.py","new_file":"get_input_data.py","old_contents":"#!\/usr\/bin\/env python\n\nimport sys\nimport os, errno\nimport shlex\nimport shutil\nimport subprocess as sp\nimport argparse\n\ndata_filename = 'input_486a8dd0.tar.gz'\n\ndata_path = '\/short\/public\/access-om2\/' + data_filename\ndata_url = 'http:\/\/s3-ap-southeast-2.amazonaws.com\/dp-drop\/access-om2\/' + data_filename\n\ndef main():\n\n my_dir = os.path.dirname(os.path.realpath(__file__))\n tarball = os.path.join(my_dir, data_filename)\n\n if os.path.exists('input'):\n return 0\n\n # Download input data.\n if not os.path.exists(data_filename):\n if not os.path.exists(data_path):\n sp.check_call(['wget', '-P', my_dir, data_url])\n else:\n shutil.copy(data_path, my_dir)\n\n sp.check_call(['tar', 'zxvf', tarball, '-C', my_dir])\n\n return 0\n\nif __name__ == '__main__':\n sys.exit(main())\n","new_contents":"#!\/usr\/bin\/env python\n\nimport sys\nimport os, errno\nimport shlex\nimport shutil\nimport subprocess as sp\nimport argparse\n\ndata_filename = 'input_f7a40d5b.tar.gz'\n\ndata_path = '\/short\/public\/access-om2\/' + data_filename\ndata_url = 'http:\/\/s3-ap-southeast-2.amazonaws.com\/dp-drop\/access-om2\/' + data_filename\n\ndef main():\n\n my_dir = os.path.dirname(os.path.realpath(__file__))\n tarball = os.path.join(my_dir, data_filename)\n\n if os.path.exists('input'):\n return 0\n\n # Download input data.\n if not os.path.exists(data_filename):\n if not os.path.exists(data_path):\n sp.check_call(['wget', '-P', my_dir, data_url])\n else:\n shutil.copy(data_path, my_dir)\n\n sp.check_call(['tar', 'zxvf', tarball, '-C', my_dir])\n\n return 0\n\nif __name__ == '__main__':\n sys.exit(main())\n","subject":"Update data sources for https:\/\/github.com\/OceansAus\/access-om2\/issues\/37","message":"Update data sources for https:\/\/github.com\/OceansAus\/access-om2\/issues\/37\n","lang":"Python","license":"apache-2.0","repos":"CWSL\/access-om"} {"commit":"7b88a2e65b0010ceef49fdbce61949ee10420cd8","old_file":"desertbot\/modules\/utils\/CommandHandler.py","new_file":"desertbot\/modules\/utils\/CommandHandler.py","old_contents":"\"\"\"\nCreated on Feb 28, 2018\n\n@author: StarlitGhost\n\"\"\"\n\nfrom twisted.plugin import IPlugin\nfrom desertbot.moduleinterface import IModule, BotModule\nfrom zope.interface import implementer\n\n\n@implementer(IPlugin, IModule)\nclass CommandHandler(BotModule):\n def actions(self):\n return super(CommandHandler, self).actions() + [('message-channel', 1, self.handleCommand),\n ('message-user', 1, self.handleCommand)]\n\n def handleCommand(self, message):\n if message.command:\n return self.bot.moduleHandler.runGatheringAction('botmessage', message)\n\n\ncommandhandler = CommandHandler()\n","new_contents":"\"\"\"\nCreated on Feb 28, 2018\n\n@author: StarlitGhost\n\"\"\"\n\nfrom twisted.plugin import IPlugin\nfrom desertbot.moduleinterface import IModule, BotModule\nfrom zope.interface import implementer\n\n\n@implementer(IPlugin, IModule)\nclass CommandHandler(BotModule):\n def __init__(self):\n BotModule.__init__(self)\n self.loadingPriority = 10\n\n def actions(self):\n return super(CommandHandler, self).actions() + [('message-channel', 1, self.handleCommand),\n ('message-user', 1, self.handleCommand)]\n\n def handleCommand(self, message):\n if message.command:\n return self.bot.moduleHandler.runGatheringAction('botmessage', message)\n\n\ncommandhandler = CommandHandler()\n","subject":"Load the command handler before the commands","message":"Load the command handler before the commands\n","lang":"Python","license":"mit","repos":"DesertBot\/DesertBot"} {"commit":"31921ce5ca7ccbaa2db8b8fa11b2b9a6caa14aeb","old_file":"daisyproducer\/settings.py","new_file":"daisyproducer\/settings.py","old_contents":"from settings_common import *\n\nPACKAGE_VERSION = \"0.5\"\n\nDEBUG = TEMPLATE_DEBUG = True\n\nDAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', '..', 'tmp', 'pipeline')\nEXTERNAL_PATH = os.path.join(PROJECT_DIR, '..', '..', '..', 'tmp')\n\nSERVE_STATIC_FILES = True\n\n# the following is an idea from https:\/\/code.djangoproject.com\/wiki\/SplitSettings\n# We have both local settings and common settings. They are used as follows:\n# - common settings are shared data between normal settings and unit test settings\n# - local settings are used on productive servers to keep the local\n# settings such as db passwords, etc out of version control\ntry:\n from settings_local import *\nexcept ImportError:\n pass\n","new_contents":"from settings_common import *\n\nPACKAGE_VERSION = \"0.5\"\n\nDEBUG = TEMPLATE_DEBUG = True\n\nDAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', '..', 'tmp', 'pipeline')\nEXTERNAL_PATH = os.path.join(PROJECT_DIR, '..', '..', '..', 'tmp')\n\nSERVE_STATIC_FILES = True\n\n# the following is an idea from https:\/\/code.djangoproject.com\/wiki\/SplitSettings\n# We have both local settings and common settings. They are used as follows:\n# - common settings are shared data between normal settings and unit test settings\n# - local settings are used on productive servers to keep the local\n# settings such as db passwords, etc out of version control\ntry:\n from settings_local import *\nexcept ImportError:\n pass\n\nif SERVE_STATIC_FILES:\n INSTALLED_APPS += ('django.contrib.staticfiles',)\n","subject":"Use django.contrib.staticfiles when running locally","message":"Use django.contrib.staticfiles when running locally\n\nso that the admin interface uses the proper css when running on the\ndev machine\n","lang":"Python","license":"agpl-3.0","repos":"sbsdev\/daisyproducer,sbsdev\/daisyproducer,sbsdev\/daisyproducer,sbsdev\/daisyproducer"} {"commit":"3b75a6f3654e8f325060779ca56b6df93fe0cabe","old_file":"genome_designer\/main\/demo_view_overrides.py","new_file":"genome_designer\/main\/demo_view_overrides.py","old_contents":"\"\"\"View overrides for demo mode.\n\"\"\"\n\nfrom django.contrib.auth import authenticate\nfrom django.contrib.auth import login\nfrom django.http import HttpResponseRedirect\n\n\ndef login_demo_account(request):\n new_user = authenticate(username='gmcdev',\n password='g3n3d3z')\n login(request, new_user)\n return HttpResponseRedirect(\"\/\")\n","new_contents":"\"\"\"View overrides for demo mode.\n\"\"\"\n\nfrom django.contrib.auth import authenticate\nfrom django.contrib.auth import login\nfrom django.http import HttpResponseRedirect\n\n\ndef login_demo_account(request):\n new_user = authenticate(username='gmcdev',\n password='g3n3d3z')\n login(request, new_user)\n redirect_url = request.GET.get('next', '\/')\n return HttpResponseRedirect(redirect_url)\n","subject":"Handle redirect_url in demo login bypass.","message":"Handle redirect_url in demo login bypass.\n","lang":"Python","license":"mit","repos":"churchlab\/millstone,churchlab\/millstone,churchlab\/millstone,woodymit\/millstone_accidental_source,woodymit\/millstone_accidental_source,woodymit\/millstone,woodymit\/millstone_accidental_source,woodymit\/millstone_accidental_source,woodymit\/millstone,woodymit\/millstone,woodymit\/millstone,churchlab\/millstone"} {"commit":"dbe622d2297d62f61adf34e17de7c84d0cffbeaf","old_file":"project\/project\/local_settings_example.py","new_file":"project\/project\/local_settings_example.py","old_contents":"DEBUG = True\n\nADMINS = (\n ('Zaphod Beeblebrox', 'hoopyfrood@heartofgold.com'),\n)\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = '^lkajsdlfkjaoif09ijoi23092309i02[93ip2j3[r29u3[0923jorij'\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.mysql',\n 'NAME': 'db_name_here',\n 'USER': 'db_user_here',\n 'PASSWORD': 'db_password_here',\n 'HOST': 'localhost',\n 'PORT': '3306',\n }\n}\n\nGOOGLE_ACCOUNT_CODE = \"UA-XXXXXXX-XX\"","new_contents":"DEBUG = True\n\nADMINS = (\n ('Zaphod Beeblebrox', 'hoopyfrood@heartofgold.com'),\n)\n\n# SECURITY WARNING: keep the secret key used in production secret!\nSECRET_KEY = '^lkajsdlfkjaoif09ijoi23092309i02[93ip2j3[r29u3[0923jorij'\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.mysql',\n 'NAME': 'db_name_here',\n 'USER': 'db_user_here',\n 'PASSWORD': 'db_password_here',\n 'HOST': 'localhost',\n 'PORT': '3306',\n }\n}\n\nEMAIL_HOST = ''\nEMAIL_HOST_USER = ''\nEMAIL_HOST_PASSWORD = ''\nEMAIL_PORT = 587\nDEFAULT_FROM_EMAIL = 'My Site Admin '\n\nGOOGLE_ACCOUNT_CODE = \"UA-XXXXXXX-XX\"","subject":"Add email settings example to localsettings","message":"Add email settings example to localsettings\n","lang":"Python","license":"mit","repos":"colbypalmer\/cp-project-template,colbypalmer\/cp-project-template,colbypalmer\/cp-project-template"} {"commit":"b6f3c619e8c3fa375ac9b66e7ce555c77f02f152","old_file":"pytest_raisesregexp\/plugin.py","new_file":"pytest_raisesregexp\/plugin.py","old_contents":"import re\nimport py.code\nimport pytest\n\n\ndef pytest_namespace():\n return {'raises_regexp': raises_regexp}\n\n\nclass raises_regexp(object):\n def __init__(self, expected_exception, regexp):\n self.exception = expected_exception\n self.regexp = regexp\n self.excinfo = None\n\n def __enter__(self):\n self.excinfo = object.__new__(py.code.ExceptionInfo)\n return self.excinfo\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n __tracebackhide__ = True\n if exc_type is None:\n pytest.fail('DID NOT RAISE %s' % self.exception)\n\n self.excinfo.__init__((exc_type, exc_val, exc_tb))\n\n if not issubclass(exc_type, self.exception):\n pytest.fail('%s RAISED instead of %s' % (exc_type, self.exception))\n\n if not re.search(self.regexp, str(exc_val)):\n pytest.fail('pattern \"%s\" not found in \"%s\"' % (self.regexp, str(exc_val)))\n\n return True\n","new_contents":"import re\nimport py.code\nimport pytest\n\n\ndef pytest_namespace():\n return {'raises_regexp': raises_regexp}\n\n\nclass raises_regexp(object):\n def __init__(self, expected_exception, regexp):\n self.exception = expected_exception\n self.regexp = regexp\n self.excinfo = None\n\n def __enter__(self):\n self.excinfo = object.__new__(py.code.ExceptionInfo)\n return self.excinfo\n\n def __exit__(self, exc_type, exc_val, exc_tb):\n __tracebackhide__ = True\n if exc_type is None:\n pytest.fail('DID NOT RAISE %s' % self.exception)\n\n self.excinfo.__init__((exc_type, exc_val, exc_tb))\n\n if not issubclass(exc_type, self.exception):\n pytest.fail('%s RAISED instead of %s\\n%s' % (exc_type, self.exception, repr(exc_val)))\n\n if not re.search(self.regexp, str(exc_val)):\n pytest.fail('pattern \"%s\" not found in \"%s\"' % (self.regexp, str(exc_val)))\n\n return True\n","subject":"Add originally raised exception value to pytest error message","message":"Add originally raised exception value to pytest error message\n","lang":"Python","license":"mit","repos":"kissgyorgy\/pytest-raisesregexp"} {"commit":"69cf5602ba9dd9d7e0a89c169682ac72e2e18a67","old_file":"everywhere\/base.py","new_file":"everywhere\/base.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n\ndef fib(number: int) -> int:\n '''\n >>> fib(10)\n 55\n '''\n if number < 2:\n return number\n else:\n return fib(number-1) + fib(number-2)\n\n\ndef hello() -> None:\n '''\n >>> hello()\n 'Hello World'\n '''\n return 'Hello World'\n\n\ndef add42(number: int) -> int:\n '''\n >>> add42(100)\n 142\n '''\n return number + 42\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n\ndef fib(number: int) -> int:\n '''\n >>> fib(10)\n 55\n '''\n if number < 2:\n return number\n else:\n return fib(number-1) + fib(number-2)\n\n\ndef hello() -> str:\n '''\n >>> hello()\n 'Hello World'\n '''\n return 'Hello World'\n\n\ndef add42(number: int) -> int:\n '''\n >>> add42(100)\n 142\n '''\n return number + 42\n","subject":"Fix return type of hello","message":"Fix return type of hello\n","lang":"Python","license":"bsd-2-clause","repos":"wdv4758h\/python-everywhere,wdv4758h\/python-everywhere,wdv4758h\/python-everywhere"} {"commit":"5b8da0d318d7b37b3f1a3d868980507b15aa4213","old_file":"salt\/renderers\/json.py","new_file":"salt\/renderers\/json.py","old_contents":"from __future__ import absolute_import\nimport json\n\ndef render(json_data, env='', sls='', **kws):\n if not isinstance(json_data, basestring):\n json_data = json_data.read()\n\n if json_data.startswith('#!'):\n json_data = json_data[json_data.find('\\n')+1:]\n\n return json.loads(json_data)\n\n","new_contents":"from __future__ import absolute_import\nimport json\n\ndef render(json_data, env='', sls='', **kws):\n if not isinstance(json_data, basestring):\n json_data = json_data.read()\n\n if json_data.startswith('#!'):\n json_data = json_data[json_data.find('\\n')+1:]\n if not json_data.strip():\n return {}\n return json.loads(json_data)\n\n","subject":"Add missed changes for the previous commit.","message":"Add missed changes for the previous commit.\n","lang":"Python","license":"apache-2.0","repos":"saltstack\/salt,saltstack\/salt,saltstack\/salt,saltstack\/salt,saltstack\/salt"} {"commit":"3ad64c06f917efdaa94ad5debc23941f4d95105a","old_file":"fuzzy_happiness\/attributes.py","new_file":"fuzzy_happiness\/attributes.py","old_contents":"#!\/usr\/bin\/python\n#\n# Copyright 2013 Rackspace Australia\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\n# Work out what fields to anonymize. To run a simple example, do this:\n# cd \n# . .tox\/py27\/bin\/activate\n# \/attributes.py\n\n\nimport inspect\n\nfrom nova.db.sqlalchemy import models\n\n\ndef load_configuration():\n configs = {}\n\n for name, obj in inspect.getmembers(models):\n if not inspect.isclass(obj):\n continue\n\n if not issubclass(obj, models.NovaBase):\n continue\n\n if not hasattr(obj, '__confidential__'):\n continue\n\n configs[name] = obj.__confidential__\n\n return configs\n\n\nif __name__ == '__main__':\n print load_configuration()\n","new_contents":"#!\/usr\/bin\/python\n#\n# Copyright 2013 Rackspace Australia\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\n# Work out what fields to anonymize. To run a simple example, do this:\n# cd \n# . .tox\/py27\/bin\/activate\n# \/attributes.py\n\n\nimport inspect\n\nfrom nova.db.sqlalchemy import models\n\n\ndef load_configuration():\n configs = {}\n\n for name, obj in inspect.getmembers(models):\n if not inspect.isclass(obj):\n continue\n\n if not issubclass(obj, models.NovaBase):\n continue\n\n attrs_missing = []\n for required_attr in ['__tablename__', '__confidential__']:\n if not hasattr(obj, required_attr):\n attrs_missing.append(required_attr)\n\n if attrs_missing:\n print ('Required attributes %s missing from %s'\n %(', '.join(attrs_missing), name))\n continue\n\n configs[obj.__tablename__] = obj.__confidential__\n\n return configs\n\n\nif __name__ == '__main__':\n print load_configuration()\n","subject":"Return table name not model object name.","message":"Return table name not model object name.\n","lang":"Python","license":"apache-2.0","repos":"rcbau\/fuzzy-happiness"} {"commit":"484e0b77aeb5beb08328a6d6d4155f449107b710","old_file":"conanfile.py","new_file":"conanfile.py","old_contents":"from conans import ConanFile\nfrom conans.tools import download, unzip\nimport os\n\nVERSION = \"0.0.1\"\n\n\nclass IWYUCTargetCmakeConan(ConanFile):\n name = \"iwyu-target-cmake\"\n version = os.environ.get(\"CONAN_VERSION_OVERRIDE\", VERSION)\n generators = \"cmake\"\n requires = (\"cmake-include-guard\/master@smspillaz\/cmake-include-guard\",\n \"tooling-find-pkg-util\/master@smspillaz\/tooling-find-pkg-util\",\n \"tooling-cmake-util\/master@smspillaz\/tooling-cmake-util\",\n \"cmake-unit\/master@smspillaz\/cmake-unit\")\n url = \"http:\/\/github.com\/polysquare\/iwyu-target-cmake\"\n license = \"MIT\"\n\n def source(self):\n zip_name = \"iwyu-target-cmake.zip\"\n download(\"https:\/\/github.com\/polysquare\/\"\n \"iwyu-target-cmake\/archive\/{version}.zip\"\n \"\".format(version=\"v\" + VERSION),\n zip_name)\n unzip(zip_name)\n os.unlink(zip_name)\n\n def package(self):\n self.copy(pattern=\"*.cmake\",\n dst=\"cmake\/iwyu-target-cmake\",\n src=\"iwyu-target-cmake-\" + VERSION,\n keep_path=True)\n","new_contents":"from conans import ConanFile\nfrom conans.tools import download, unzip\nimport os\n\nVERSION = \"0.0.2\"\n\n\nclass IWYUCTargetCmakeConan(ConanFile):\n name = \"iwyu-target-cmake\"\n version = os.environ.get(\"CONAN_VERSION_OVERRIDE\", VERSION)\n generators = \"cmake\"\n requires = (\"cmake-include-guard\/master@smspillaz\/cmake-include-guard\",\n \"tooling-find-pkg-util\/master@smspillaz\/tooling-find-pkg-util\",\n \"tooling-cmake-util\/master@smspillaz\/tooling-cmake-util\",\n \"cmake-unit\/master@smspillaz\/cmake-unit\")\n url = \"http:\/\/github.com\/polysquare\/iwyu-target-cmake\"\n license = \"MIT\"\n\n def source(self):\n zip_name = \"iwyu-target-cmake.zip\"\n download(\"https:\/\/github.com\/polysquare\/\"\n \"iwyu-target-cmake\/archive\/{version}.zip\"\n \"\".format(version=\"v\" + VERSION),\n zip_name)\n unzip(zip_name)\n os.unlink(zip_name)\n\n def package(self):\n self.copy(pattern=\"*.cmake\",\n dst=\"cmake\/iwyu-target-cmake\",\n src=\"iwyu-target-cmake-\" + VERSION,\n keep_path=True)\n","subject":"Bump version: 0.0.1 -> 0.0.2","message":"Bump version: 0.0.1 -> 0.0.2\n\n[ci skip]\n","lang":"Python","license":"mit","repos":"polysquare\/iwyu-target-cmake,polysquare\/include-what-you-use-target-cmake"} {"commit":"dbbd6e1e87964db6b2279a661a63751da31213e5","old_file":"millipede.py","new_file":"millipede.py","old_contents":"#!\/usr\/bin\/env python3\n\n\nclass millipede:\n def __init__(self, size, comment=None):\n self._millipede = \"\"\n if comment:\n self._millipede = comment + \"\\n\\n\"\n\n self._millipede += \" ╚⊙ ⊙╝ \\n\"\n padding = 2\n direction = -1\n while (size):\n for i in range(0, padding):\n self._millipede += \" \"\n self._millipede += \"╚═(███)═╝\\n\"\n padding += direction\n\n if padding == 0:\n direction = 1\n elif padding == 4:\n padding = 3\n direction = -1\n size -= 1\n\n def __str__(self):\n return self._millipede\n\nif __name__ == \"__main__\":\n import argparse\n\n parser = argparse.ArgumentParser(description='Millipede generator')\n parser.add_argument('size', metavar='s', type=int, help='the size of the millipede')\n parser.add_argument('comment', metavar='c', type=str, help='the comment', nargs=\"?\")\n args = parser.parse_args()\n\n print(millipede(args.size, args.comment))\n","new_contents":"#!\/usr\/bin\/env python3\n\n\nclass millipede:\n\n def __init__(self, size, comment=None, reverse=False):\n self._padding_offsets = [2, 1, 0, 1, 2, 3, 4, 4, 3]\n\n head = \" ╔⊙ ⊙╗\\n\" if reverse else \" ╚⊙ ⊙╝\\n\"\n body = \"\".join([\n \"{}{}\\n\".format(\n \" \" * self._padding_offsets[(x + 3) % 9 if reverse else x % 9],\n \"╔═(███)═╗\" if reverse else \"╚═(███)═╝\"\n )\n for x in range(size)\n ])\n\n self._millipede = \"\"\n if reverse:\n self._millipede += body + head\n if comment:\n self._millipede += \"\\n\" + comment\n else:\n if comment:\n self._millipede += comment + \"\\n\\n\"\n self._millipede += head + body\n\n def __str__(self):\n return self._millipede\n\nif __name__ == \"__main__\":\n import argparse\n\n parser = argparse.ArgumentParser(description='Millipede generator')\n parser.add_argument('size', metavar='s', type=int, help='the size of the millipede')\n parser.add_argument('comment', metavar='c', type=str, help='the comment', nargs=\"?\")\n parser.add_argument('-r', '--reverse', action='store_true', help='reverse the millipede')\n args = parser.parse_args()\n\n print(millipede(args.size, comment=args.comment, reverse=args.reverse))\n","subject":"Rewrite body generation and add reverse option","message":"Rewrite body generation and add reverse option\n","lang":"Python","license":"bsd-3-clause","repos":"evadot\/millipede-python,getmillipede\/millipede-python,moul\/millipede-python,EasonYi\/millipede-python,EasonYi\/millipede-python,evadot\/millipede-python,moul\/millipede-python,getmillipede\/millipede-python"} {"commit":"8233abab6084db39df064b87d256fd0caffecb89","old_file":"simpy\/test\/test_simulation.py","new_file":"simpy\/test\/test_simulation.py","old_contents":"from simpy import Simulation, InterruptedException\n\ndef test_simple_process():\n def pem(ctx, result):\n while True:\n result.append(ctx.now)\n yield ctx.wait(1)\n\n result = []\n Simulation(pem, result).simulate(until=4)\n\n assert result == [0, 1, 2, 3]\n\ndef test_interrupt():\n def pem(ctx):\n try:\n yield ctx.wait(10)\n raise RuntimeError('Expected an interrupt')\n except InterruptedException:\n pass\n\n def root(ctx):\n process = ctx.fork(pem)\n yield ctx.wait(5)\n process.interrupt()\n\n Simulation(root).simulate(until=20)\n\ndef test_wait_for_process():\n def pem(ctx):\n yield ctx.wait(10)\n\n def root(ctx):\n yield ctx.wait(ctx.fork(pem))\n assert ctx.now == 10\n\n Simulation(root).simulate(until=20)\n\ndef test_process_result():\n def pem(ctx):\n yield ctx.wait(10)\n ctx.exit('oh noes, i am dead x_x')\n\n def root(ctx):\n result = yield ctx.wait(ctx.fork(pem))\n assert result == 'oh noes, i am dead x_x'\n\n Simulation(root).simulate(until=20)\n","new_contents":"from simpy import Simulation, InterruptedException\n\ndef test_simple_process():\n def pem(ctx, result):\n while True:\n result.append(ctx.now)\n yield ctx.wait(1)\n\n result = []\n Simulation(pem, result).simulate(until=4)\n\n assert result == [0, 1, 2, 3]\n\ndef test_interrupt():\n def root(ctx):\n def pem(ctx):\n try:\n yield ctx.wait(10)\n raise RuntimeError('Expected an interrupt')\n except InterruptedException:\n pass\n\n process = ctx.fork(pem)\n yield ctx.wait(5)\n process.interrupt()\n\n Simulation(root).simulate(until=20)\n\ndef test_wait_for_process():\n def root(ctx):\n def pem(ctx):\n yield ctx.wait(10)\n\n yield ctx.wait(ctx.fork(pem))\n assert ctx.now == 10\n\n Simulation(root).simulate(until=20)\n\ndef test_process_result():\n def root(ctx):\n def pem(ctx):\n yield ctx.wait(10)\n ctx.exit('oh noes, i am dead x_x')\n\n result = yield ctx.wait(ctx.fork(pem))\n assert result == 'oh noes, i am dead x_x'\n\n Simulation(root).simulate(until=20)\n","subject":"Define subprocesses in the context of the root process. Maybe this is more readable?","message":"Define subprocesses in the context of the root process. Maybe this is more readable?\n","lang":"Python","license":"mit","repos":"Uzere\/uSim"} {"commit":"a51f5e108f6fb81fce5d99b53888f2a2954fb9a6","old_file":"server_app\/__main__.py","new_file":"server_app\/__main__.py","old_contents":"import sys\nimport os\nimport logging\nimport time\n\nif not os.path.exists(os.path.expanduser(\"~\/.chatserver\")):\n os.makedirs(os.path.expanduser(\"~\/.chatserver\"))\nlogging.basicConfig(filename=os.path.expanduser(\"~\/.chatserver\/chat-\"+time.strftime(\"%d-%m-%Y.log\"), level=logging.DEBUG))\n\nsys.stderr.close()\nsys.stdout.close()\nsys.stdin.close()\n\nfrom app import app, db, main, socketio\ndb.create_all()\napp.register_blueprint(main)\n\nport = app.config['PORT']\nif len(sys.argv) == 2:\n port = int(sys.argv[1])\n\nlogging.info(\"Chat server is now running on 0.0.0.0:%r\" % port)\nsocketio.run(app, host=\"0.0.0.0\", port=port)\n","new_contents":"import sys\nimport os\nimport logging\nimport time\n\nif not os.path.exists(os.path.expanduser(\"~\/.chatserver\")):\n os.makedirs(os.path.expanduser(\"~\/.chatserver\"))\nlogging.basicConfig(filename=os.path.expanduser(\"~\/.chatserver\/chat-\"+time.strftime(\"%d-%m-%Y.log\")), level=logging.DEBUG)\n\nsys.stderr.close()\nsys.stdout.close()\nsys.stdin.close()\n\nfrom app import app, db, main, socketio\ndb.create_all()\napp.register_blueprint(main)\n\nport = app.config['PORT']\nif len(sys.argv) == 2:\n port = int(sys.argv[1])\n\nlogging.info(\"Chat server is now running on 0.0.0.0:%r\" % port)\nsocketio.run(app, host=\"0.0.0.0\", port=port)\n","subject":"Make logger sort by date","message":"Make logger sort by date\n","lang":"Python","license":"bsd-3-clause","repos":"jos0003\/Chat,jos0003\/Chat,jos0003\/Chat,jos0003\/Chat,jos0003\/Chat"} {"commit":"1b0edd2eeb722397e9c6c7da04ab6cbd3865a476","old_file":"reddit_adzerk\/adzerkads.py","new_file":"reddit_adzerk\/adzerkads.py","old_contents":"from urllib import quote\n\nfrom pylons import c, g\n\nfrom r2.lib.pages import Ads as BaseAds\nfrom r2.models.subreddit import DefaultSR\n\n\nclass Ads(BaseAds):\n def __init__(self):\n BaseAds.__init__(self)\n\n url_key = \"adzerk_https_url\" if c.secure else \"adzerk_url\"\n site_name = getattr(c.site, \"analytics_name\", c.site.name)\n\n # adzerk reporting is easier when not using a space in the tag\n if isinstance(c.site, DefaultSR):\n site_name = \"-reddit.com\"\n\n self.ad_url = g.config[url_key].format(\n subreddit=quote(site_name.lower()),\n origin=c.request_origin,\n )\n self.frame_id = \"ad_main\"\n","new_contents":"from urllib import quote\n\nfrom pylons import c, g\n\nfrom r2.lib.pages import Ads as BaseAds\nfrom r2.models.subreddit import DefaultSR\n\n\nFRONTPAGE_NAME = \"-reddit.com\"\n\nclass Ads(BaseAds):\n def __init__(self):\n BaseAds.__init__(self)\n\n url_key = \"adzerk_https_url\" if c.secure else \"adzerk_url\"\n site_name = getattr(c.site, \"analytics_name\", c.site.name)\n\n # adzerk reporting is easier when not using a space in the tag\n if isinstance(c.site, DefaultSR):\n site_name = FRONTPAGE_NAME\n\n self.ad_url = g.config[url_key].format(\n subreddit=quote(site_name.lower()),\n origin=c.request_origin,\n )\n self.frame_id = \"ad_main\"\n","subject":"Move special frontpage name to variable.","message":"Move special frontpage name to variable.\n","lang":"Python","license":"bsd-3-clause","repos":"madbook\/reddit-plugin-adzerk,madbook\/reddit-plugin-adzerk,madbook\/reddit-plugin-adzerk"} {"commit":"2cb03ff8c3d21f36b95103eaf9ae0fb3e43077bd","old_file":"pinax_theme_bootstrap\/templatetags\/pinax_theme_bootstrap_tags.py","new_file":"pinax_theme_bootstrap\/templatetags\/pinax_theme_bootstrap_tags.py","old_contents":"from django import template\nfrom django.contrib.messages.utils import get_level_tags\n\n\nLEVEL_TAGS = get_level_tags()\n\nregister = template.Library()\n\n\n@register.simple_tag()\ndef get_message_tags(message):\n \"\"\"\n Returns the message's level_tag prefixed with Bootstrap's \"alert-\" prefix\n along with any tags included in message.extra_tags\n\n Messages in Django >= 1.7 have a message.level_tag attr\n \"\"\"\n level_tag = LEVEL_TAGS[message.level]\n if level_tag == u\"error\":\n level_tag = u\"danger\"\n\n alert_level_tag = u\"alert-{tag}\".format(tag=level_tag)\n\n tags = [alert_level_tag]\n if message.extra_tags:\n tags.append(message.extra_tags)\n\n return u\" \".join(tags)\n","new_contents":"from django import template\nfrom django.contrib.messages.utils import get_level_tags\nfrom django.utils.encoding import force_text\n\n\nLEVEL_TAGS = get_level_tags()\n\nregister = template.Library()\n\n\n@register.simple_tag()\ndef get_message_tags(message):\n \"\"\"\n Returns the message's level_tag prefixed with Bootstrap's \"alert-\" prefix\n along with any tags included in message.extra_tags\n\n Messages in Django >= 1.7 have a message.level_tag attr\n \"\"\"\n level_tag = force_text(LEVEL_TAGS.get(message.level, ''), strings_only=True)\n if level_tag == u\"error\":\n level_tag = u\"danger\"\n\n alert_level_tag = u\"alert-{tag}\".format(tag=level_tag)\n\n tags = [alert_level_tag]\n extra_tags = force_text(message.extra_tags, strings_only=True)\n if extra_tags:\n tags.append(extra_tags)\n\n return u\" \".join(tags)\n","subject":"Allow for lazy translation of message tags","message":"Allow for lazy translation of message tags\n","lang":"Python","license":"mit","repos":"foraliving\/foraliving,druss16\/danslist,grahamu\/pinax-theme-bootstrap,foraliving\/foraliving,grahamu\/pinax-theme-bootstrap,druss16\/danslist,jacobwegner\/pinax-theme-bootstrap,jacobwegner\/pinax-theme-bootstrap,druss16\/danslist,grahamu\/pinax-theme-bootstrap,jacobwegner\/pinax-theme-bootstrap,foraliving\/foraliving"} {"commit":"d20f9d5e56c4f430f6ca7b4ab03a279e34bbbd45","old_file":"importer\/tasks.py","new_file":"importer\/tasks.py","old_contents":"from .models import FileImport\nfrom .importers import ImportFailure\nfrom django.db import transaction\nimport celery\n\n\nassuming_failure_message = '{0} did not return True. Assuming failure.'\n\n\nprocessing_status = 'processing'\nprocessing_description = 'Processing the data in {filename}.'\n\n\nsuccess_status = 'success'\nsuccess_description = 'The import appears to have completed successfully.'\n\n\n# The description for failures is the contents of the exception message.\nfailure_status = 'failure'\n\n\n@celery.task\n@transaction.atomic\ndef importer_asynchronous_task(import_pk, *args, **kwargs):\n logger = importer_asynchronous_task.get_logger()\n\n import_instance = FileImport.objects.get(pk=import_pk)\n ImportType = import_instance.get_related_importer(**kwargs)\n\n if ImportType is None:\n import_instance.status = 30\n return False\n\n importer = ImportType()\n\n import_instance.status = processing_status\n import_instance.status_description = 'Currently processing file'\n import_instance.save()\n\n import_context = import_instance.get_context()\n\n try:\n if importer.process(import_context, logger) is True:\n import_instance.status = success_status\n import_instance.status_description = success_description\n import_instance.save()\n\n else:\n raise ImportFailure(assuming_failure_message.format(\n importer.__class__.__name__\n ))\n\n except ImportFailure, e:\n import_instance.status = failure_status\n import_instance.status_description = e.message\n import_instance.save()\n\n return True\n","new_contents":"from .models import FileImport\nfrom .importers import ImportFailure\nfrom django.db import transaction\nimport celery\n\n\nassuming_failure_message = '{0} did not return True. Assuming failure.'\n\n\nprocessing_status = 'processing'\nprocessing_description = 'Processing the data in {filename}.'\n\n\nsuccess_status = 'success'\nsuccess_description = 'The import appears to have completed successfully.'\n\n\n# The description for failures is the contents of the exception message.\nfailure_status = 'failure'\n\n\n@celery.shared_task\n@transaction.atomic\ndef importer_asynchronous_task(import_pk, *args, **kwargs):\n logger = importer_asynchronous_task.get_logger()\n\n import_instance = FileImport.objects.get(pk=import_pk)\n ImportType = import_instance.get_related_importer(**kwargs)\n\n if ImportType is None:\n import_instance.status = 30\n return False\n\n importer = ImportType()\n\n import_instance.status = processing_status\n import_instance.status_description = 'Currently processing file'\n import_instance.save()\n\n import_context = import_instance.get_context()\n\n try:\n if importer.process(import_context, logger) is True:\n import_instance.status = success_status\n import_instance.status_description = success_description\n import_instance.save()\n\n else:\n raise ImportFailure(assuming_failure_message.format(\n importer.__class__.__name__\n ))\n\n except ImportFailure, e:\n import_instance.status = failure_status\n import_instance.status_description = e.message\n import_instance.save()\n\n return True\n","subject":"Use shared_task instead of task.","message":"Use shared_task instead of task.\n","lang":"Python","license":"mit","repos":"monokrome\/django-drift"} {"commit":"3973e0d2591b2554e96da0a22b2d723a71d2423e","old_file":"imgaug\/augmenters\/__init__.py","new_file":"imgaug\/augmenters\/__init__.py","old_contents":"from __future__ import absolute_import\nfrom imgaug.augmenters.arithmetic import *\nfrom imgaug.augmenters.blur import *\nfrom imgaug.augmenters.color import *\nfrom imgaug.augmenters.contrast import GammaContrast, SigmoidContrast, LogContrast, LinearContrast\nfrom imgaug.augmenters.convolutional import *\nfrom imgaug.augmenters.flip import *\nfrom imgaug.augmenters.geometric import *\nfrom imgaug.augmenters.meta import *\nfrom imgaug.augmenters.overlay import *\nfrom imgaug.augmenters.segmentation import *\nfrom imgaug.augmenters.size import *\n","new_contents":"from __future__ import absolute_import\nfrom imgaug.augmenters.arithmetic import *\nfrom imgaug.augmenters.blur import *\nfrom imgaug.augmenters.color import *\nfrom imgaug.augmenters.contrast import *\nfrom imgaug.augmenters.convolutional import *\nfrom imgaug.augmenters.flip import *\nfrom imgaug.augmenters.geometric import *\nfrom imgaug.augmenters.meta import *\nfrom imgaug.augmenters.overlay import *\nfrom imgaug.augmenters.segmentation import *\nfrom imgaug.augmenters.size import *\n","subject":"Switch import from contrast to all","message":"Switch import from contrast to all\n\nChange import from contrast.py in\naugmenters\/__init__.py to * instead of\nselective, as * should not import private\nmethods anyways.\n","lang":"Python","license":"mit","repos":"aleju\/ImageAugmenter,aleju\/imgaug,aleju\/imgaug"} {"commit":"2c7634f3af71014547437bd78239acef9977468a","old_file":"endpoints_management\/__init__.py","new_file":"endpoints_management\/__init__.py","old_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import absolute_import\n\nfrom . import auth, config, control, gen\n\n__version__ = '1.2.1'\n\nUSER_AGENT = u'ESP'\nSERVICE_AGENT = u'EF_PYTHON\/' + __version__\n\n__all__ = ['auth', 'config', 'control', 'gen']\n","new_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import absolute_import\n\nfrom . import auth, config, control, gen\n\n__version__ = '1.3.0'\n\nUSER_AGENT = u'ESP'\nSERVICE_AGENT = u'EF_PYTHON\/' + __version__\n\n__all__ = ['auth', 'config', 'control', 'gen']\n","subject":"Bump minor version (1.2.1 -> 1.3.0)","message":"Bump minor version (1.2.1 -> 1.3.0)\n\nRationale: support for quotas\/ratelimiting.\n","lang":"Python","license":"apache-2.0","repos":"inklesspen\/endpoints-management-python,cloudendpoints\/endpoints-management-python"} {"commit":"ad558a5acc93e1e5206ed27b2dc679089b277890","old_file":"me_api\/app.py","new_file":"me_api\/app.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import absolute_import, unicode_literals\n\n\nfrom flask import Flask\n\nfrom .middleware.me import me\nfrom .middleware import github, keybase, medium\nfrom .cache import cache\n\n\ndef create_app(config):\n app = Flask(__name__)\n app.config.from_object(config)\n cache.init_app(app)\n\n modules = config.modules['modules']\n blueprints = {\n 'github': github.github_api,\n 'keybase': keybase.keybase_api,\n 'medium': medium.medium_api\n }\n\n app.register_blueprint(me)\n for module in modules.keys():\n app.register_blueprint(blueprints[module])\n\n return app\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import absolute_import, unicode_literals\n\n\nfrom flask import Flask\n\nfrom .middleware.me import me\nfrom .cache import cache\n\n\ndef _register_module(app, module):\n if module == 'github':\n from .middleware import github\n app.register_blueprint(github.github_api)\n elif module == 'keybase':\n from .middleware import keybase\n app.register_blueprint(keybase.keybase_api)\n elif module == 'medium':\n from .middleware import medium\n app.register_blueprint(medium.medium_api)\n\n\ndef create_app(config):\n app = Flask(__name__)\n app.config.from_object(config)\n cache.init_app(app)\n\n modules = config.modules['modules']\n app.register_blueprint(me)\n for module in modules.keys():\n _register_module(app, module)\n\n return app\n","subject":"Fix giant bug: crash when don't config all modules","message":"Fix giant bug: crash when don't config all modules\n\nthat's bacause you import all the modules\n> from .middleware import github, keybase, medium\nwhile each module need to get configurations from modules.json, e.g.\n> config = Config.modules['modules']['github']\nbut can't get anything at all, so it will crash.\n\nthat's not the correct behaviour, so we just import the desired module\nand then register it. The solution isn't very elegant.\n","lang":"Python","license":"mit","repos":"lord63\/me-api"} {"commit":"42b330e5629b25db45e7a0f3f08bdb21e608b106","old_file":"skimage\/viewer\/qt.py","new_file":"skimage\/viewer\/qt.py","old_contents":"has_qt = True\n\ntry:\n from matplotlib.backends.qt_compat import QtGui, QtCore, QtWidgets\nexcept ImportError:\n try:\n from matplotlib.backends.qt4_compat import QtGui, QtCore\n QtWidgets = QtGui\n except ImportError:\n # Mock objects\n class QtGui(object):\n QMainWindow = object\n QDialog = object\n QWidget = object\n\n class QtCore_cls(object):\n class Qt(object):\n TopDockWidgetArea = None\n BottomDockWidgetArea = None\n LeftDockWidgetArea = None\n RightDockWidgetArea = None\n\n def Signal(self, *args, **kwargs): \n pass\n\n QtCore = QtWidgets = QtCore_cls()\n\n has_qt = False\n\nQt = QtCore.Qt\nSignal = QtCore.Signal\n","new_contents":"has_qt = True\n\ntry:\n from matplotlib.backends.qt_compat import QtGui, QtCore, QtWidgets\nexcept ImportError:\n try:\n from matplotlib.backends.qt4_compat import QtGui, QtCore\n QtWidgets = QtGui\n except ImportError:\n # Mock objects\n class QtGui(object):\n QMainWindow = object\n QDialog = object\n QWidget = object\n\n class QtCore_cls(object):\n class Qt(object):\n TopDockWidgetArea = None\n BottomDockWidgetArea = None\n LeftDockWidgetArea = None\n RightDockWidgetArea = None\n\n def Signal(self, *args, **kwargs): \n pass\n\n QWidget = object\n\n QtCore = QtWidgets = QtCore_cls()\n\n has_qt = False\n\nQt = QtCore.Qt\nSignal = QtCore.Signal\n","subject":"Add QWidget to the mock Qt","message":"Add QWidget to the mock Qt\n","lang":"Python","license":"bsd-3-clause","repos":"ClinicalGraphics\/scikit-image,bsipocz\/scikit-image,rjeli\/scikit-image,juliusbierk\/scikit-image,vighneshbirodkar\/scikit-image,pratapvardhan\/scikit-image,juliusbierk\/scikit-image,keflavich\/scikit-image,warmspringwinds\/scikit-image,Britefury\/scikit-image,oew1v07\/scikit-image,bennlich\/scikit-image,ClinicalGraphics\/scikit-image,rjeli\/scikit-image,ajaybhat\/scikit-image,ajaybhat\/scikit-image,keflavich\/scikit-image,michaelaye\/scikit-image,blink1073\/scikit-image,paalge\/scikit-image,Midafi\/scikit-image,michaelpacer\/scikit-image,michaelaye\/scikit-image,GaZ3ll3\/scikit-image,chriscrosscutler\/scikit-image,michaelpacer\/scikit-image,WarrenWeckesser\/scikits-image,youprofit\/scikit-image,oew1v07\/scikit-image,bsipocz\/scikit-image,Hiyorimi\/scikit-image,newville\/scikit-image,youprofit\/scikit-image,paalge\/scikit-image,robintw\/scikit-image,chriscrosscutler\/scikit-image,vighneshbirodkar\/scikit-image,dpshelio\/scikit-image,Midafi\/scikit-image,bennlich\/scikit-image,warmspringwinds\/scikit-image,robintw\/scikit-image,Britefury\/scikit-image,ofgulban\/scikit-image,GaZ3ll3\/scikit-image,dpshelio\/scikit-image,emon10005\/scikit-image,WarrenWeckesser\/scikits-image,blink1073\/scikit-image,pratapvardhan\/scikit-image,ofgulban\/scikit-image,emon10005\/scikit-image,rjeli\/scikit-image,ofgulban\/scikit-image,vighneshbirodkar\/scikit-image,Hiyorimi\/scikit-image,jwiggins\/scikit-image,paalge\/scikit-image,jwiggins\/scikit-image,newville\/scikit-image"} {"commit":"bd3d97cefe61886ab8c2fa24eecd624ca1c6f751","old_file":"profile_collection\/startup\/90-settings.py","new_file":"profile_collection\/startup\/90-settings.py","old_contents":"import logging\n\n# metadata set at startup\nRE.md['owner'] = 'xf11id'\nRE.md['beamline_id'] = 'CHX'\n# removing 'custom' as it is raising an exception in 0.3.2\n# gs.RE.md['custom'] = {}\n\n\n\ndef print_scanid(name, doc):\n if name == 'start':\n print('Scan ID:', doc['scan_id'])\n print('Unique ID:', doc['uid'])\n\ndef print_md(name, doc):\n if name == 'start':\n print('Metadata:\\n', repr(doc))\n\nRE.subscribe(print_scanid)\n\n#from eiger_io.fs_handler import LazyEigerHandler\n#db.fs.register_handler(\"AD_EIGER\", LazyEigerHandler)\n","new_contents":"import logging\n\n# metadata set at startup\nRE.md['owner'] = 'xf11id'\nRE.md['beamline_id'] = 'CHX'\n# removing 'custom' as it is raising an exception in 0.3.2\n# gs.RE.md['custom'] = {}\n\ndef print_md(name, doc):\n if name == 'start':\n print('Metadata:\\n', repr(doc))\n\nRE.subscribe(print_scanid)\n\n#from eiger_io.fs_handler import LazyEigerHandler\n#db.fs.register_handler(\"AD_EIGER\", LazyEigerHandler)\n","subject":"Remove redundant Scan ID printing (there is another one elsewhere)","message":"Remove redundant Scan ID printing (there is another one elsewhere)\n","lang":"Python","license":"bsd-2-clause","repos":"NSLS-II-CHX\/ipython_ophyd,NSLS-II-CHX\/ipython_ophyd"} {"commit":"ad60b0cd3326c0729237afbb094d22f4415fb422","old_file":"laboratory\/experiment.py","new_file":"laboratory\/experiment.py","old_contents":"import traceback\n\nfrom laboratory.observation import Observation, Test\nfrom laboratory import exceptions\n\n\nclass Experiment(object):\n\n def __init__(self, name='Experiment', raise_on_mismatch=False):\n self.name = name\n self.raise_on_mismatch = raise_on_mismatch\n\n self._control = None\n self.observations = []\n\n def control(self):\n self._control = Observation('Control')\n return Test('Control', True, self._control)\n\n def candidate(self, name='Candidate'):\n observation = Observation(name)\n self.observations.append(observation)\n return Test(name, False, observation)\n\n def run(self):\n control = self._control\n if control is None:\n raise exceptions.LaboratoryException(\n 'Your experiment must record a control case'\n )\n\n match = self.compare(control, *self.observations)\n return control.value\n\n def compare(self, control, *candidates):\n for observation in candidates:\n if observation.failure or control.value != observation.value:\n return self._comparison_mismatch(control, observation)\n\n return True\n\n def _comparison_mismatch(self, control, observation):\n if self.raise_on_mismatch:\n if observation.failure:\n msg = '%s raised an exception:\\n%s' % (\n observation.name, traceback.format_exc(observation.exception)\n )\n else:\n msg = '%s does not match control value (%s != %s)' % (\n observation.name, control.value, observation.value\n )\n raise exceptions.MismatchException(msg)\n\n return False\n\n def publish(self):\n raise NotImplementedError\n","new_contents":"import traceback\n\nfrom laboratory.observation import Observation, Test\nfrom laboratory import exceptions\n\n\nclass Experiment(object):\n\n def __init__(self, name='Experiment', raise_on_mismatch=False):\n self.name = name\n self.raise_on_mismatch = raise_on_mismatch\n\n self._control = None\n self.observations = []\n\n def control(self):\n self._control = Observation('Control')\n return Test('Control', True, self._control)\n\n def candidate(self, name='Candidate'):\n observation = Observation(name)\n self.observations.append(observation)\n return Test(name, False, observation)\n\n def run(self):\n control = self._control\n if control is None:\n raise exceptions.LaboratoryException(\n 'Your experiment must record a control case'\n )\n\n match = self.compare(control, *self.observations)\n self.publish(match)\n return control.value\n\n def compare(self, control, *candidates):\n for observation in candidates:\n if observation.failure or control.value != observation.value:\n return self._comparison_mismatch(control, observation)\n\n return True\n\n def _comparison_mismatch(self, control, observation):\n if self.raise_on_mismatch:\n if observation.failure:\n msg = '%s raised an exception:\\n%s' % (\n observation.name, traceback.format_exc(observation.exception)\n )\n else:\n msg = '%s does not match control value (%s != %s)' % (\n observation.name, control.value, observation.value\n )\n raise exceptions.MismatchException(msg)\n\n return False\n\n def publish(self, match):\n return\n","subject":"Call Experiment.publish in run method","message":"Call Experiment.publish in run method\n","lang":"Python","license":"mit","repos":"joealcorn\/laboratory,shaunvxc\/laboratory"} {"commit":"9560ccf476a887c20b2373eca52f38f186b6ed58","old_file":"conanfile.py","new_file":"conanfile.py","old_contents":"from conans import ConanFile, CMake\n\nclass NostalgiaConan(ConanFile):\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n requires = \"jsoncpp\/1.9.2\", \"sdl2\/2.0.10@bincrafters\/stable\", \"qt\/5.14.1@bincrafters\/stable\", \"sqlite3\/3.31.0\", \"libiconv\/1.16\"\n generators = \"cmake\", \"cmake_find_package\", \"cmake_paths\"\n #default_options = {\n # \"sdl2:nas\": False\n #}\n\n def requirements(self):\n pass\n","new_contents":"from conans import ConanFile, CMake\n\nclass NostalgiaConan(ConanFile):\n settings = \"os\", \"compiler\", \"build_type\", \"arch\"\n requires = \"jsoncpp\/1.9.2\", \"sdl2\/2.0.10@bincrafters\/stable\"\n generators = \"cmake\", \"cmake_find_package\", \"cmake_paths\"\n #default_options = {\n # \"sdl2:nas\": False\n #}\n","subject":"Remove conan Qt, as it is currently being ignored","message":"[nostalgia] Remove conan Qt, as it is currently being ignored\n","lang":"Python","license":"mpl-2.0","repos":"wombatant\/nostalgia,wombatant\/nostalgia,wombatant\/nostalgia"} {"commit":"a284a69432b2e0052fd2da4121cf4512fc9423da","old_file":"lemon\/dashboard\/admin.py","new_file":"lemon\/dashboard\/admin.py","old_contents":"from django.conf import settings\n\nfrom lemon import extradmin as admin\nfrom lemon.dashboard import views\nfrom lemon.dashboard.base import dashboard, Widget\n\n\nclass DashboardAdmin(admin.AppAdmin):\n\n instance = dashboard\n\n @property\n def urls(self):\n return self.instance.get_urls(self), 'dashboard', 'dashboard'\n\n\nadmin.site.register_app('dashboard', DashboardAdmin)\n","new_contents":"from django.conf import settings\n\nfrom lemon import extradmin as admin\nfrom lemon.dashboard import views\nfrom lemon.dashboard.base import dashboard, Widget\n\n\nclass DashboardAdmin(admin.AppAdmin):\n\n dashboard = dashboard\n\n @property\n def urls(self):\n return self.dashboard.get_urls(self), 'dashboard', 'dashboard'\n\n\nadmin.site.register_app('dashboard', DashboardAdmin)\n","subject":"Rename instance to dashboard in DashboardAdmin","message":"Rename instance to dashboard in DashboardAdmin\n","lang":"Python","license":"bsd-3-clause","repos":"trilan\/lemon,trilan\/lemon,trilan\/lemon"} {"commit":"8541ec09e237f1401095d31177bdde9ac1adaa39","old_file":"util\/linkJS.py","new_file":"util\/linkJS.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\n\n\ndef linkJS(target_fn, file_list_fn, source_base, prologue=\"\", module_dirs=[]):\n with open(target_fn, \"wb\") as target:\n target.write(prologue)\n\n # Add files listed in file_list_fn\n with open(file_list_fn) as file_list:\n for source_fn in file_list:\n source_fn = source_fn.replace(\"\/\", os.path.sep).strip()\n if len(source_fn) > 0 and source_fn[0] != \"#\":\n addContents(source_base, source_fn, target)\n\n # Add all *.js files in module_dirs\n for module_base in module_dirs:\n for module_fn in os.listdir(module_base):\n if module_fn.endswith(\".js\"):\n addContents(module_base, module_fn, target)\n\n\ndef addContents(source_base, source_fn, target):\n target.write(\"\\n\\n\/\/ \" + source_fn + \"\\n\\n\")\n with open(os.path.join(source_base, source_fn)) as source:\n for line in source:\n target.write(line)\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os\n\n\ndef linkJS(target_fn, file_list_fn, source_base, prologue=\"\", module_dirs=[]):\n with open(target_fn, \"wb\") as target:\n target.write(prologue)\n\n # Add files listed in file_list_fn\n with open(file_list_fn) as file_list:\n for source_fn in file_list:\n source_fn = source_fn.replace(\"\/\", os.path.sep).strip()\n if len(source_fn) > 0 and source_fn[0] != \"#\":\n addContents(os.path.join(source_base, source_fn), target)\n\n # Add all *.js files in module_dirs\n for module_base in module_dirs:\n for module_fn in os.listdir(module_base):\n if module_fn.endswith(\".js\"):\n addContents(os.path.join(module_base, module_fn), target)\n\n\ndef addContents(source_fn, target):\n target.write(\"\\n\\n\/\/ \" + source_fn + \"\\n\\n\")\n with open(source_fn) as source:\n for line in source:\n target.write(line)\n","subject":"Include full path to original files","message":"Include full path to original files\n","lang":"Python","license":"mpl-2.0","repos":"MozillaSecurity\/funfuzz,nth10sd\/funfuzz,MozillaSecurity\/funfuzz,nth10sd\/funfuzz,MozillaSecurity\/funfuzz,nth10sd\/funfuzz"} {"commit":"72dd10849190fb191fdab4962996ea537322e103","old_file":"tests\/TestPluginManager.py","new_file":"tests\/TestPluginManager.py","old_contents":"import socket\nimport time\nfrom unittest import TestCase\nfrom PluginManager import PluginManager\n\nclass TestPluginManager(TestCase):\n def test_stop(self):\n class Plugin:\n def __init__(self):\n sock = socket.socket()\n sock.bind(('', 0)) # bind to any available port\n self._port = sock.getsockname()[1]\n sock.close()\n def get_port(self):\n return self._port\n plugin_manager = PluginManager(Plugin(), lambda: None)\n plugin_manager.start()\n time.sleep(1)\n plugin_manager.stop()\n plugin_manager.join()\n self.assertFalse(plugin_manager.is_alive())\n\n","new_contents":"import socket\nimport time\nfrom unittest import TestCase\nfrom PluginManager import PluginManager\n\nclass TestPluginManager(TestCase):\n def test_stop(self):\n class Plugin:\n def __init__(self):\n sock = socket.socket()\n sock.bind(('', 0)) # bind to any available port\n self._port = sock.getsockname()[1]\n sock.close()\n def get_port(self):\n return self._port\n plugin_manager = PluginManager(Plugin(), lambda: None)\n plugin_manager.start()\n time.sleep(0.01)\n plugin_manager.stop()\n plugin_manager.join()\n self.assertFalse(plugin_manager.is_alive())\n\n","subject":"Reduce sleep duration in PluginManager.stop() test","message":"Reduce sleep duration in PluginManager.stop() test\n","lang":"Python","license":"mit","repos":"ckaz18\/honeypot,laurenmalone\/honeypot,theplue\/honeypot,ckaz18\/honeypot,laurenmalone\/honeypot,theplue\/honeypot,coyle5280\/honeypot,coyle5280\/honeypot,ckaz18\/honeypot,theplue\/honeypot,theplue\/honeypot,coyle5280\/honeypot,ckaz18\/honeypot,laurenmalone\/honeypot,laurenmalone\/honeypot,coyle5280\/honeypot"} {"commit":"2117778d777120293e506eca9743f97619b5ad5c","old_file":"kiwi\/interface.py","new_file":"kiwi\/interface.py","old_contents":"class Menu(object):\n def __init__(self, dialog, items, title, caller = None):\n self.d = dialog\n self.caller = caller\n\n self.entries = []\n self.dispatch_table = {}\n tag = 1\n\n self.title = title\n\n for entry, func in items:\n self.entries.append(tuple([str(tag), entry]))\n self.dispatch_table[str(tag)] = func\n tag += 1\n\n def run(self, ret=None):\n code, tag = self.d.menu(self.title, choices=self.entries)\n if code == self.d.OK: self.dispatch(tag)\n if ret: ret()\n\n def dispatch(self, tag):\n if tag in self.dispatch_table:\n func = self.dispatch_table[tag]\n if isinstance(func, Menu):\n func.run(ret=self.run)\n else: func()\n\n","new_contents":"class MenuItem(object):\n def __init__(self, func=None):\n if func: self.function = func\n\n # Wrapper for child.function() that creates a call stack\n def run(self, ret=None):\n self.function()\n if ret: ret()\n\nclass Menu(MenuItem):\n def __init__(self, dialog, items, title):\n self.d = dialog\n\n self.entries = []\n self.dispatch_table = {}\n tag = 1\n\n self.title = title\n\n for entry, func in items:\n self.entries.append(tuple([str(tag), entry]))\n self.dispatch_table[str(tag)] = func\n tag += 1\n\n def function(self):\n code, tag = self.d.menu(self.title, choices=self.entries)\n if code == self.d.OK: self._dispatch(tag)\n\n def _dispatch(self, tag):\n if tag in self.dispatch_table:\n func = self.dispatch_table[tag]\n if isinstance(func, MenuItem):\n func.run(ret=self.run)\n else: func()\n\n","subject":"Create object MenuItem that wraps functions to create a call stack","message":"Create object MenuItem that wraps functions to create a call stack\n","lang":"Python","license":"mit","repos":"jakogut\/KiWI"} {"commit":"ccf285c30a0110f2ff59b91ec0166f9b5306239d","old_file":"dukpy\/evaljs.py","new_file":"dukpy\/evaljs.py","old_contents":"import json\nfrom . import _dukpy\n\ntry:\n from collections.abc import Iterable\nexcept ImportError:\n from collections import Iterable\n\ntry: # pragma: no cover\n unicode\n string_types = (str, unicode)\nexcept NameError: # pragma: no cover\n string_types = (bytes, str)\n\n\nclass JSInterpreter(object):\n \"\"\"JavaScript Interpreter\"\"\"\n def __init__(self):\n self._ctx = _dukpy.create_context()\n\n def evaljs(self, code, **kwargs):\n \"\"\"Runs JavaScript code in the context of the interpreter.\n\n All arguments will be converted to plain javascript objects\n through the JSON encoder and will be available in `dukpy`\n global object.\n\n Returns the last object on javascript stack.\n \"\"\"\n jsvars = json.dumps(kwargs)\n jscode = code\n\n if not isinstance(code, string_types):\n jscode = ';\\n'.join(code)\n\n if not isinstance(jscode, bytes):\n jscode = jscode.encode('utf-8')\n\n res = _dukpy.eval_string(self._ctx, jscode, jsvars)\n if res is None:\n return None\n\n return json.loads(res.decode('utf-8'))\n\n\ndef evaljs(code, **kwargs):\n \"\"\"Evaluates the given ``code`` as JavaScript and returns the result\"\"\"\n return JSInterpreter().evaljs(code, **kwargs)","new_contents":"import json\nfrom . import _dukpy\n\ntry:\n from collections.abc import Iterable\nexcept ImportError:\n from collections import Iterable\n\ntry: # pragma: no cover\n unicode\n string_types = (str, unicode)\n jscode_type = str\nexcept NameError: # pragma: no cover\n string_types = (bytes, str)\n jscode_type = str\n\n\nclass JSInterpreter(object):\n \"\"\"JavaScript Interpreter\"\"\"\n def __init__(self):\n self._ctx = _dukpy.create_context()\n\n def evaljs(self, code, **kwargs):\n \"\"\"Runs JavaScript code in the context of the interpreter.\n\n All arguments will be converted to plain javascript objects\n through the JSON encoder and will be available in `dukpy`\n global object.\n\n Returns the last object on javascript stack.\n \"\"\"\n jsvars = json.dumps(kwargs)\n jscode = code\n\n if not isinstance(code, string_types):\n jscode = ';\\n'.join(code)\n\n if not isinstance(jscode, str):\n # Source code must be str on both Py2 and Py3\n # so it must be encoded on Py2 and decoded on Py3\n if isinstance(jscode, bytes):\n jscode = jscode.decode('utf-8')\n else:\n jscode = jscode.encode('utf-8')\n\n res = _dukpy.eval_string(self._ctx, jscode, jsvars)\n if res is None:\n return None\n\n return json.loads(res.decode('utf-8'))\n\n\ndef evaljs(code, **kwargs):\n \"\"\"Evaluates the given ``code`` as JavaScript and returns the result\"\"\"\n return JSInterpreter().evaljs(code, **kwargs)","subject":"Fix unicode source code on py3","message":"Fix unicode source code on py3\n","lang":"Python","license":"mit","repos":"amol-\/dukpy,amol-\/dukpy,amol-\/dukpy"} {"commit":"fdcdb5416bccf85a1745ccd07915e15629128ff9","old_file":"es_config.py","new_file":"es_config.py","old_contents":"\n# A list of ES hosts\nES_HOSTS = ['https:\/\/c3d581bfab179c1101d5b7a9e22a5f95.us-east-1.aws.found.io:9243']\nES_HTTP_AUTH = (\"elastic:u3Mk8jjADYJ4NzUmPTn15MNx\")\n","new_contents":"import os\nimport ast\n\n\n# A list of ES hosts\n# Uncomment the following for debugging\n# ES_HOSTS = ['https:\/\/c3d581bfab179c1101d5b7a9e22a5f95.us-east-1.aws.found.io:9243']\n# ES_HTTP_AUTH = (\"elastic:u3Mk8jjADYJ4NzUmPTn15MNx\")\n\n# Comment the following for debugging,\n# or set corresponding environment variables\ntry:\n ES_HOSTS = ast.literal_eval(os.environ['ES_HOSTS'])\n ES_HTTP_AUTH = ast.literal_eval(os.environ['ES_HTTP_AUTH'])\nexcept Exception as err:\n print(err)\n print(\"Please set ES_HOSTS and ES_HTTP_AUTH correctly.\")\n","subject":"Update ES server config to make use of environment variables","message":"Update ES server config to make use of environment variables\n","lang":"Python","license":"mit","repos":"justinchuby\/cmu-courseapi-flask"} {"commit":"08fe9e7beb4285feec9205012a62d464b3489bcf","old_file":"natasha\/grammars\/person\/interpretation.py","new_file":"natasha\/grammars\/person\/interpretation.py","old_contents":"from enum import Enum\nfrom collections import Counter\nfrom yargy.interpretation import InterpretationObject\n\n\nclass PersonObject(InterpretationObject):\n\n class Attributes(Enum):\n\n Firstname = 0 # владимир\n Middlename = 1 # владимирович\n Lastname = 2 # путин\n Descriptor = 3 # президент\n DescriptorDestination = 4 # российской федерации\n\n @property\n def gender(self):\n '''\n Very simple gender prediction algorithm\n '''\n counter = Counter()\n for field, token in self.__dict__.items():\n if not token:\n continue\n for form in token.forms:\n grammemes = set()\n if ('Ms-f' in form['grammemes']) or ('Fixd' in form['grammemes']):\n continue\n elif 'femn' in form['grammemes']:\n grammemes |= {'femn'}\n elif 'masc' in form['grammemes']:\n grammemes |= {'masc'}\n counter.update(grammemes)\n return counter\n","new_contents":"# coding: utf-8\nfrom __future__ import unicode_literals\n\nfrom enum import Enum\nfrom collections import Counter\nfrom yargy.interpretation import InterpretationObject\n\n\nclass PersonObject(InterpretationObject):\n\n class Attributes(Enum):\n\n Firstname = 0 # владимир\n Middlename = 1 # владимирович\n Lastname = 2 # путин\n Descriptor = 3 # президент\n DescriptorDestination = 4 # российской федерации\n\n @property\n def gender(self):\n '''\n Very simple gender prediction algorithm\n '''\n counter = Counter()\n for field, token in self.__dict__.items():\n if not token:\n continue\n for form in token.forms:\n grammemes = set()\n if ('Ms-f' in form['grammemes']) or ('Fixd' in form['grammemes']):\n continue\n elif 'femn' in form['grammemes']:\n grammemes |= {'femn'}\n elif 'masc' in form['grammemes']:\n grammemes |= {'masc'}\n counter.update(grammemes)\n return counter\n","subject":"Fix encoding for python 2.x","message":"Fix encoding for python 2.x\n","lang":"Python","license":"mit","repos":"natasha\/natasha"} {"commit":"294c251d83bec3738ce54a67d718c2ba959a7b4b","old_file":"git.py","new_file":"git.py","old_contents":"\nimport os\nimport subprocess\n\n\nclass cd:\n \"\"\"Context manager for changing the current working directory.\"\"\"\n def __init__(self, new_path):\n self.new_path = os.path.expanduser(new_path)\n\n def __enter__(self):\n self.previous_path = os.getcwd()\n os.chdir(self.new_path)\n\n def __exit__(self, etype, value, traceback):\n os.chdir(self.previous_path)\n\n\n# TODO use same context for all methods\nclass git:\n \"\"\"Minimal git wrapper, providing only funtions to init add and commit.\"\"\"\n def __init__(self, path):\n assert os.path.isabs(path)\n self.path = path\n\n def init(self):\n with cd(self.path):\n subprocess.call('git init')\n\n\n def add(self, filenames):\n for filename in filenames:\n assert os.path.isfile(filename) # needs entire path\n subprocess.call('git add ' + filename)\n\n def commit(self, date, message=''):\n subprocess.call(\n \"git commit -m '{m}' --date {d}\".format(m=message, d=date)\n )\n","new_contents":"\nimport os\nimport subprocess\nfrom contextlib import ContextDecorator\n\n\nclass cd(ContextDecorator):\n \"\"\"Context manager\/decorator for changing the current working directory.\"\"\"\n def __init__(self, new_path):\n self.new_path = os.path.expanduser(new_path)\n\n def __enter__(self):\n self.previous_path = os.getcwd()\n os.chdir(self.new_path)\n return self\n\n def __exit__(self, *exc):\n os.chdir(self.previous_path)\n return False\n\n\n# TODO use same context for all methods\nclass git:\n \"\"\"Minimal git wrapper, providing only funtions to init add and commit.\"\"\"\n path = ''\n def __init__(self, path):\n assert os.path.isabs(path)\n self.path = path\n\n @cd(path)\n def init(self):\n print('current dir: {}'.format(os.getcwd()))\n subprocess.call('git init')\n\n\n def add(self, filenames):\n for filename in filenames:\n assert os.path.isfile(filename) # needs entire path\n subprocess.call('git add ' + filename)\n\n def commit(self, date, message=''):\n subprocess.call(\n \"git commit -m '{m}' --date {d}\".format(m=message, d=date)\n )\n","subject":"Extend cd context manager to decorator.","message":"Extend cd context manager to decorator.\n","lang":"Python","license":"mit","repos":"0xfoo\/punchcard"} {"commit":"2e6efd4ecf22a1e7c673f90f113bfae47b98d294","old_file":"medical_insurance_us\/models\/__init__.py","new_file":"medical_insurance_us\/models\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n##############################################################################\n#\n# Author: Dave Lasley \n# Copyright: 2015 LasLabs, Inc [https:\/\/laslabs.com]\n#\n# This program is free software: you can redistribute it and\/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n##############################################################################\n\nfrom . import medical_insurance_plan\nfrom . import medical_insurance_company\n","new_contents":"# -*- coding: utf-8 -*-\n##############################################################################\n#\n# Author: Dave Lasley \n# Copyright: 2015 LasLabs, Inc [https:\/\/laslabs.com]\n#\n# This program is free software: you can redistribute it and\/or modify\n# it under the terms of the GNU Affero General Public License as\n# published by the Free Software Foundation, either version 3 of the\n# License, or (at your option) any later version.\n#\n# This program is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\n# GNU Affero General Public License for more details.\n#\n# You should have received a copy of the GNU Affero General Public License\n# along with this program. If not, see .\n#\n##############################################################################\n\nfrom . import medical_insurance_plan\nfrom . import medical_insurance_template\n","subject":"Add template and remove company from insurance_us imports","message":"Add template and remove company from insurance_us imports\n","lang":"Python","license":"agpl-3.0","repos":"ShaheenHossain\/eagle-medical,laslabs\/vertical-medical,ShaheenHossain\/eagle-medical,laslabs\/vertical-medical"} {"commit":"3d3a81efc36e39888929e62287b9d895922d8615","old_file":"tests\/sentry\/filters\/test_web_crawlers.py","new_file":"tests\/sentry\/filters\/test_web_crawlers.py","old_contents":"from __future__ import absolute_import\n\nfrom sentry.filters.web_crawlers import WebCrawlersFilter\nfrom sentry.testutils import TestCase\n\n\nclass WebCrawlersFilterTest(TestCase):\n filter_cls = WebCrawlersFilter\n\n def apply_filter(self, data):\n return self.filter_cls(self.project).test(data)\n\n def get_mock_data(self, user_agent):\n return {\n 'sentry.interfaces.Http': {\n 'url': 'http:\/\/example.com',\n 'method': 'GET',\n 'headers': [\n ['User-Agent', user_agent],\n ]\n }\n }\n\n def test_filters_googlebot(self):\n data = self.get_mock_data('Googlebot')\n assert self.apply_filter(data)\n\n def test_does_not_filter_chrome(self):\n data = self.get_mock_data('Mozilla\/5.0 (Windows NT 6.1) AppleWebKit\/537.36 (KHTML, like Gecko) Chrome\/41.0.2228.0 Safari\/537.36')\n assert not self.apply_filter(data)\n","new_contents":"from __future__ import absolute_import\n\nfrom sentry.filters.web_crawlers import WebCrawlersFilter\nfrom sentry.testutils import TestCase\n\n\nclass WebCrawlersFilterTest(TestCase):\n filter_cls = WebCrawlersFilter\n\n def apply_filter(self, data):\n return self.filter_cls(self.project).test(data)\n\n def get_mock_data(self, user_agent):\n return {\n 'sentry.interfaces.Http': {\n 'url': 'http:\/\/example.com',\n 'method': 'GET',\n 'headers': [\n ['User-Agent', user_agent],\n ]\n }\n }\n\n def test_filters_googlebot(self):\n data = self.get_mock_data('Googlebot')\n assert self.apply_filter(data)\n\n def test_does_not_filter_chrome(self):\n data = self.get_mock_data('Mozilla\/5.0 (Windows NT 6.1) AppleWebKit\/537.36 (KHTML, like Gecko) Chrome\/41.0.2228.0 Safari\/537.36')\n assert not self.apply_filter(data)\n\n def test_filters_twitterbot(self):\n data = self.get_mock_data('Twitterbot\/1.0')\n assert self.apply_filter(data)\n\n def test_filters_slack(self):\n data = self.get_mock_data('Slackbot-LinkExpanding 1.0 (+https:\/\/api.slack.com\/robots)')\n assert self.apply_filter(data)\n\n data = self.get_mock_data('Slack-ImgProxy 0.19 (+https:\/\/api.slack.com\/robots)')\n assert self.apply_filter(data)\n\n data = self.get_mock_data('Slackbot 1.0(+https:\/\/api.slack.com\/robots)')\n assert self.apply_filter(data)\n","subject":"Add unit tests for filtering Twitterbot and Slack.","message":"Add unit tests for filtering Twitterbot and Slack.\n","lang":"Python","license":"bsd-3-clause","repos":"ifduyue\/sentry,mvaled\/sentry,gencer\/sentry,gencer\/sentry,mvaled\/sentry,JackDanger\/sentry,beeftornado\/sentry,mvaled\/sentry,mvaled\/sentry,jean\/sentry,looker\/sentry,looker\/sentry,jean\/sentry,looker\/sentry,mvaled\/sentry,looker\/sentry,ifduyue\/sentry,ifduyue\/sentry,mvaled\/sentry,JackDanger\/sentry,gencer\/sentry,gencer\/sentry,jean\/sentry,jean\/sentry,JackDanger\/sentry,ifduyue\/sentry,beeftornado\/sentry,gencer\/sentry,beeftornado\/sentry,looker\/sentry,ifduyue\/sentry,jean\/sentry"} {"commit":"6bd088acd0ec0cfa5298051e286ce76e42430067","old_file":"shuup\/front\/themes\/views\/_product_preview.py","new_file":"shuup\/front\/themes\/views\/_product_preview.py","old_contents":"# -*- coding: utf-8 -*-\n# This file is part of Shuup.\n#\n# Copyright (c) 2012-2016, Shoop Ltd. All rights reserved.\n#\n# This source code is licensed under the AGPLv3 license found in the\n# LICENSE file in the root directory of this source tree.\nfrom shuup.front.views.product import ProductDetailView\n\n\nclass ProductPreviewView(ProductDetailView):\n template_name = \"shuup\/front\/product\/product_preview.jinja\"\n\n def get_context_data(self, **kwargs):\n # By default the template rendering the basket add form\n # uses the `request.path` as its' `next` value.\n # This is fine if you are on product page but here in\n # preview, we cannot redirect back to `\/xtheme\/product_preview`.\n\n context = super(ProductPreviewView, self).get_context_data(**kwargs)\n # Add `return_url` to context to avoid usage of `request.path` in\n # `classic_gray\/shuup\/front\/product\/_detail_order_section.jinja`\n context[\"return_url\"] = \"\/xtheme\/products\"\n return context\n\n\ndef product_preview(request):\n return ProductPreviewView.as_view()(request, pk=request.GET[\"id\"])\n","new_contents":"# -*- coding: utf-8 -*-\n# This file is part of Shuup.\n#\n# Copyright (c) 2012-2016, Shoop Ltd. All rights reserved.\n#\n# This source code is licensed under the AGPLv3 license found in the\n# LICENSE file in the root directory of this source tree.\nfrom shuup.front.views.product import ProductDetailView\n\n\nclass ProductPreviewView(ProductDetailView):\n template_name = \"shuup\/front\/product\/product_preview.jinja\"\n\n def get_context_data(self, **kwargs):\n # By default the template rendering the basket add form\n # uses the `request.path` as its' `next` value.\n # This is fine if you are on product page but here in\n # preview, we cannot redirect back to `\/xtheme\/product_preview`.\n\n context = super(ProductPreviewView, self).get_context_data(**kwargs)\n # Add `return_url` to context to avoid usage of `request.path`\n context[\"return_url\"] = \"\/xtheme\/products\"\n return context\n\n\ndef product_preview(request):\n return ProductPreviewView.as_view()(request, pk=request.GET[\"id\"])\n","subject":"Remove reference to nonexistent file","message":"Front: Remove reference to nonexistent file\n","lang":"Python","license":"agpl-3.0","repos":"shoopio\/shoop,shoopio\/shoop,shawnadelic\/shuup,suutari-ai\/shoop,shawnadelic\/shuup,hrayr-artunyan\/shuup,suutari\/shoop,suutari\/shoop,shoopio\/shoop,suutari-ai\/shoop,hrayr-artunyan\/shuup,shawnadelic\/shuup,suutari\/shoop,suutari-ai\/shoop,hrayr-artunyan\/shuup"} {"commit":"5fdedac2eae25d88d2595f0fe79ca3a332f24dfe","old_file":"pysuru\/api.py","new_file":"pysuru\/api.py","old_contents":"# coding: utf-8\n\"\"\"\nPublic endpoint to import API classes\n\nInstead of importing each module individually (eg.\n``from pysuru.apps import AppsAPI``), import from this module.\n\"\"\"\nfrom __future__ import absolute_imports\n\nfrom .apps import AppsAPI\nfrom .services import ServicesAPI\n","new_contents":"# coding: utf-8\n\"\"\"\nPublic endpoint to import API classes\n\nInstead of importing each module individually (eg.\n``from pysuru.apps import AppsAPI``), import from this module.\n\"\"\"\nfrom __future__ import absolute_imports\n\nfrom .apps import AppsAPI\nfrom .services import ServiceInstanceAPI\n","subject":"Fix service instance API class name","message":"Fix service instance API class name\n","lang":"Python","license":"mit","repos":"rcmachado\/pysuru"} {"commit":"a347c699be3ce5659db4b76a26ce253a209e232e","old_file":"webapp_health_monitor\/verificators\/base.py","new_file":"webapp_health_monitor\/verificators\/base.py","old_contents":"from webapp_health_monitor import errors\n\n\nclass Verificator(object):\n verificator_name = None\n\n def __init__(self, **kwargs):\n pass\n\n def run(self):\n raise NotImplementedError()\n\n def __str__(self):\n if self.verificator_name:\n return self.verificator_name\n else:\n return self.__class__.__name__\n\n\nclass RangeVerificator(Verificator):\n value_extractor = None\n upper_bound = None\n lower_bound = None\n\n def run(self):\n self._check_configuration()\n self._check_value()\n\n def _check_configuration(self):\n if not self._are_bounds_configured():\n raise errors.BadConfigurationError()\n\n def _are_bounds_configured(self):\n if self.lower_bound is None:\n return self.upper_bound is not None\n elif self.upper_bound is not None:\n return self.lower_bound <= self.upper_bound\n else:\n return True\n\n def _check_value(self):\n value = self.get_value()\n self._check_lower_bound(value)\n self._check_upper_bound(value)\n\n def get_value(self):\n raise NotImplementedError\n\n def _check_lower_bound(self, value):\n if self.lower_bound is not None:\n if value < self.lower_bound:\n raise errors.VerificationFailure()\n\n def _check_upper_bound(self, value):\n if self.upper_bound is not None:\n if value > self.upper_bound:\n raise errors.VerificationFailure()\n","new_contents":"from webapp_health_monitor import errors\n\n\nclass Verificator(object):\n verificator_name = None\n\n def __init__(self, **kwargs):\n pass\n\n def run(self):\n raise NotImplementedError()\n\n def __str__(self):\n if self.verificator_name:\n return self.verificator_name\n else:\n return self.__class__.__name__\n\n\nclass RangeVerificator(Verificator):\n upper_bound = None\n lower_bound = None\n\n def run(self):\n self._check_configuration()\n self._check_value()\n\n def _check_configuration(self):\n if not self._are_bounds_configured():\n raise errors.BadConfigurationError()\n\n def _are_bounds_configured(self):\n if self.lower_bound is None:\n return self.upper_bound is not None\n elif self.upper_bound is not None:\n return self.lower_bound <= self.upper_bound\n else:\n return True\n\n def _check_value(self):\n value = self.get_value()\n self._check_lower_bound(value)\n self._check_upper_bound(value)\n\n def get_value(self):\n raise NotImplementedError\n\n def _check_lower_bound(self, value):\n if self.lower_bound is not None:\n if value < self.lower_bound:\n raise errors.VerificationFailure()\n\n def _check_upper_bound(self, value):\n if self.upper_bound is not None:\n if value > self.upper_bound:\n raise errors.VerificationFailure()\n","subject":"Delete unused value extractor attribute.","message":"Delete unused value extractor attribute.\n","lang":"Python","license":"mit","repos":"pozytywnie\/webapp-health-monitor,serathius\/webapp-health-monitor"} {"commit":"47672fe44673fe9cae54a736bdc9eb496494ab58","old_file":"UI\/utilities\/synchronization_core.py","new_file":"UI\/utilities\/synchronization_core.py","old_contents":"# -*- coding: utf-8 -*-\n# Synchronization core module for Storj GUI Client #\n\n\nclass StorjFileSynchronization():\n def start_sync_thread(self):\n return 1\n\n def reload_sync_configuration(self):\n return 1\n\n def add_file_to_sync_queue(self):\n return 1\n","new_contents":"# -*- coding: utf-8 -*-\n# Synchronization core module for Storj GUI Client #\n\nimport time\nfrom watchdog.observers import Observer\nfrom watchdog.events import PatternMatchingEventHandler\nimport threading\n\nHANDLE_ON_MOVE_EVENT = True\nHANDLE_ON_DELETE_EVENT = True\n\nclass StorjFileSynchronization():\n\n def start_sync_thread(self):\n return 1\n\n def reload_sync_configuration(self):\n return 1\n\n def add_file_to_sync_queue(self, file_path):\n return 1\n\nclass FileChangesHandler(PatternMatchingEventHandler):\n #patterns = [\"*.xml\", \"*.lxml\"]\n\n def __init__(self):\n self.storj_file_synchronization_core = StorjFileSynchronization()\n\n\n def process(self, event):\n \"\"\"\n event.event_type\n 'modified' | 'created' | 'moved' | 'deleted'\n event.is_directory\n True | False\n event.src_path\n path\/to\/observed\/file\n \"\"\"\n # the file will be processed there\n self.storj_file_synchronization_core.add_file_to_sync_queue(file_path=str(event.src_path))\n print str(event)\n #print str(event.src_path) + str(event.event_type) + \"event\" # print now only for degug\n\n def on_deleted(self, event):\n if HANDLE_ON_DELETE_EVENT:\n self.process(event)\n\n def on_moved(self, event):\n if HANDLE_ON_MOVE_EVENT:\n self.process(event)\n\n def on_modified(self, event):\n self.process(event)\n\n def on_created(self, event):\n self.process(event)\n\nclass SyncObserverWorker():\n def start_observing_thread(self):\n observing_main_thread = threading.Thread(\n target=self.start_observing)\n observing_main_thread.start()\n\n def start_observing(self):\n paths_to_observe = []\n paths_to_observe.append(\"\/home\/lakewik\/storjsync\")\n self.observer = Observer()\n for path in paths_to_observe:\n self.observer.schedule(FileChangesHandler(), path=str(path))\n self.observer.start()\n print \"Synchronization directories observing started!\"\n\n def stop_observers(self):\n self.observer.stop()\n return 1\n #try:\n # while True:\n # time.sleep(1)\n #except KeyboardInterrupt:\n # observer.stop()\n\n #observer.join()\n\n","subject":"Add synchronization directory observer and handler","message":"Add synchronization directory observer and handler","lang":"Python","license":"mit","repos":"lakewik\/storj-gui-client"} {"commit":"23af33b7ca48c59ff58638b733437d8f348b279b","old_file":"openapi_core\/__init__.py","new_file":"openapi_core\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"OpenAPI core module\"\"\"\nfrom openapi_core.shortcuts import (\n create_spec, validate_parameters, validate_body, validate_data,\n)\n\n__author__ = 'Artur Maciąg'\n__email__ = 'maciag.artur@gmail.com'\n__version__ = '0.5.0'\n__url__ = 'https:\/\/github.com\/p1c2u\/openapi-core'\n__license__ = 'BSD 3-Clause License'\n\n__all__ = [\n 'create_spec', 'validate_parameters', 'validate_body', 'validate_data',\n]\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"OpenAPI core module\"\"\"\nfrom openapi_core.shortcuts import (\n create_spec, validate_parameters, validate_body, validate_data,\n)\n\n__author__ = 'Artur Maciag'\n__email__ = 'maciag.artur@gmail.com'\n__version__ = '0.5.0'\n__url__ = 'https:\/\/github.com\/p1c2u\/openapi-core'\n__license__ = 'BSD 3-Clause License'\n\n__all__ = [\n 'create_spec', 'validate_parameters', 'validate_body', 'validate_data',\n]\n","subject":"Replace unicode character for RPM build.","message":"Replace unicode character for RPM build.\n\nTo make building RPMs of package easier when using ascii by\ndefault.\n","lang":"Python","license":"bsd-3-clause","repos":"p1c2u\/openapi-core"} {"commit":"f4c9482e41ec2ee6c894a413e8fcb0349a9edbd1","old_file":"tapiriik\/web\/templatetags\/displayutils.py","new_file":"tapiriik\/web\/templatetags\/displayutils.py","old_contents":"from django import template\nimport json\nregister = template.Library()\n\n\n@register.filter(name=\"format_meters\")\ndef meters_to_kms(value):\n try:\n return round(value \/ 1000)\n except:\n return \"NaN\"\n\n@register.filter(name='json')\ndef jsonit(obj):\n return json.dumps(obj)\n\n@register.filter(name='dict_get')\ndef dict_get(tdict, key):\n if type(tdict) is not dict:\n tdict = tdict.__dict__\n return tdict.get(key, None)\n\n\n@register.filter(name='format')\ndef format(format, var):\n return format.format(var)\n\n@register.simple_tag\ndef stringformat(value, *args):\n return value.format(*args)\n\n@register.filter(name=\"percentage\")\ndef percentage(value, *args):\n if not value:\n return \"NaN\"\n return str(round(float(value) * 100)) + \"%\"\n","new_contents":"from django import template\nimport json\nregister = template.Library()\n\n\n@register.filter(name=\"format_meters\")\ndef meters_to_kms(value):\n try:\n return round(value \/ 1000)\n except:\n return \"NaN\"\n\n@register.filter(name='json')\ndef jsonit(obj):\n return json.dumps(obj)\n\n@register.filter(name='dict_get')\ndef dict_get(tdict, key):\n if type(tdict) is not dict:\n tdict = tdict.__dict__\n return tdict.get(key, None)\n\n\n@register.filter(name='format')\ndef format(format, var):\n return format.format(var)\n\n@register.simple_tag\ndef stringformat(value, *args):\n return value.format(*args)\n\n@register.filter(name=\"percentage\")\ndef percentage(value, *args):\n if not value:\n return \"NaN\"\n try:\n return str(round(float(value) * 100)) + \"%\"\n except ValueError:\n return value\n","subject":"Fix broken diagnostic dashboard with new sync progress values","message":"Fix broken diagnostic dashboard with new sync progress values\n","lang":"Python","license":"apache-2.0","repos":"campbellr\/tapiriik,niosus\/tapiriik,gavioto\/tapiriik,cheatos101\/tapiriik,cheatos101\/tapiriik,brunoflores\/tapiriik,abhijit86k\/tapiriik,mjnbike\/tapiriik,dlenski\/tapiriik,abhijit86k\/tapiriik,cpfair\/tapiriik,marxin\/tapiriik,abhijit86k\/tapiriik,dlenski\/tapiriik,cheatos101\/tapiriik,abs0\/tapiriik,niosus\/tapiriik,dmschreiber\/tapiriik,gavioto\/tapiriik,cmgrote\/tapiriik,campbellr\/tapiriik,abs0\/tapiriik,cheatos101\/tapiriik,brunoflores\/tapiriik,mjnbike\/tapiriik,cmgrote\/tapiriik,cpfair\/tapiriik,marxin\/tapiriik,dmschreiber\/tapiriik,mduggan\/tapiriik,cmgrote\/tapiriik,mduggan\/tapiriik,dmschreiber\/tapiriik,cgourlay\/tapiriik,brunoflores\/tapiriik,dlenski\/tapiriik,mjnbike\/tapiriik,cpfair\/tapiriik,abs0\/tapiriik,gavioto\/tapiriik,campbellr\/tapiriik,marxin\/tapiriik,campbellr\/tapiriik,dlenski\/tapiriik,mduggan\/tapiriik,gavioto\/tapiriik,brunoflores\/tapiriik,mduggan\/tapiriik,olamy\/tapiriik,niosus\/tapiriik,marxin\/tapiriik,cgourlay\/tapiriik,dmschreiber\/tapiriik,abhijit86k\/tapiriik,mjnbike\/tapiriik,cmgrote\/tapiriik,cgourlay\/tapiriik,cpfair\/tapiriik,niosus\/tapiriik,olamy\/tapiriik,cgourlay\/tapiriik,olamy\/tapiriik,olamy\/tapiriik,abs0\/tapiriik"} {"commit":"03eb0081a4037e36775271fb2373277f8e89835b","old_file":"src\/mcedit2\/resourceloader.py","new_file":"src\/mcedit2\/resourceloader.py","old_contents":"\"\"\"\n ${NAME}\n\"\"\"\nfrom __future__ import absolute_import, division, print_function, unicode_literals\nimport logging\nimport zipfile\n\nlog = logging.getLogger(__name__)\n\nclass ResourceNotFound(KeyError):\n pass\n\nclass ResourceLoader(object):\n def __init__(self):\n super(ResourceLoader, self).__init__()\n self.zipFiles = []\n\n def addZipFile(self, zipPath):\n try:\n zf = zipfile.ZipFile(zipPath)\n except zipfile.BadZipfile as e:\n raise IOError(\"Could not read %s as a zip file.\" % zipPath)\n self.zipFiles.append(zf)\n\n def openStream(self, path):\n for zipFile in self.zipFiles:\n try:\n stream = zipFile.open(path)\n break\n except KeyError: # Not found in zip file\n continue\n else:\n raise ResourceNotFound(\"Resource %s not found in search path\" % path)\n\n return stream\n","new_contents":"\"\"\"\n ${NAME}\n\"\"\"\nfrom __future__ import absolute_import, division, print_function, unicode_literals\nimport logging\nimport zipfile\n\nlog = logging.getLogger(__name__)\n\nclass ResourceNotFound(KeyError):\n pass\n\nclass ResourceLoader(object):\n def __init__(self):\n super(ResourceLoader, self).__init__()\n self.zipFiles = []\n\n def addZipFile(self, zipPath):\n try:\n zf = zipfile.ZipFile(zipPath)\n except zipfile.BadZipfile as e:\n raise IOError(\"Could not read %s as a zip file.\" % zipPath)\n self.zipFiles.append(zf)\n\n def openStream(self, path):\n for zipFile in self.zipFiles:\n try:\n stream = zipFile.open(path)\n break\n except KeyError: # Not found in zip file\n continue\n else:\n raise ResourceNotFound(\"Resource %s not found in search path\" % path)\n\n return stream\n\n def blockModelPaths(self):\n for zf in self.zipFiles:\n for name in zf.namelist():\n if name.startswith(\"assets\/minecraft\/models\/block\"):\n yield name\n","subject":"Add function to ResourceLoader for listing all block models","message":"Add function to ResourceLoader for listing all block models\n\nxxx only lists Vanilla models. haven't looked at mods with models yet.\n","lang":"Python","license":"bsd-3-clause","repos":"vorburger\/mcedit2,vorburger\/mcedit2,Rubisk\/mcedit2,Rubisk\/mcedit2"} {"commit":"de4f43613b5f3a8b6f49ace6b8e9585a242d7cb2","old_file":"src\/build.py","new_file":"src\/build.py","old_contents":"# Copyright 2007 Pompeu Fabra University (Computational Imaging Laboratory), Barcelona, Spain. Web: www.cilab.upf.edu.\n# This software is distributed WITHOUT ANY WARRANTY;\n# without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n# \n# Script to automate CSnake calls from cruise control\n\nimport sys\nimport csnGUIHandler\nimport csnGUIOptions\nimport csnGenerator\n\n# Check command line arguments\nif len(sys.argv) != 3:\n sys.exit(\"Error: not enough arguments. You need to provide an option and a configuration file.\")\n\n# Command line inputs\noptions_file = sys.argv[1]\nconfig_file = sys.argv[2]\n\n# Create GUI handler\nhandler = csnGUIHandler.Handler()\n# Read options\noptions = csnGUIOptions.Options()\noptions.Load( options_file )\n# Read settings\nsettings = csnGenerator.Settings()\nsettings.Load( config_file )\n# Set the options\nhandler.SetOptions( options )\n# Configure the project with the settings\nif settings.instance == \"thirdParty\":\n res = handler.ConfigureThirdPartyFolder(settings)\nelse:\n res = handler.ConfigureProjectToBinFolder( settings, 1 )\nsys.exit(res)\n\n","new_contents":"# Copyright 2007 Pompeu Fabra University (Computational Imaging Laboratory), Barcelona, Spain. Web: www.cilab.upf.edu.\n# This software is distributed WITHOUT ANY WARRANTY;\n# without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n# \n# Script to automate CSnake calls from cruise control\n\nimport sys\nimport csnGUIHandler\nimport csnGUIOptions\nimport csnGenerator\n\n# Check command line arguments\nif len(sys.argv) != 3:\n sys.exit(\"Error: not enough arguments. You need to provide an option and a configuration file.\")\n\n# Command line inputs\noptions_file = sys.argv[1]\nconfig_file = sys.argv[2]\n\n# Create GUI handler\nhandler = csnGUIHandler.Handler()\n# Read options\noptions = csnGUIOptions.Options()\noptions.Load( options_file )\n# Read settings\nsettings = csnGenerator.Settings()\nsettings.Load( config_file )\n# Set the options\nhandler.SetOptions( options )\n# Configure the project with the settings\nif settings.instance == \"thirdParty\":\n res = handler.ConfigureThirdPartyFolder(settings)\nelse:\n res = handler.ConfigureProjectToBinFolder( settings, 1 )\n# exit with error if there was a problem\nif res == false:\n sys.exit(1)\n\n","subject":"Exit with the proper value.","message":"Exit with the proper value.\n\ngit-svn-id: a26c1b3dc012bc7b166f1b96505d8277332098eb@265 9ffc3505-93cb-cd4b-9e5d-8a77f6415fcf\n","lang":"Python","license":"bsd-3-clause","repos":"csnake-org\/CSnake,csnake-org\/CSnake,msteghofer\/CSnake,msteghofer\/CSnake,csnake-org\/CSnake,msteghofer\/CSnake"} {"commit":"55464daa00ca68b07737433b0983df4667432a9c","old_file":"system\/plugins\/info.py","new_file":"system\/plugins\/info.py","old_contents":"__author__ = 'Gareth Coles'\n\nimport weakref\n\n\nclass Info(object):\n\n data = None\n\n core = None\n info = None\n\n def __init__(self, yaml_data, plugin_object=None):\n \"\"\"\n\n :param yaml_data:\n :type yaml_data: dict\n :return:\n \"\"\"\n\n self.data = yaml_data\n\n if plugin_object:\n self._plugin_object = weakref.ref(plugin_object)\n\n for key in yaml_data.keys():\n obj = yaml_data[key]\n\n if isinstance(obj, dict):\n setattr(self, key, Info(obj))\n else:\n setattr(self, key, obj)\n\n if self.core is not None:\n self.name = self.core.name\n self.module = self.core.module\n if hasattr(self.core, \"dependencies\"):\n self.dependencies = self.core.dependencies\n else:\n self.dependencies = []\n\n if self.info is not None:\n self.version = self.info.version\n self.description = self.info.description\n self.author = self.info.author\n self.website = self.info.website\n self.copyright = self.info.copyright\n\n @property\n def plugin_object(self):\n if hasattr(self, \"_plugin_object\"):\n return self._plugin_object()\n return None\n\n def get_module(self):\n if hasattr(self, \"module\"):\n return \"plugins.%s\" % self.module\n return None\n\n def set_plugin_object(self, obj):\n self._plugin_object = weakref.ref(obj)\n","new_contents":"__author__ = 'Gareth Coles'\n\nimport weakref\n\n\nclass Info(object):\n\n data = None\n\n core = None\n info = None\n\n def __init__(self, yaml_data, plugin_object=None):\n \"\"\"\n\n :param yaml_data:\n :type yaml_data: dict\n :return:\n \"\"\"\n\n self.data = yaml_data\n\n if plugin_object:\n self._plugin_object = weakref.ref(plugin_object)\n\n for key in yaml_data.keys():\n obj = yaml_data[key]\n\n if isinstance(obj, dict):\n setattr(self, key, Info(obj))\n else:\n setattr(self, key, obj)\n\n if self.core is not None:\n self.name = self.core.name\n self.module = self.core.module\n if hasattr(self.core, \"dependencies\"):\n self.dependencies = self.core.dependencies\n else:\n self.core.dependencies = []\n self.dependencies = []\n\n if self.info is not None:\n self.version = self.info.version\n self.description = self.info.description\n self.author = self.info.author\n self.website = self.info.website\n self.copyright = self.info.copyright\n\n @property\n def plugin_object(self):\n if hasattr(self, \"_plugin_object\"):\n return self._plugin_object()\n return None\n\n def get_module(self):\n if hasattr(self, \"module\"):\n return \"plugins.%s\" % self.module\n return None\n\n def set_plugin_object(self, obj):\n self._plugin_object = weakref.ref(obj)\n","subject":"Fix missing dependencies on core","message":"Fix missing dependencies on core\n","lang":"Python","license":"artistic-2.0","repos":"UltrosBot\/Ultros,UltrosBot\/Ultros"} {"commit":"985dd1ada1b2ad9ceaae111fa32b1d8e54b61786","old_file":"mailqueue\/tasks.py","new_file":"mailqueue\/tasks.py","old_contents":"from celery.task import task\nfrom .models import MailerMessage\n\n@task(name=\"tasks.send_mail\")\ndef send_mail(pk):\n message = MailerMessage.objects.get(pk=pk)\n message._send()\n\n@task()\ndef clear_sent_messages():\n from mailqueue.models import MailerMessage\n MailerMessage.objects.clear_sent_messages()\n","new_contents":"from celery.task import task\nfrom .models import MailerMessage\n\n@task(name=\"tasks.send_mail\", default_retry_delay=5, max_retries=5)\ndef send_mail(pk):\n message = MailerMessage.objects.get(pk=pk)\n message._send()\n \n # Retry when message is not sent\n if not message.sent:\n send_mail.retry([message.pk,])\n\n@task()\ndef clear_sent_messages():\n from mailqueue.models import MailerMessage\n MailerMessage.objects.clear_sent_messages()\n","subject":"Add retry to celery task","message":"Add retry to celery task\n\nMessages do not always get delivered. Built in a retry when message is not sent.\r\nMax retry count could also be a setting.","lang":"Python","license":"mit","repos":"Goury\/django-mail-queue,dstegelman\/django-mail-queue,winfieldco\/django-mail-queue,Goury\/django-mail-queue,styrmis\/django-mail-queue,dstegelman\/django-mail-queue"} {"commit":"8d8798554d996776eecc61b673adcbc2680f327a","old_file":"mastermind\/main.py","new_file":"mastermind\/main.py","old_contents":"from __future__ import (absolute_import, print_function, division)\nfrom itertools import repeat\nfrom mitmproxy.main import mitmdump\nimport os\n\nfrom . import (cli, proxyswitch, say)\n\ndef main():\n parser = cli.args()\n args, extra_args = parser.parse_known_args()\n\n try:\n config = cli.config(args)\n except IOError as err:\n parser.error(err)\n except toml.core.TomlError as err:\n parser.error(\"Errors found in the config file:\\n\\n\", err)\n\n mitm_args = cli.mitm_args(config)\n is_sudo = os.getuid() == 0\n\n if type(mitm_args) == Exception:\n parser.error(mitm_args.message)\n\n say.level(config[\"core\"][\"verbose\"])\n\n try:\n if config[\"os\"][\"proxy-settings\"]:\n if not is_sudo:\n parser.error(\"proxy-settings is enabled, please provide sudo in order to change the OSX proxy configuration.\")\n\n proxyswitch.enable(config[\"core\"][\"host\"],\n str(config[\"core\"][\"port\"]))\n\n mitmdump(mitm_args + extra_args)\n finally:\n if config[\"os\"][\"proxy-settings\"] and is_sudo:\n proxyswitch.disable()\n","new_contents":"from __future__ import (absolute_import, print_function, division)\nfrom itertools import repeat\nfrom mitmproxy.main import mitmdump\nimport os\n\nfrom . import (cli, proxyswitch, say)\n\ndef main():\n parser = cli.args()\n args, extra_args = parser.parse_known_args()\n\n try:\n config = cli.config(args)\n except IOError as err:\n parser.error(err)\n except toml.core.TomlError as err:\n parser.error(\"Errors found in the config file:\\n\\n\", err)\n\n mitm_args = cli.mitm_args(config)\n is_sudo = os.getuid() == 0\n\n if type(mitm_args) == Exception:\n parser.error(mitm_args.message)\n\n say.level(config[\"core\"][\"verbose\"])\n\n host= config[\"core\"][\"host\"]\n port = config[\"core\"][\"port\"]\n pid_filename = \"\/var\/tmp\/mastermind.{}{}.pid\".format(host.replace('.', ''), port)\n\n try:\n if config[\"os\"][\"proxy-settings\"]:\n if not is_sudo:\n parser.error(\"proxy-settings is enabled, please provide sudo in order to change the OSX proxy configuration.\")\n\n proxyswitch.enable(host, str(port))\n\n with open(pid_filename, \"w\") as f:\n f.write(str(os.getpid()))\n\n mitmdump(mitm_args + extra_args)\n finally:\n os.remove(pid_filename)\n\n if config[\"os\"][\"proxy-settings\"] and is_sudo:\n proxyswitch.disable()\n","subject":"Write PID to a file","message":"Write PID to a file\n","lang":"Python","license":"mit","repos":"ustwo\/mastermind,ustwo\/mastermind"} {"commit":"4ef5d9ae7a571f97242cf2cc44e539d039486549","old_file":"runserver.py","new_file":"runserver.py","old_contents":"#!\/usr\/bin\/env python\n\"\"\"\n # # #### ##### # # ##### # # # #\n # # # # # ## ## # # # ## # #\n ### #### #### # # # # # # # # #####\n # # # # # # # # ## # # #\n # # # ##### # # # # # # # #\n\n Kremlin Magical Everything System\n Glasnost Image Board and Boredom Inhibitor\n\n\"\"\"\n\nfrom kremlin import app\n\ndef main():\n print \"Launching kremlin in development mode.\"\n print \"--------------------------------------\"\n app.run(debug=True)\n\nif __name__ == '__main__':\n main()\n","new_contents":"#!\/usr\/bin\/env python\n\"\"\"\n # # #### ##### # # ##### # # # #\n # # # # # ## ## # # # ## # #\n ### #### #### # # # # # # # # #####\n # # # # # # # # ## # # #\n # # # ##### # # # # # # # #\n\n Kremlin Magical Everything System\n Glasnost Image Board and Boredom Inhibitor\n\n\"\"\"\n\nfrom kremlin import app\n\ndef main():\n print \"Kremlin Magical Everything System v 0.0.0-None\"\n print \"Copyright (c) Glasnost 2010-2011\"\n print \"-----------------------------------------------\"\n print \"RUNNING IN DEVELOPMENT MODE! ** NOT FOR PRODUCTION **\"\n print \"Connect to http:\/\/127.0.0.1:5000 to access.\"\n app.run(debug=True)\n\nif __name__ == '__main__':\n main()\n","subject":"Add proper banner to dev launch script","message":"Add proper banner to dev launch script\n","lang":"Python","license":"bsd-2-clause","repos":"glasnost\/kremlin,glasnost\/kremlin,glasnost\/kremlin"} {"commit":"928cbc47cb7430d8a2fef924b61179cd30f5ca34","old_file":"linter.py","new_file":"linter.py","old_contents":"#\n# linter.py\n# Linter for SublimeLinter3, a code checking framework for Sublime Text 3\n#\n# Written by Ethan Zimmerman\n# Copyright (c) 2014 Ethan Zimmerman\n#\n# License: MIT\n#\n\n\"\"\"This module exports the RamlCop plugin class.\"\"\"\n\nfrom SublimeLinter.lint import NodeLinter\n\n\nclass RamlCop(NodeLinter):\n\n \"\"\"Provides an interface to raml-cop.\"\"\"\n\n syntax = 'raml'\n cmd = 'raml-cop --no-color'\n version_requirement = '>= 1.0.0'\n regex = (\n r'^\\[.+:(?P\\d+):(?P\\d+)\\] '\n r'(?P.+)'\n )\n line_col_base = (0, 0)\n tempfile_suffix = '-'\n","new_contents":"#\n# linter.py\n# Linter for SublimeLinter3, a code checking framework for Sublime Text 3\n#\n# Written by Ethan Zimmerman\n# Copyright (c) 2014 Ethan Zimmerman\n#\n# License: MIT\n#\n\n\"\"\"This module exports the RamlCop plugin class.\"\"\"\n\nfrom SublimeLinter.lint import NodeLinter\n\n\nclass RamlCop(NodeLinter):\n \"\"\"Provides an interface to raml-cop.\"\"\"\n\n syntax = 'raml'\n cmd = 'raml-cop --no-color'\n version_requirement = '>= 1.0.0'\n regex = (\n r'^\\[.+:(?P\\d+):(?P\\d+)\\] '\n r'(?P.+)'\n )\n line_col_base = (0, 0)\n tempfile_suffix = '-'\n","subject":"Remove empty line before class docstring","message":"Remove empty line before class docstring\n","lang":"Python","license":"mit","repos":"thebinarypenguin\/SublimeLinter-contrib-raml-cop"} {"commit":"9ee0f3f7be90046f796f3395b2149288a2b52a26","old_file":"src\/zeit\/magazin\/preview.py","new_file":"src\/zeit\/magazin\/preview.py","old_contents":"# Copyright (c) 2013 gocept gmbh & co. kg\n# See also LICENSE.txt\n\nimport grokcore.component as grok\nimport zeit.cms.browser.preview\nimport zeit.magazin.interfaces\n\n\n@grok.adapter(zeit.magazin.interfaces.IZMOContent, basestring)\n@grok.implementer(zeit.cms.browser.interfaces.IPreviewURL)\ndef preview_url(content, preview_type):\n return zeit.cms.browser.preview.prefixed_url(\n 'zmo-%s-prefix' % preview_type, content.uniqueId)\n\n\n# XXX there also is a (basestring, basestring) variant of the adapter\n# which is used by zeit.find to caluclate preview-urls for search results\n# without looking up the content object first. What do we do about that?\n","new_contents":"# Copyright (c) 2013 gocept gmbh & co. kg\n# See also LICENSE.txt\n\nimport grokcore.component as grok\nimport zeit.cms.browser.preview\nimport zeit.magazin.interfaces\n\n\n@grok.adapter(zeit.magazin.interfaces.IZMOContent, basestring)\n@grok.implementer(zeit.cms.browser.interfaces.IPreviewURL)\ndef preview_url(content, preview_type):\n return zeit.cms.browser.preview.prefixed_url(\n 'zmo-%s-prefix' % preview_type, content.uniqueId)\n","subject":"Remove oddity marker, it's been resolved in zeit.find now","message":"Remove oddity marker, it's been resolved in zeit.find now\n","lang":"Python","license":"bsd-3-clause","repos":"ZeitOnline\/zeit.magazin"} {"commit":"ee0f28abd70396bf1e094592028aa693e5d6fe6c","old_file":"rechunker\/executors\/python.py","new_file":"rechunker\/executors\/python.py","old_contents":"import itertools\nfrom functools import partial\nimport math\n\nfrom typing import Any, Callable, Iterable\n\nfrom rechunker.types import CopySpec, StagedCopySpec, Executor\n\n\nThunk = Callable[[], None]\n\n\nclass PythonExecutor(Executor[Thunk]):\n \"\"\"An execution engine based on Python loops.\n\n Supports copying between any arrays that implement ``__getitem__`` and\n ``__setitem__`` for tuples of ``slice`` objects.\n\n Execution plans for PythonExecutor are functions that accept no arguments.\n \"\"\"\n\n def prepare_plan(self, specs: Iterable[StagedCopySpec]) -> Thunk:\n tasks = []\n for staged_copy_spec in specs:\n for copy_spec in staged_copy_spec.stages:\n tasks.append(partial(_direct_copy_array, copy_spec))\n return partial(_execute_all, tasks)\n\n def execute_plan(self, plan: Thunk):\n plan()\n\n\ndef _direct_copy_array(copy_spec: CopySpec) -> None:\n \"\"\"Direct copy between zarr arrays.\"\"\"\n source_array, target_array, chunks = copy_spec\n shape = source_array.shape\n ranges = [range(math.ceil(s \/ c)) for s, c in zip(shape, chunks)]\n for indices in itertools.product(*ranges):\n key = tuple(slice(c * i, c * (i + 1)) for i, c in zip(indices, chunks))\n target_array[key] = source_array[key]\n\n\ndef _execute_all(tasks: Iterable[Callable[[], Any]]) -> None:\n for task in tasks:\n task()\n","new_contents":"import itertools\nfrom functools import partial\nimport math\n\nfrom typing import Callable, Iterable\n\nfrom rechunker.types import CopySpec, StagedCopySpec, Executor\n\n\n# PythonExecutor represents delayed execution tasks as functions that require\n# no arguments.\nTask = Callable[[], None]\n\n\nclass PythonExecutor(Executor[Task]):\n \"\"\"An execution engine based on Python loops.\n\n Supports copying between any arrays that implement ``__getitem__`` and\n ``__setitem__`` for tuples of ``slice`` objects.\n\n Execution plans for PythonExecutor are functions that accept no arguments.\n \"\"\"\n\n def prepare_plan(self, specs: Iterable[StagedCopySpec]) -> Task:\n tasks = []\n for staged_copy_spec in specs:\n for copy_spec in staged_copy_spec.stages:\n tasks.append(partial(_direct_copy_array, copy_spec))\n return partial(_execute_all, tasks)\n\n def execute_plan(self, plan: Task):\n plan()\n\n\ndef _direct_copy_array(copy_spec: CopySpec) -> None:\n \"\"\"Direct copy between zarr arrays.\"\"\"\n source_array, target_array, chunks = copy_spec\n shape = source_array.shape\n ranges = [range(math.ceil(s \/ c)) for s, c in zip(shape, chunks)]\n for indices in itertools.product(*ranges):\n key = tuple(slice(c * i, c * (i + 1)) for i, c in zip(indices, chunks))\n target_array[key] = source_array[key]\n\n\ndef _execute_all(tasks: Iterable[Task]) -> None:\n for task in tasks:\n task()\n","subject":"Remove 'thunk' jargon from PythonExecutor","message":"Remove 'thunk' jargon from PythonExecutor\n","lang":"Python","license":"mit","repos":"pangeo-data\/rechunker"} {"commit":"01dc78bc4cea6c11744879c0f2066ab627314625","old_file":"django_stackoverflow_trace\/__init__.py","new_file":"django_stackoverflow_trace\/__init__.py","old_contents":"from django.views import debug\n\n\ndef _patch_django_debug_view():\n\n new_data = \"\"\"\n

\n View in Stackoverflow<\/a>\n <\/h3>\n \"\"\"\n\n replace_point = ''\n replacement = new_data + replace_point\n\n # monkey patch the built-in template.\n debug.TECHNICAL_500_TEMPLATE = debug.TECHNICAL_500_TEMPLATE.replace(\n replace_point,\n replacement,\n 1 # replace the first occurence\n )\n\n\nclass DjangoStackoverTraceMiddleware(object):\n\n def __init__(self):\n _patch_django_debug_view()\n\n def process_response(self, request, response):\n return response\n","new_contents":"from django.views import debug\nfrom django.conf import settings\n\n\ndef get_search_link():\n default_choice = \"stackoverflow\"\n\n search_urls = {\n \"stackoverflow\": \"http:\/\/stackoverflow.com\/search?q=[python] or \"\n \"[django]+{{ exception_value|force_escape }}\",\n \"googlesearch\": \"https:\/\/www.google.com.tr\/#q=site:stackoverflow.com\"\n \"+django+{{ exception_value|force_escape }}\"\n }\n\n search_url = getattr(\n settings,\n 'DJANGO_STACKOVERFLOW_TRACE_SEARCH_SITE',\n default_choice\n )\n\n return search_urls.get(search_url, search_urls[default_choice])\n\n\ndef _patch_django_debug_view():\n\n new_data = \"\"\"\n

\n View in Stackoverflow<\/a>\n <\/h3>\n \"\"\" % get_search_link()\n\n replace_point = '

'\n replacement = new_data + replace_point\n\n # monkey patch the built-in template.\n debug.TECHNICAL_500_TEMPLATE = debug.TECHNICAL_500_TEMPLATE.replace(\n replace_point,\n replacement,\n 1 # replace the first occurence\n )\n\n\nclass DjangoStackoverTraceMiddleware(object):\n\n def __init__(self):\n _patch_django_debug_view()\n\n def process_response(self, request, response):\n return response\n","subject":"Add a google search option","message":"Add a google search option\n","lang":"Python","license":"mit","repos":"emre\/django-stackoverflow-trace"} {"commit":"ddc03637b19059f6fb06d72dc380afaf4fba57c2","old_file":"indra\/tests\/test_context.py","new_file":"indra\/tests\/test_context.py","old_contents":"from indra.databases import context_client\n\ndef test_get_protein_expression():\n res = context_client.get_protein_expression('EGFR', 'BT20_BREAST')\n assert(res is not None)\n assert(res.get('EGFR') is not None)\n assert(res['EGFR'].get('BT20_BREAST') is not None)\n assert(res['EGFR']['BT20_BREAST'] > 1000)\n\ndef test_get_mutations():\n res = context_client.get_mutations('BRAF', 'A375_SKIN')\n assert(res is not None)\n assert(res.get('BRAF') is not None)\n assert(res['BRAF'].get('A375_SKIN') is not None)\n assert(res['BRAF']['A375_SKIN'] == 1.0)\n\ndef test_send_request_bad():\n res = context_client.send_request('xyz', None)\n assert(res is None)\n","new_contents":"from indra.databases import context_client\n\ndef test_get_protein_expression():\n res = context_client.get_protein_expression('EGFR', 'BT20_BREAST')\n assert(res is not None)\n assert(res.get('EGFR') is not None)\n assert(res['EGFR'].get('BT20_BREAST') is not None)\n assert(res['EGFR']['BT20_BREAST'] > 1000)\n\ndef test_get_mutations():\n res = context_client.get_mutations('BRAF', 'A375_SKIN')\n assert(res is not None)\n assert(res.get('BRAF') is not None)\n assert(res['BRAF'].get('A375_SKIN') is not None)\n assert(res['BRAF']['A375_SKIN'] == 1.0)\n","subject":"Remove deprecated context client test","message":"Remove deprecated context client test\n","lang":"Python","license":"bsd-2-clause","repos":"johnbachman\/belpy,sorgerlab\/indra,johnbachman\/belpy,sorgerlab\/indra,pvtodorov\/indra,bgyori\/indra,sorgerlab\/indra,sorgerlab\/belpy,johnbachman\/indra,jmuhlich\/indra,jmuhlich\/indra,johnbachman\/belpy,sorgerlab\/belpy,pvtodorov\/indra,pvtodorov\/indra,jmuhlich\/indra,bgyori\/indra,johnbachman\/indra,sorgerlab\/belpy,johnbachman\/indra,pvtodorov\/indra,bgyori\/indra"} {"commit":"a1f5a392d5270dd6f80a40e45c5e25b6ae04b7c3","old_file":"embed_video\/fields.py","new_file":"embed_video\/fields.py","old_contents":"from django.db import models\nfrom django import forms\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom .backends import detect_backend, UnknownIdException, \\\n UnknownBackendException\n\n__all__ = ('EmbedVideoField', 'EmbedVideoFormField')\n\n\nclass EmbedVideoField(models.URLField):\n \"\"\"\n Model field for embeded video. Descendant of\n :py:class:`django.db.models.URLField`.\n \"\"\"\n\n def formfield(self, **kwargs):\n defaults = {'form_class': EmbedVideoFormField}\n defaults.update(kwargs)\n return super(EmbedVideoField, self).formfield(**defaults)\n\n def south_field_triple(self):\n from south.modelsinspector import introspector\n cls_name = '%s.%s' % (\n self.__class__.__module__,\n self.__class__.__name__\n )\n args, kwargs = introspector(self)\n return (cls_name, args, kwargs)\n\n\nclass EmbedVideoFormField(forms.URLField):\n \"\"\"\n Form field for embeded video. Descendant of\n :py:class:`django.forms.URLField`\n \"\"\"\n\n def validate(self, url):\n super(EmbedVideoFormField, self).validate(url)\n\n if url:\n try:\n detect_backend(url)\n except UnknownBackendException:\n raise forms.ValidationError(_(u'URL could not be recognized.'))\n except UnknownIdException:\n raise forms.ValidationError(_(u'ID of this video could not be \\\n recognized.'))\n\n return url\n","new_contents":"from django.db import models\nfrom django import forms\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom .backends import detect_backend, UnknownIdException, \\\n UnknownBackendException\n\n__all__ = ('EmbedVideoField', 'EmbedVideoFormField')\n\n\nclass EmbedVideoField(models.URLField):\n \"\"\"\n Model field for embeded video. Descendant of\n :py:class:`django.db.models.URLField`.\n \"\"\"\n\n def formfield(self, **kwargs):\n defaults = {'form_class': EmbedVideoFormField}\n defaults.update(kwargs)\n return super(EmbedVideoField, self).formfield(**defaults)\n\n def south_field_triple(self):\n from south.modelsinspector import introspector\n cls_name = '%s.%s' % (\n self.__class__.__module__,\n self.__class__.__name__\n )\n args, kwargs = introspector(self)\n return (cls_name, args, kwargs)\n\n\nclass EmbedVideoFormField(forms.URLField):\n \"\"\"\n Form field for embeded video. Descendant of\n :py:class:`django.forms.URLField`\n \"\"\"\n\n def validate(self, url):\n # if empty url is not allowed throws an exception\n super(EmbedVideoFormField, self).validate(url)\n \n if not url:\n return\n\n try:\n detect_backend(url)\n except UnknownBackendException:\n raise forms.ValidationError(_(u'URL could not be recognized.'))\n except UnknownIdException:\n raise forms.ValidationError(_(u'ID of this video could not be \\\n recognized.'))\n return url\n","subject":"Simplify validate method in FormField.","message":"Simplify validate method in FormField.","lang":"Python","license":"mit","repos":"yetty\/django-embed-video,jazzband\/django-embed-video,jazzband\/django-embed-video,mpachas\/django-embed-video,yetty\/django-embed-video,mpachas\/django-embed-video"} {"commit":"7f6167ef9f62b9b79e3c30b358c796caae69a2e6","old_file":"PyWXSB\/exceptions_.py","new_file":"PyWXSB\/exceptions_.py","old_contents":"\"\"\"Extensions of standard exceptions for PyWXSB events.\n\nYeah, I'd love this module to be named exceptions.py, but it can't\nbecause the standard library has one of those, and we need to\nreference it below.\n\"\"\"\n\nimport exceptions\n\nclass PyWXSBException (exceptions.Exception):\n \"\"\"Base class for exceptions that indicate a problem that the user should fix.\"\"\"\n pass\n\nclass SchemaValidationError (PyWXSBException):\n \"\"\"Raised when the XML hierarchy does not appear to be valid for an XML schema.\"\"\"\n pass\n\nclass BadTypeValueError (PyWXSBException):\n \"\"\"Raised when a value in an XML attribute does not conform to the simple type.\"\"\"\n pass\n\nclass NotInNamespaceError (PyWXSBException):\n '''Raised when a name is referenced that is not defined in the appropriate namespace.'''\n __namespace = None\n __ncName = None\n\nclass BadPropertyError (PyWXSBException):\n \"\"\"Raised when a schema component property is accessed on a component instance that does not define that property.\"\"\"\n pass\n\nclass PyWXSBError (exceptions.Exception):\n \"\"\"Base class for exceptions that indicate a problem that the user probably can't fix.\"\"\"\n pass\n \nclass LogicError (PyWXSBError):\n \"\"\"Raised when the code detects an implementation problem.\"\"\"\n\nclass IncompleteImplementationError (LogicError):\n \"\"\"Raised when a code branch is taken that has not yet been implemented.\"\"\"\n","new_contents":"\"\"\"Extensions of standard exceptions for PyWXSB events.\n\nYeah, I'd love this module to be named exceptions.py, but it can't\nbecause the standard library has one of those, and we need to\nreference it below.\n\"\"\"\n\nimport exceptions\n\nclass PyWXSBException (exceptions.Exception):\n \"\"\"Base class for exceptions that indicate a problem that the user should fix.\"\"\"\n pass\n\nclass SchemaValidationError (PyWXSBException):\n \"\"\"Raised when the XML hierarchy does not appear to be valid for an XML schema.\"\"\"\n pass\n\nclass BadTypeValueError (PyWXSBException):\n \"\"\"Raised when a value in an XML attribute does not conform to the simple type.\"\"\"\n pass\n\nclass NotInNamespaceError (PyWXSBException):\n '''Raised when a name is referenced that is not defined in the appropriate namespace.'''\n __namespace = None\n __ncName = None\n\nclass BadPropertyError (PyWXSBException):\n \"\"\"Raised when a schema component property is accessed on a component instance that does not define that property.\"\"\"\n pass\n\nclass BadDocumentError (PyWXSBException):\n \"\"\"Raised when processing document content and an error is encountered.\"\"\"\n pass\n\nclass PyWXSBError (exceptions.Exception):\n \"\"\"Base class for exceptions that indicate a problem that the user probably can't fix.\"\"\"\n pass\n \nclass LogicError (PyWXSBError):\n \"\"\"Raised when the code detects an implementation problem.\"\"\"\n\nclass IncompleteImplementationError (LogicError):\n \"\"\"Raised when a code branch is taken that has not yet been implemented.\"\"\"\n","subject":"Add an exception to throw when a document does have the expected structure","message":"Add an exception to throw when a document does have the expected structure\n","lang":"Python","license":"apache-2.0","repos":"jonfoster\/pyxb-upstream-mirror,jonfoster\/pyxb-upstream-mirror,balanced\/PyXB,jonfoster\/pyxb2,pabigot\/pyxb,CantemoInternal\/pyxb,jonfoster\/pyxb2,balanced\/PyXB,CantemoInternal\/pyxb,jonfoster\/pyxb1,pabigot\/pyxb,jonfoster\/pyxb2,CantemoInternal\/pyxb,jonfoster\/pyxb-upstream-mirror,jonfoster\/pyxb1,balanced\/PyXB"} {"commit":"3bb0e65eac5c93fa6e331d22252fd7b17ecdf964","old_file":"__main__.py","new_file":"__main__.py","old_contents":"# Copyright 2016 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport sys\n\nfrom tensorflow.tensorboard.tensorboard import main\n\nif __name__ == '__main__':\n sys.exit(main())\n","new_contents":"# Copyright 2016 The TensorFlow Authors. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n# ==============================================================================\n\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport sys\n\nfrom tensorflow.tensorboard.tensorboard import main\n\nif __name__ == '__main__':\n sys.exit(main())\n","subject":"Add pylint indentation check to sanity and fix existing indentation Change: 132840696","message":"Add pylint indentation check to sanity and fix existing indentation\nChange: 132840696\n","lang":"Python","license":"apache-2.0","repos":"francoisluus\/tensorboard-supervise,qiuminxu\/tensorboard,qiuminxu\/tensorboard,tensorflow\/tensorboard,shakedel\/tensorboard,qiuminxu\/tensorboard,ioeric\/tensorboard,shakedel\/tensorboard,qiuminxu\/tensorboard,tensorflow\/tensorboard,qiuminxu\/tensorboard,francoisluus\/tensorboard-supervise,agrubb\/tensorboard,tensorflow\/tensorboard,shakedel\/tensorboard,ioeric\/tensorboard,tensorflow\/tensorboard,tensorflow\/tensorboard,shakedel\/tensorboard,francoisluus\/tensorboard-supervise,agrubb\/tensorboard,francoisluus\/tensorboard-supervise,qiuminxu\/tensorboard,tensorflow\/tensorboard,ioeric\/tensorboard,ioeric\/tensorboard,agrubb\/tensorboard,ioeric\/tensorboard,ioeric\/tensorboard,tensorflow\/tensorboard,shakedel\/tensorboard,agrubb\/tensorboard,agrubb\/tensorboard,francoisluus\/tensorboard-supervise,francoisluus\/tensorboard-supervise,shakedel\/tensorboard,agrubb\/tensorboard"} {"commit":"36298a9f9a7a373a716e44cac6226a0ec8c8c40c","old_file":"__main__.py","new_file":"__main__.py","old_contents":"from twisted.internet.endpoints import TCP4ServerEndpoint\nfrom twisted.internet import reactor\n\nimport editorFactory\n\nif __name__ == \"__main__\":\n server = editorFactory.EditorFactory()\n TCP4ServerEndpoint(reactor, 4567).listen(server)\n reactor.run()\n","new_contents":"from twisted.internet.endpoints import TCP4ServerEndpoint\nfrom twisted.internet import reactor\n\nimport editorFactory\n\nif __name__ == \"__main__\":\n server = editorFactory.EditorFactory()\n TCP4ServerEndpoint(reactor, 4567).listen(server)\n print('Starting up...')\n reactor.run()\n","subject":"Print something at the start","message":"Print something at the start\n","lang":"Python","license":"apache-2.0","repos":"Floobits\/floobits-emacs"} {"commit":"602c01caa23df0c6dad5963412a340087012f692","old_file":"thinc\/tests\/integration\/test_shape_check.py","new_file":"thinc\/tests\/integration\/test_shape_check.py","old_contents":"import pytest\nimport numpy\n\nfrom ...neural._classes.model import Model\n\n\n\ndef test_mismatched_shapes_raise_ShapeError():\n X = numpy.ones((3, 4))\n model = Model(10, 5)\n with pytest.raises(ValueError):\n y = model.begin_training(X)\n \n","new_contents":"import pytest\nimport numpy\n\nfrom ...neural._classes.model import Model\nfrom ...exceptions import UndefinedOperatorError, DifferentLengthError\nfrom ...exceptions import ExpectedTypeError, ShapeMismatchError\n\n\ndef test_mismatched_shapes_raise_ShapeError():\n X = numpy.ones((3, 4))\n model = Model(10, 5)\n with pytest.raises(ShapeMismatchError):\n y = model.begin_training(X, X)\n\n","subject":"Update test and import errors","message":"Update test and import errors\n","lang":"Python","license":"mit","repos":"explosion\/thinc,explosion\/thinc,explosion\/thinc,spacy-io\/thinc,spacy-io\/thinc,spacy-io\/thinc,explosion\/thinc"} {"commit":"3e413b9f0afea5e33f8698e13984fe5dcf4783dd","old_file":"src\/core\/homepage_elements\/about\/hooks.py","new_file":"src\/core\/homepage_elements\/about\/hooks.py","old_contents":"__copyright__ = \"Copyright 2017 Birkbeck, University of London\"\n__author__ = \"Martin Paul Eve & Andy Byers\"\n__license__ = \"AGPL v3\"\n__maintainer__ = \"Birkbeck Centre for Technology and Publishing\"\n\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom utils.setting_handler import get_plugin_setting\nfrom core.homepage_elements.about import plugin_settings\n\n\ndef yield_homepage_element_context(request, homepage_elements):\n if homepage_elements is not None and homepage_elements.filter(name='About').exists():\n\n try:\n title = get_plugin_setting(\n plugin_settings.get_self(),\n 'about_title',\n request.journal,\n )\n title_value = title.value if title.value else ''\n except IndexError:\n title_value = _('About this Journal')\n\n return {\n 'about_content': request.journal.description,\n 'title_value': title_value,\n }\n else:\n return {}\n","new_contents":"__copyright__ = \"Copyright 2017 Birkbeck, University of London\"\n__author__ = \"Martin Paul Eve & Andy Byers\"\n__license__ = \"AGPL v3\"\n__maintainer__ = \"Birkbeck Centre for Technology and Publishing\"\n\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom utils.setting_handler import get_plugin_setting\nfrom core.homepage_elements.about import plugin_settings\n\n\ndef yield_homepage_element_context(request, homepage_elements):\n if homepage_elements is not None and homepage_elements.filter(name='About').exists():\n\n try:\n title = get_plugin_setting(\n plugin_settings.get_self(),\n 'about_title',\n request.journal,\n )\n title_value = title.value if title.value else ''\n except AttributeError:\n title_value = _('About this Journal')\n\n return {\n 'about_content': request.journal.description,\n 'title_value': title_value,\n }\n else:\n return {}\n","subject":"Swap IndexError for AttributeError as a result of the swap from HVAD to MT","message":"Swap IndexError for AttributeError as a result of the swap from HVAD to MT\n","lang":"Python","license":"agpl-3.0","repos":"BirkbeckCTP\/janeway,BirkbeckCTP\/janeway,BirkbeckCTP\/janeway,BirkbeckCTP\/janeway"} {"commit":"838012c457d6c963707bb16259cd72d28c231672","old_file":"cellcounter\/accounts\/decorators.py","new_file":"cellcounter\/accounts\/decorators.py","old_contents":"__author__ = 'jvc26'\n","new_contents":"from functools import wraps\n\nfrom ratelimit.exceptions import Ratelimited\nfrom ratelimit.helpers import is_ratelimited\n\n\ndef registration_ratelimit(ip=True, block=False, method=['POST'], field=None, rate='1\/h',\n skip_if=None, keys=None):\n def decorator(fn):\n @wraps(fn)\n def _wrapped(request, *args, **kw):\n request.limited = getattr(request, 'limited', False)\n if skip_if is None or not skip_if(request):\n ratelimited = is_ratelimited(request=request, increment=False,\n ip=ip, method=method, field=field,\n rate=rate, keys=keys)\n if ratelimited and block:\n raise Ratelimited()\n return_val, success = fn(request, *args, **kw)\n if success:\n is_ratelimited(request=request, increment=True, ip=ip,\n method=method, field=field, rate=rate, keys=keys)\n return return_val\n return _wrapped\n return decorator","subject":"Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors","message":"Use custom decorator to allow ratelimiting only on successful POST - prevents blocking form errors\n","lang":"Python","license":"mit","repos":"haematologic\/cellcounter,cellcounter\/cellcounter,cellcounter\/cellcounter,haematologic\/cellcounter,cellcounter\/cellcounter,cellcounter\/cellcounter,haematologic\/cellcounter"} {"commit":"6ec3b50a087e68373f71162b3dd2421ce7655e4f","old_file":"neuroimaging\/testing\/__init__.py","new_file":"neuroimaging\/testing\/__init__.py","old_contents":"\"\"\"The testing directory contains a small set of imaging files to be used\nfor doctests only. More thorough tests and example data will be stored in\na nipy-data-suite to be created later and downloaded separately.\n\nExamples\n--------\n\n>>> from neuroimaging.testing import funcfile\n>>> from neuroimaging.core.image import image\n>>> img = image.load(funcfile)\n>>> img.shape\n(20, 2, 20, 20)\n\nNotes\n-----\nBUG: anatomical.nii.gz is a copy of functional.nii.gz. This is a place-holder\n until we build a proper anatomical test image.\n\n\"\"\"\n\nimport os\n\n#__all__ = ['funcfile', 'anatfile']\n\n# Discover directory path\nfilepath = os.path.abspath(__file__)\nbasedir = os.path.dirname(filepath)\n\nfuncfile = os.path.join(basedir, 'functional.nii.gz')\nanatfile = os.path.join(basedir, 'anatomical.nii.gz')\n\nfrom numpy.testing import *\nimport decorators as dec\n","new_contents":"\"\"\"The testing directory contains a small set of imaging files to be used\nfor doctests only. More thorough tests and example data will be stored in\na nipy-data-suite to be created later and downloaded separately.\n\nExamples\n--------\n\n>>> from neuroimaging.testing import funcfile\n>>> from neuroimaging.core.image import image\n>>> img = image.load(funcfile)\n>>> img.shape\n(20, 2, 20, 20)\n\nNotes\n-----\nBUG: anatomical.nii.gz is a copy of functional.nii.gz. This is a place-holder\n until we build a proper anatomical test image.\n\n\"\"\"\n\nimport os\n\n#__all__ = ['funcfile', 'anatfile']\n\n# Discover directory path\nfilepath = os.path.abspath(__file__)\nbasedir = os.path.dirname(filepath)\n\nfuncfile = os.path.join(basedir, 'functional.nii.gz')\nanatfile = os.path.join(basedir, 'anatomical.nii.gz')\n\nfrom numpy.testing import *\nimport decorators as dec\nfrom nose.tools import assert_true, assert_false\n\n","subject":"Add some nose.tools to testing imports.","message":"Add some nose.tools to testing imports.","lang":"Python","license":"bsd-3-clause","repos":"alexis-roche\/nipy,nipy\/nipy-labs,bthirion\/nipy,alexis-roche\/niseg,arokem\/nipy,alexis-roche\/nipy,arokem\/nipy,bthirion\/nipy,nipy\/nireg,nipy\/nipy-labs,alexis-roche\/niseg,nipy\/nireg,alexis-roche\/nipy,arokem\/nipy,arokem\/nipy,alexis-roche\/register,alexis-roche\/nireg,alexis-roche\/register,bthirion\/nipy,alexis-roche\/register,alexis-roche\/nipy,alexis-roche\/nireg,bthirion\/nipy"} {"commit":"bda420a0f9abd31b78decdc43359d0dcff36381f","old_file":"zephyr\/management\/commands\/dump_pointers.py","new_file":"zephyr\/management\/commands\/dump_pointers.py","old_contents":"from optparse import make_option\nfrom django.core.management.base import BaseCommand\nfrom zephyr.models import Realm, UserProfile\nimport simplejson\n\ndef dump():\n pointers = []\n for u in UserProfile.objects.select_related(\"user__email\").all():\n pointers.append((u.user.email, u.pointer))\n file(\"dumped-pointers\", \"w\").write(simplejson.dumps(pointers) + \"\\n\")\n\ndef restore(change):\n for (email, pointer) in simplejson.loads(file(\"dumped-pointers\").read()):\n u = UserProfile.objects.get(user__email=email)\n print \"%s: pointer %s => %s\" % (email, u.pointer, pointer)\n if change:\n u.pointer = pointer\n u.save()\n\nclass Command(BaseCommand):\n option_list = BaseCommand.option_list + (\n make_option('--restore', default=False, action='store_true'),\n make_option('--dry-run', '-n', default=False, action='store_true'),)\n\n def handle(self, *args, **options):\n if options[\"restore\"]:\n restore(change=not options['dry_run'])\n else:\n dump()\n","new_contents":"from optparse import make_option\nfrom django.core.management.base import BaseCommand\nfrom zephyr.models import Realm, UserProfile\nimport simplejson\n\ndef dump():\n pointers = []\n for u in UserProfile.objects.select_related(\"user__email\").all():\n pointers.append((u.user.email, u.pointer))\n file(\"dumped-pointers\", \"w\").write(simplejson.dumps(pointers) + \"\\n\")\n\ndef restore(change):\n for (email, pointer) in simplejson.loads(file(\"dumped-pointers\").read()):\n u = UserProfile.objects.get(user__email__iexact=email)\n print \"%s: pointer %s => %s\" % (email, u.pointer, pointer)\n if change:\n u.pointer = pointer\n u.save()\n\nclass Command(BaseCommand):\n option_list = BaseCommand.option_list + (\n make_option('--restore', default=False, action='store_true'),\n make_option('--dry-run', '-n', default=False, action='store_true'),)\n\n def handle(self, *args, **options):\n if options[\"restore\"]:\n restore(change=not options['dry_run'])\n else:\n dump()\n","subject":"Fix email case issues when restoring user pointers.","message":"Fix email case issues when restoring user pointers.\n\n(imported from commit 84d3288dffc1cb010d8cd2a749fe71aa2a4d0df3)\n","lang":"Python","license":"apache-2.0","repos":"KingxBanana\/zulip,zachallaun\/zulip,themass\/zulip,avastu\/zulip,bastianh\/zulip,bitemyapp\/zulip,EasonYi\/zulip,lfranchi\/zulip,levixie\/zulip,pradiptad\/zulip,adnanh\/zulip,hengqujushi\/zulip,vaidap\/zulip,ashwinirudrappa\/zulip,reyha\/zulip,brainwane\/zulip,arpitpanwar\/zulip,reyha\/zulip,grave-w-grave\/zulip,KJin99\/zulip,Gabriel0402\/zulip,yocome\/zulip,Drooids\/zulip,proliming\/zulip,jessedhillon\/zulip,mohsenSy\/zulip,souravbadami\/zulip,bssrdf\/zulip,tbutter\/zulip,hafeez3000\/zulip,grave-w-grave\/zulip,joyhchen\/zulip,technicalpickles\/zulip,mansilladev\/zulip,bowlofstew\/zulip,kokoar\/zulip,zorojean\/zulip,dawran6\/zulip,kokoar\/zulip,ApsOps\/zulip,Qgap\/zulip,hafeez3000\/zulip,qq1012803704\/zulip,Drooids\/zulip,mahim97\/zulip,voidException\/zulip,kokoar\/zulip,showell\/zulip,EasonYi\/zulip,johnny9\/zulip,jerryge\/zulip,hafeez3000\/zulip,dxq-git\/zulip,tbutter\/zulip,tbutter\/zulip,amanharitsh123\/zulip,jphilipsen05\/zulip,tiansiyuan\/zulip,LeeRisk\/zulip,reyha\/zulip,brainwane\/zulip,PhilSk\/zulip,deer-hope\/zulip,dwrpayne\/zulip,synicalsyntax\/zulip,willingc\/zulip,jrowan\/zulip,showell\/zulip,KingxBanana\/zulip,eastlhu\/zulip,saitodisse\/zulip,ashwinirudrappa\/zulip,amyliu345\/zulip,eastlhu\/zulip,noroot\/zulip,joshisa\/zulip,christi3k\/zulip,developerfm\/zulip,natanovia\/zulip,developerfm\/zulip,wweiradio\/zulip,armooo\/zulip,hengqujushi\/zulip,blaze225\/zulip,ufosky-server\/zulip,shaunstanislaus\/zulip,dwrpayne\/zulip,JanzTam\/zulip,luyifan\/zulip,susansls\/zulip,dhcrzf\/zulip,tdr130\/zulip,aps-sids\/zulip,hackerkid\/zulip,jessedhillon\/zulip,hackerkid\/zulip,tommyip\/zulip,atomic-labs\/zulip,j831\/zulip,ryanbackman\/zulip,zhaoweigg\/zulip,shubhamdhama\/zulip,gkotian\/zulip,rht\/zulip,isht3\/zulip,amyliu345\/zulip,zachallaun\/zulip,zulip\/zulip,Drooids\/zulip,ryansnowboarder\/zulip,AZtheAsian\/zulip,LeeRisk\/zulip,proliming\/zulip,pradiptad\/zulip,adnanh\/zulip,hj3938\/zulip,kaiyuanheshang\/zulip,sharmaeklavya2\/zulip,littledogboy\/zulip,sup95\/zulip,tbutter\/zulip,Vallher\/zulip,luyifan\/zulip,AZtheAsian\/zulip,wangdeshui\/zulip,he15his\/zulip,itnihao\/zulip,vabs22\/zulip,Vallher\/zulip,yuvipanda\/zulip,saitodisse\/zulip,RobotCaleb\/zulip,adnanh\/zulip,hayderimran7\/zulip,fw1121\/zulip,lfranchi\/zulip,bluesea\/zulip,ikasumiwt\/zulip,Galexrt\/zulip,karamcnair\/zulip,m1ssou\/zulip,jrowan\/zulip,lfranchi\/zulip,karamcnair\/zulip,dawran6\/zulip,ryanbackman\/zulip,jackrzhang\/zulip,zulip\/zulip,joshisa\/zulip,hj3938\/zulip,levixie\/zulip,krtkmj\/zulip,aps-sids\/zulip,bowlofstew\/zulip,LeeRisk\/zulip,hj3938\/zulip,KJin99\/zulip,eastlhu\/zulip,Vallher\/zulip,tiansiyuan\/zulip,Drooids\/zulip,ufosky-server\/zulip,zulip\/zulip,easyfmxu\/zulip,calvinleenyc\/zulip,shubhamdhama\/zulip,dwrpayne\/zulip,joyhchen\/zulip,timabbott\/zulip,jimmy54\/zulip,timabbott\/zulip,MayB\/zulip,proliming\/zulip,kaiyuanheshang\/zulip,KJin99\/zulip,wdaher\/zulip,eastlhu\/zulip,johnny9\/zulip,MayB\/zulip,Jianchun1\/zulip,zacps\/zulip,jimmy54\/zulip,littledogboy\/zulip,DazWorrall\/zulip,xuxiao\/zulip,saitodisse\/zulip,alliejones\/zulip,hafeez3000\/zulip,synicalsyntax\/zulip,nicholasbs\/zulip,shrikrishnaholla\/zulip,bssrdf\/zulip,amanharitsh123\/zulip,timabbott\/zulip,swinghu\/zulip,babbage\/zulip,niftynei\/zulip,jrowan\/zulip,bastianh\/zulip,he15his\/zulip,developerfm\/zulip,JPJPJPOPOP\/zulip,schatt\/zulip,JanzTam\/zulip,mohsenSy\/zulip,Qgap\/zulip,gigawhitlocks\/zulip,xuxiao\/zulip,codeKonami\/zulip,Frouk\/zulip,hengqujushi\/zulip,themass\/zulip,jainayush975\/zulip,developerfm\/zulip,SmartPeople\/zulip,aps-sids\/zulip,ashwinirudrappa\/zulip,tommyip\/zulip,joyhchen\/zulip,hengqujushi\/zulip,wweiradio\/zulip,blaze225\/zulip,esander91\/zulip,sharmaeklavya2\/zulip,DazWorrall\/zulip,ahmadassaf\/zulip,vakila\/zulip,proliming\/zulip,willingc\/zulip,Qgap\/zulip,Suninus\/zulip,tbutter\/zulip,krtkmj\/zulip,tdr130\/zulip,dwrpayne\/zulip,Qgap\/zulip,moria\/zulip,alliejones\/zulip,Gabriel0402\/zulip,jessedhillon\/zulip,Qgap\/zulip,sharmaeklavya2\/zulip,JPJPJPOPOP\/zulip,kou\/zulip,bastianh\/zulip,Juanvulcano\/zulip,krtkmj\/zulip,kou\/zulip,littledogboy\/zulip,kokoar\/zulip,hackerkid\/zulip,hj3938\/zulip,Cheppers\/zulip,dawran6\/zulip,zwily\/zulip,natanovia\/zulip,mansilladev\/zulip,Qgap\/zulip,wavelets\/zulip,so0k\/zulip,wdaher\/zulip,karamcnair\/zulip,technicalpickles\/zulip,dawran6\/zulip,Juanvulcano\/zulip,ryansnowboarder\/zulip,bssrdf\/zulip,ipernet\/zulip,peguin40\/zulip,krtkmj\/zulip,peiwei\/zulip,samatdav\/zulip,amanharitsh123\/zulip,mdavid\/zulip,mansilladev\/zulip,willingc\/zulip,peiwei\/zulip,vaidap\/zulip,xuxiao\/zulip,avastu\/zulip,joshisa\/zulip,dhcrzf\/zulip,tdr130\/zulip,peguin40\/zulip,xuanhan863\/zulip,kou\/zulip,ericzhou2008\/zulip,verma-varsha\/zulip,EasonYi\/zulip,he15his\/zulip,firstblade\/zulip,tommyip\/zulip,jerryge\/zulip,vikas-parashar\/zulip,bastianh\/zulip,natanovia\/zulip,brockwhittaker\/zulip,swinghu\/zulip,Cheppers\/zulip,atomic-labs\/zulip,shrikrishnaholla\/zulip,udxxabp\/zulip,fw1121\/zulip,wavelets\/zulip,dotcool\/zulip,bowlofstew\/zulip,zacps\/zulip,qq1012803704\/zulip,armooo\/zulip,moria\/zulip,sup95\/zulip,jainayush975\/zulip,itnihao\/zulip,hafeez3000\/zulip,wavelets\/zulip,voidException\/zulip,tommyip\/zulip,jonesgithub\/zulip,praveenaki\/zulip,dwrpayne\/zulip,xuanhan863\/zulip,isht3\/zulip,johnny9\/zulip,andersk\/zulip,cosmicAsymmetry\/zulip,paxapy\/zulip,suxinde2009\/zulip,JPJPJPOPOP\/zulip,vikas-parashar\/zulip,suxinde2009\/zulip,glovebx\/zulip,showell\/zulip,seapasulli\/zulip,esander91\/zulip,rishig\/zulip,fw1121\/zulip,joshisa\/zulip,itnihao\/zulip,jerryge\/zulip,xuanhan863\/zulip,stamhe\/zulip,so0k\/zulip,esander91\/zulip,Galexrt\/zulip,Jianchun1\/zulip,noroot\/zulip,TigorC\/zulip,zorojean\/zulip,stamhe\/zulip,gkotian\/zulip,johnnygaddarr\/zulip,zulip\/zulip,yocome\/zulip,MariaFaBella85\/zulip,codeKonami\/zulip,codeKonami\/zulip,dxq-git\/zulip,hj3938\/zulip,fw1121\/zulip,dawran6\/zulip,AZtheAsian\/zulip,paxapy\/zulip,synicalsyntax\/zulip,PhilSk\/zulip,hayderimran7\/zulip,noroot\/zulip,calvinleenyc\/zulip,andersk\/zulip,Drooids\/zulip,ahmadassaf\/zulip,Gabriel0402\/zulip,bluesea\/zulip,suxinde2009\/zulip,mdavid\/zulip,deer-hope\/zulip,samatdav\/zulip,shubhamdhama\/zulip,jonesgithub\/zulip,ikasumiwt\/zulip,zhaoweigg\/zulip,he15his\/zulip,avastu\/zulip,seapasulli\/zulip,eeshangarg\/zulip,lfranchi\/zulip,noroot\/zulip,KingxBanana\/zulip,rht\/zulip,sharmaeklavya2\/zulip,arpith\/zulip,dotcool\/zulip,m1ssou\/zulip,rishig\/zulip,themass\/zulip,LeeRisk\/zulip,jrowan\/zulip,karamcnair\/zulip,hayderimran7\/zulip,vabs22\/zulip,zulip\/zulip,brockwhittaker\/zulip,thomasboyt\/zulip,vakila\/zulip,Batterfii\/zulip,ApsOps\/zulip,firstblade\/zulip,amyliu345\/zulip,sup95\/zulip,udxxabp\/zulip,souravbadami\/zulip,ikasumiwt\/zulip,joshisa\/zulip,itnihao\/zulip,glovebx\/zulip,atomic-labs\/zulip,johnny9\/zulip,wangdeshui\/zulip,dattatreya303\/zulip,luyifan\/zulip,EasonYi\/zulip,peiwei\/zulip,RobotCaleb\/zulip,jimmy54\/zulip,so0k\/zulip,brockwhittaker\/zulip,guiquanz\/zulip,ApsOps\/zulip,JanzTam\/zulip,eeshangarg\/zulip,lfranchi\/zulip,shaunstanislaus\/zulip,dotcool\/zulip,Frouk\/zulip,LAndreas\/zulip,natanovia\/zulip,calvinleenyc\/zulip,gigawhitlocks\/zulip,wweiradio\/zulip,ericzhou2008\/zulip,ipernet\/zulip,Vallher\/zulip,tiansiyuan\/zulip,dnmfarrell\/zulip,technicalpickles\/zulip,verma-varsha\/zulip,easyfmxu\/zulip,timabbott\/zulip,technicalpickles\/zulip,eeshangarg\/zulip,krtkmj\/zulip,Jianchun1\/zulip,isht3\/zulip,johnnygaddarr\/zulip,dhcrzf\/zulip,calvinleenyc\/zulip,dwrpayne\/zulip,firstblade\/zulip,zhaoweigg\/zulip,niftynei\/zulip,huangkebo\/zulip,Frouk\/zulip,verma-varsha\/zulip,RobotCaleb\/zulip,jphilipsen05\/zulip,souravbadami\/zulip,niftynei\/zulip,ipernet\/zulip,synicalsyntax\/zulip,schatt\/zulip,rht\/zulip,hafeez3000\/zulip,punchagan\/zulip,levixie\/zulip,ipernet\/zulip,samatdav\/zulip,Galexrt\/zulip,mahim97\/zulip,luyifan\/zulip,umkay\/zulip,peguin40\/zulip,arpitpanwar\/zulip,gigawhitlocks\/zulip,tdr130\/zulip,Juanvulcano\/zulip,showell\/zulip,EasonYi\/zulip,jeffcao\/zulip,bluesea\/zulip,amanharitsh123\/zulip,bluesea\/zulip,jeffcao\/zulip,ikasumiwt\/zulip,reyha\/zulip,sharmaeklavya2\/zulip,amyliu345\/zulip,zwily\/zulip,noroot\/zulip,ahmadassaf\/zulip,fw1121\/zulip,JanzTam\/zulip,bssrdf\/zulip,voidException\/zulip,Batterfii\/zulip,dotcool\/zulip,rishig\/zulip,j831\/zulip,natanovia\/zulip,glovebx\/zulip,brainwane\/zulip,krtkmj\/zulip,atomic-labs\/zulip,andersk\/zulip,dnmfarrell\/zulip,bowlofstew\/zulip,Suninus\/zulip,dwrpayne\/zulip,armooo\/zulip,Gabriel0402\/zulip,tiansiyuan\/zulip,karamcnair\/zulip,jerryge\/zulip,sup95\/zulip,peiwei\/zulip,johnnygaddarr\/zulip,qq1012803704\/zulip,eastlhu\/zulip,PaulPetring\/zulip,hustlzp\/zulip,ipernet\/zulip,ikasumiwt\/zulip,yocome\/zulip,xuanhan863\/zulip,j831\/zulip,aakash-cr7\/zulip,wweiradio\/zulip,Drooids\/zulip,LAndreas\/zulip,ryansnowboarder\/zulip,tdr130\/zulip,natanovia\/zulip,MariaFaBella85\/zulip,suxinde2009\/zulip,jainayush975\/zulip,synicalsyntax\/zulip,KJin99\/zulip,arpith\/zulip,verma-varsha\/zulip,wweiradio\/zulip,hengqujushi\/zulip,wavelets\/zulip,itnihao\/zulip,amallia\/zulip,karamcnair\/zulip,shaunstanislaus\/zulip,willingc\/zulip,RobotCaleb\/zulip,SmartPeople\/zulip,samatdav\/zulip,nicholasbs\/zulip,m1ssou\/zulip,bluesea\/zulip,swinghu\/zulip,huangkebo\/zulip,jessedhillon\/zulip,alliejones\/zulip,Suninus\/zulip,j831\/zulip,akuseru\/zulip,yuvipanda\/zulip,arpitpanwar\/zulip,bitemyapp\/zulip,moria\/zulip,lfranchi\/zulip,kaiyuanheshang\/zulip,AZtheAsian\/zulip,so0k\/zulip,dhcrzf\/zulip,qq1012803704\/zulip,umkay\/zulip,ahmadassaf\/zulip,JPJPJPOPOP\/zulip,punchagan\/zulip,glovebx\/zulip,themass\/zulip,zacps\/zulip,timabbott\/zulip,paxapy\/zulip,jrowan\/zulip,dotcool\/zulip,ApsOps\/zulip,LAndreas\/zulip,noroot\/zulip,avastu\/zulip,Diptanshu8\/zulip,zorojean\/zulip,so0k\/zulip,arpitpanwar\/zulip,j831\/zulip,MariaFaBella85\/zulip,seapasulli\/zulip,jimmy54\/zulip,rht\/zulip,stamhe\/zulip,wdaher\/zulip,eeshangarg\/zulip,ryansnowboarder\/zulip,bluesea\/zulip,SmartPeople\/zulip,Frouk\/zulip,mdavid\/zulip,paxapy\/zulip,qq1012803704\/zulip,shubhamdhama\/zulip,PaulPetring\/zulip,DazWorrall\/zulip,Batterfii\/zulip,andersk\/zulip,babbage\/zulip,umkay\/zulip,hafeez3000\/zulip,themass\/zulip,Diptanshu8\/zulip,krtkmj\/zulip,aakash-cr7\/zulip,johnny9\/zulip,punchagan\/zulip,cosmicAsymmetry\/zulip,kaiyuanheshang\/zulip,joyhchen\/zulip,joyhchen\/zulip,sonali0901\/zulip,jphilipsen05\/zulip,dnmfarrell\/zulip,TigorC\/zulip,dattatreya303\/zulip,mansilladev\/zulip,eeshangarg\/zulip,shrikrishnaholla\/zulip,moria\/zulip,guiquanz\/zulip,zorojean\/zulip,bowlofstew\/zulip,Jianchun1\/zulip,KingxBanana\/zulip,zofuthan\/zulip,Cheppers\/zulip,ipernet\/zulip,zorojean\/zulip,armooo\/zulip,littledogboy\/zulip,yuvipanda\/zulip,vabs22\/zulip,alliejones\/zulip,ikasumiwt\/zulip,firstblade\/zulip,guiquanz\/zulip,bowlofstew\/zulip,yocome\/zulip,Juanvulcano\/zulip,yocome\/zulip,wavelets\/zulip,SmartPeople\/zulip,akuseru\/zulip,bluesea\/zulip,huangkebo\/zulip,dnmfarrell\/zulip,showell\/zulip,jimmy54\/zulip,zachallaun\/zulip,mohsenSy\/zulip,ericzhou2008\/zulip,brockwhittaker\/zulip,praveenaki\/zulip,gkotian\/zulip,fw1121\/zulip,TigorC\/zulip,codeKonami\/zulip,kou\/zulip,dattatreya303\/zulip,zwily\/zulip,ericzhou2008\/zulip,stamhe\/zulip,mohsenSy\/zulip,udxxabp\/zulip,Diptanshu8\/zulip,luyifan\/zulip,showell\/zulip,zorojean\/zulip,Suninus\/zulip,TigorC\/zulip,arpitpanwar\/zulip,isht3\/zulip,xuxiao\/zulip,Vallher\/zulip,xuanhan863\/zulip,zorojean\/zulip,saitodisse\/zulip,JPJPJPOPOP\/zulip,jonesgithub\/zulip,adnanh\/zulip,proliming\/zulip,jonesgithub\/zulip,souravbadami\/zulip,udxxabp\/zulip,guiquanz\/zulip,AZtheAsian\/zulip,zhaoweigg\/zulip,zwily\/zulip,AZtheAsian\/zulip,hackerkid\/zulip,amanharitsh123\/zulip,babbage\/zulip,bitemyapp\/zulip,saitodisse\/zulip,cosmicAsymmetry\/zulip,Juanvulcano\/zulip,voidException\/zulip,deer-hope\/zulip,huangkebo\/zulip,dxq-git\/zulip,vaidap\/zulip,Galexrt\/zulip,mahim97\/zulip,wavelets\/zulip,cosmicAsymmetry\/zulip,dxq-git\/zulip,jeffcao\/zulip,proliming\/zulip,esander91\/zulip,PaulPetring\/zulip,kaiyuanheshang\/zulip,adnanh\/zulip,guiquanz\/zulip,blaze225\/zulip,huangkebo\/zulip,eastlhu\/zulip,Vallher\/zulip,paxapy\/zulip,huangkebo\/zulip,avastu\/zulip,nicholasbs\/zulip,jackrzhang\/zulip,Diptanshu8\/zulip,vabs22\/zulip,adnanh\/zulip,suxinde2009\/zulip,vakila\/zulip,bitemyapp\/zulip,deer-hope\/zulip,vabs22\/zulip,Cheppers\/zulip,amallia\/zulip,m1ssou\/zulip,rishig\/zulip,voidException\/zulip,ApsOps\/zulip,RobotCaleb\/zulip,alliejones\/zulip,PaulPetring\/zulip,dxq-git\/zulip,MayB\/zulip,samatdav\/zulip,samatdav\/zulip,brainwane\/zulip,DazWorrall\/zulip,bastianh\/zulip,cosmicAsymmetry\/zulip,aakash-cr7\/zulip,schatt\/zulip,hj3938\/zulip,m1ssou\/zulip,Frouk\/zulip,xuxiao\/zulip,ahmadassaf\/zulip,punchagan\/zulip,seapasulli\/zulip,JanzTam\/zulip,amyliu345\/zulip,zacps\/zulip,joshisa\/zulip,glovebx\/zulip,saitodisse\/zulip,tdr130\/zulip,ericzhou2008\/zulip,ryanbackman\/zulip,moria\/zulip,RobotCaleb\/zulip,Cheppers\/zulip,umkay\/zulip,brockwhittaker\/zulip,christi3k\/zulip,technicalpickles\/zulip,brainwane\/zulip,natanovia\/zulip,rht\/zulip,peiwei\/zulip,developerfm\/zulip,brainwane\/zulip,aps-sids\/zulip,calvinleenyc\/zulip,ashwinirudrappa\/zulip,LeeRisk\/zulip,aliceriot\/zulip,jonesgithub\/zulip,wangdeshui\/zulip,dhcrzf\/zulip,sup95\/zulip,Drooids\/zulip,ericzhou2008\/zulip,johnny9\/zulip,xuxiao\/zulip,shaunstanislaus\/zulip,PhilSk\/zulip,vikas-parashar\/zulip,praveenaki\/zulip,PhilSk\/zulip,esander91\/zulip,jeffcao\/zulip,hustlzp\/zulip,ipernet\/zulip,dattatreya303\/zulip,susansls\/zulip,bastianh\/zulip,umkay\/zulip,synicalsyntax\/zulip,schatt\/zulip,wavelets\/zulip,johnnygaddarr\/zulip,jimmy54\/zulip,pradiptad\/zulip,akuseru\/zulip,sonali0901\/zulip,niftynei\/zulip,levixie\/zulip,mahim97\/zulip,susansls\/zulip,Suninus\/zulip,mansilladev\/zulip,hengqujushi\/zulip,tommyip\/zulip,EasonYi\/zulip,suxinde2009\/zulip,JanzTam\/zulip,aliceriot\/zulip,zacps\/zulip,so0k\/zulip,yuvipanda\/zulip,Suninus\/zulip,akuseru\/zulip,zhaoweigg\/zulip,johnnygaddarr\/zulip,kokoar\/zulip,ashwinirudrappa\/zulip,gkotian\/zulip,shaunstanislaus\/zulip,dawran6\/zulip,peguin40\/zulip,thomasboyt\/zulip,so0k\/zulip,aps-sids\/zulip,amallia\/zulip,KingxBanana\/zulip,dxq-git\/zulip,wdaher\/zulip,sonali0901\/zulip,itnihao\/zulip,technicalpickles\/zulip,akuseru\/zulip,mansilladev\/zulip,zhaoweigg\/zulip,zofuthan\/zulip,bssrdf\/zulip,kokoar\/zulip,willingc\/zulip,arpitpanwar\/zulip,hustlzp\/zulip,sonali0901\/zulip,aakash-cr7\/zulip,vabs22\/zulip,verma-varsha\/zulip,arpith\/zulip,avastu\/zulip,wdaher\/zulip,shrikrishnaholla\/zulip,schatt\/zulip,hj3938\/zulip,deer-hope\/zulip,schatt\/zulip,mahim97\/zulip,alliejones\/zulip,hackerkid\/zulip,littledogboy\/zulip,udxxabp\/zulip,bssrdf\/zulip,rishig\/zulip,swinghu\/zulip,isht3\/zulip,gkotian\/zulip,dxq-git\/zulip,christi3k\/zulip,wangdeshui\/zulip,arpith\/zulip,EasonYi\/zulip,PhilSk\/zulip,suxinde2009\/zulip,joshisa\/zulip,jphilipsen05\/zulip,grave-w-grave\/zulip,babbage\/zulip,luyifan\/zulip,ryanbackman\/zulip,wdaher\/zulip,niftynei\/zulip,shubhamdhama\/zulip,ufosky-server\/zulip,swinghu\/zulip,codeKonami\/zulip,aliceriot\/zulip,amallia\/zulip,gkotian\/zulip,proliming\/zulip,dhcrzf\/zulip,Suninus\/zulip,KingxBanana\/zulip,deer-hope\/zulip,levixie\/zulip,wangdeshui\/zulip,qq1012803704\/zulip,jerryge\/zulip,jackrzhang\/zulip,jainayush975\/zulip,avastu\/zulip,zhaoweigg\/zulip,LeeRisk\/zulip,shrikrishnaholla\/zulip,Batterfii\/zulip,he15his\/zulip,thomasboyt\/zulip,levixie\/zulip,timabbott\/zulip,dattatreya303\/zulip,peguin40\/zulip,shaunstanislaus\/zulip,punchagan\/zulip,Jianchun1\/zulip,jessedhillon\/zulip,m1ssou\/zulip,easyfmxu\/zulip,voidException\/zulip,easyfmxu\/zulip,pradiptad\/zulip,ashwinirudrappa\/zulip,Cheppers\/zulip,hustlzp\/zulip,reyha\/zulip,peiwei\/zulip,isht3\/zulip,punchagan\/zulip,mdavid\/zulip,zofuthan\/zulip,hayderimran7\/zulip,dnmfarrell\/zulip,willingc\/zulip,vaidap\/zulip,arpith\/zulip,guiquanz\/zulip,amyliu345\/zulip,aliceriot\/zulip,akuseru\/zulip,bitemyapp\/zulip,souravbadami\/zulip,easyfmxu\/zulip,Jianchun1\/zulip,glovebx\/zulip,yuvipanda\/zulip,jainayush975\/zulip,jessedhillon\/zulip,atomic-labs\/zulip,cosmicAsymmetry\/zulip,zachallaun\/zulip,johnnygaddarr\/zulip,xuanhan863\/zulip,Juanvulcano\/zulip,brainwane\/zulip,itnihao\/zulip,gigawhitlocks\/zulip,zachallaun\/zulip,jackrzhang\/zulip,armooo\/zulip,calvinleenyc\/zulip,schatt\/zulip,andersk\/zulip,jessedhillon\/zulip,hayderimran7\/zulip,Batterfii\/zulip,ufosky-server\/zulip,noroot\/zulip,reyha\/zulip,littledogboy\/zulip,luyifan\/zulip,lfranchi\/zulip,armooo\/zulip,ufosky-server\/zulip,RobotCaleb\/zulip,willingc\/zulip,ikasumiwt\/zulip,bssrdf\/zulip,MariaFaBella85\/zulip,thomasboyt\/zulip,yocome\/zulip,blaze225\/zulip,nicholasbs\/zulip,hustlzp\/zulip,mohsenSy\/zulip,jainayush975\/zulip,firstblade\/zulip,umkay\/zulip,jeffcao\/zulip,arpitpanwar\/zulip,ApsOps\/zulip,LAndreas\/zulip,tbutter\/zulip,codeKonami\/zulip,ryansnowboarder\/zulip,ashwinirudrappa\/zulip,Diptanshu8\/zulip,nicholasbs\/zulip,pradiptad\/zulip,developerfm\/zulip,he15his\/zulip,zachallaun\/zulip,he15his\/zulip,yocome\/zulip,praveenaki\/zulip,rishig\/zulip,mahim97\/zulip,grave-w-grave\/zulip,xuanhan863\/zulip,rishig\/zulip,shubhamdhama\/zulip,ryansnowboarder\/zulip,Frouk\/zulip,shrikrishnaholla\/zulip,mohsenSy\/zulip,nicholasbs\/zulip,jerryge\/zulip,jonesgithub\/zulip,vakila\/zulip,vikas-parashar\/zulip,vakila\/zulip,yuvipanda\/zulip,esander91\/zulip,dnmfarrell\/zulip,jonesgithub\/zulip,KJin99\/zulip,jimmy54\/zulip,tiansiyuan\/zulip,udxxabp\/zulip,thomasboyt\/zulip,atomic-labs\/zulip,tommyip\/zulip,aakash-cr7\/zulip,DazWorrall\/zulip,PaulPetring\/zulip,Diptanshu8\/zulip,rht\/zulip,dnmfarrell\/zulip,kaiyuanheshang\/zulip,MayB\/zulip,levixie\/zulip,zwily\/zulip,jeffcao\/zulip,MayB\/zulip,karamcnair\/zulip,TigorC\/zulip,aliceriot\/zulip,paxapy\/zulip,timabbott\/zulip,gigawhitlocks\/zulip,dhcrzf\/zulip,thomasboyt\/zulip,vakila\/zulip,johnnygaddarr\/zulip,ryansnowboarder\/zulip,MariaFaBella85\/zulip,joyhchen\/zulip,akuseru\/zulip,kaiyuanheshang\/zulip,jphilipsen05\/zulip,LAndreas\/zulip,stamhe\/zulip,blaze225\/zulip,jackrzhang\/zulip,amallia\/zulip,wweiradio\/zulip,brockwhittaker\/zulip,wdaher\/zulip,deer-hope\/zulip,easyfmxu\/zulip,Qgap\/zulip,codeKonami\/zulip,mdavid\/zulip,ufosky-server\/zulip,souravbadami\/zulip,amallia\/zulip,huangkebo\/zulip,tbutter\/zulip,zofuthan\/zulip,hustlzp\/zulip,dotcool\/zulip,jackrzhang\/zulip,vaidap\/zulip,PhilSk\/zulip,pradiptad\/zulip,christi3k\/zulip,zwily\/zulip,zulip\/zulip,technicalpickles\/zulip,seapasulli\/zulip,qq1012803704\/zulip,Cheppers\/zulip,gigawhitlocks\/zulip,JanzTam\/zulip,mansilladev\/zulip,arpith\/zulip,developerfm\/zulip,littledogboy\/zulip,rht\/zulip,voidException\/zulip,PaulPetring\/zulip,tiansiyuan\/zulip,susansls\/zulip,vikas-parashar\/zulip,christi3k\/zulip,punchagan\/zulip,nicholasbs\/zulip,verma-varsha\/zulip,sonali0901\/zulip,zacps\/zulip,DazWorrall\/zulip,MariaFaBella85\/zulip,babbage\/zulip,tommyip\/zulip,vakila\/zulip,aps-sids\/zulip,Galexrt\/zulip,xuxiao\/zulip,hayderimran7\/zulip,moria\/zulip,kokoar\/zulip,wangdeshui\/zulip,udxxabp\/zulip,bowlofstew\/zulip,amanharitsh123\/zulip,grave-w-grave\/zulip,vaidap\/zulip,blaze225\/zulip,hayderimran7\/zulip,JPJPJPOPOP\/zulip,glovebx\/zulip,synicalsyntax\/zulip,shubhamdhama\/zulip,zofuthan\/zulip,atomic-labs\/zulip,DazWorrall\/zulip,Galexrt\/zulip,LAndreas\/zulip,gkotian\/zulip,alliejones\/zulip,Galexrt\/zulip,MayB\/zulip,kou\/zulip,tiansiyuan\/zulip,kou\/zulip,eeshangarg\/zulip,seapasulli\/zulip,guiquanz\/zulip,pradiptad\/zulip,mdavid\/zulip,Vallher\/zulip,hengqujushi\/zulip,adnanh\/zulip,aps-sids\/zulip,zofuthan\/zulip,praveenaki\/zulip,j831\/zulip,hackerkid\/zulip,hustlzp\/zulip,babbage\/zulip,esander91\/zulip,Batterfii\/zulip,themass\/zulip,jeffcao\/zulip,firstblade\/zulip,LAndreas\/zulip,ahmadassaf\/zulip,aliceriot\/zulip,MayB\/zulip,praveenaki\/zulip,andersk\/zulip,ryanbackman\/zulip,christi3k\/zulip,susansls\/zulip,PaulPetring\/zulip,dattatreya303\/zulip,moria\/zulip,tdr130\/zulip,dotcool\/zulip,babbage\/zulip,ahmadassaf\/zulip,zwily\/zulip,susansls\/zulip,armooo\/zulip,praveenaki\/zulip,ryanbackman\/zulip,mdavid\/zulip,gigawhitlocks\/zulip,jackrzhang\/zulip,yuvipanda\/zulip,zofuthan\/zulip,Gabriel0402\/zulip,hackerkid\/zulip,vikas-parashar\/zulip,SmartPeople\/zulip,themass\/zulip,wangdeshui\/zulip,stamhe\/zulip,stamhe\/zulip,wweiradio\/zulip,bastianh\/zulip,Batterfii\/zulip,sup95\/zulip,ericzhou2008\/zulip,KJin99\/zulip,shrikrishnaholla\/zulip,bitemyapp\/zulip,swinghu\/zulip,johnny9\/zulip,easyfmxu\/zulip,Gabriel0402\/zulip,seapasulli\/zulip,aakash-cr7\/zulip,KJin99\/zulip,saitodisse\/zulip,showell\/zulip,niftynei\/zulip,Gabriel0402\/zulip,jphilipsen05\/zulip,jrowan\/zulip,fw1121\/zulip,sonali0901\/zulip,TigorC\/zulip,MariaFaBella85\/zulip,ufosky-server\/zulip,jerryge\/zulip,eastlhu\/zulip,zachallaun\/zulip,peguin40\/zulip,LeeRisk\/zulip,m1ssou\/zulip,grave-w-grave\/zulip,sharmaeklavya2\/zulip,umkay\/zulip,ApsOps\/zulip,andersk\/zulip,firstblade\/zulip,SmartPeople\/zulip,zulip\/zulip,Frouk\/zulip,eeshangarg\/zulip,aliceriot\/zulip,kou\/zulip,amallia\/zulip,peiwei\/zulip,swinghu\/zulip,shaunstanislaus\/zulip,thomasboyt\/zulip,bitemyapp\/zulip"} {"commit":"4742f587e3e66fd1916dcb7200517e2ac06ddcf4","old_file":"uconnrcmpy\/__init__.py","new_file":"uconnrcmpy\/__init__.py","old_contents":"from .ignitiondelayexp import ExperimentalIgnitionDelay\nfrom .compare_to_sim import CompareToSimulation\nfrom .volume_trace import VolumeTraceBuilder\nfrom .nonreactive import NonReactiveExperiments\n\n__all__ = [\n 'ExperimentalIgnitionDelay',\n 'CompareToSimulation',\n 'VolumeTraceBuilder',\n 'NonReactiveExperiments',\n]\n","new_contents":"import sys\nif sys.version_info[0] < 3 and sys.version_info[1] < 4:\n raise Exception('Python 3.4 is required to use this package.')\n\nfrom .ignitiondelayexp import ExperimentalIgnitionDelay\nfrom .compare_to_sim import CompareToSimulation\nfrom .volume_trace import VolumeTraceBuilder\nfrom .nonreactive import NonReactiveExperiments\n\n__all__ = [\n 'ExperimentalIgnitionDelay',\n 'CompareToSimulation',\n 'VolumeTraceBuilder',\n 'NonReactiveExperiments',\n]\n","subject":"Enforce Python >= 3.4 on import of the package","message":"Enforce Python >= 3.4 on import of the package\n\nPython 3.4 is required for the pathlib module\n","lang":"Python","license":"bsd-3-clause","repos":"bryanwweber\/UConnRCMPy"} {"commit":"77cf2fb0f63a5520de3b8b3456ce4c9181b91d16","old_file":"spacy\/tests\/regression\/test_issue595.py","new_file":"spacy\/tests\/regression\/test_issue595.py","old_contents":"from __future__ import unicode_literals\nimport pytest\n\nfrom ...symbols import POS, VERB, VerbForm_inf\nfrom ...tokens import Doc\nfrom ...vocab import Vocab\nfrom ...lemmatizer import Lemmatizer\n\n\n@pytest.fixture\ndef index():\n return {'verb': {}}\n\n@pytest.fixture\ndef exceptions():\n return {'verb': {}}\n\n@pytest.fixture\ndef rules():\n return {\"verb\": [[\"ed\", \"e\"]]}\n\n@pytest.fixture\ndef lemmatizer(index, exceptions, rules):\n return Lemmatizer(index, exceptions, rules)\n\n\n@pytest.fixture\ndef tag_map():\n return {'VB': {POS: VERB, 'morph': VerbForm_inf}}\n\n\n@pytest.fixture\ndef vocab(lemmatizer, tag_map):\n return Vocab(lemmatizer=lemmatizer, tag_map=tag_map)\n\n\ndef test_not_lemmatize_base_forms(vocab, lemmatizer):\n doc = Doc(vocab, words=[\"Do\", \"n't\", \"feed\", \"the\", \"dog\"])\n feed = doc[2]\n feed.tag_ = u'VB'\n assert feed.text == u'feed'\n assert feed.lemma_ == u'feed'\n","new_contents":"from __future__ import unicode_literals\nimport pytest\n\nfrom ...symbols import POS, VERB, VerbForm_inf\nfrom ...tokens import Doc\nfrom ...vocab import Vocab\nfrom ...lemmatizer import Lemmatizer\n\n\n@pytest.fixture\ndef index():\n return {'verb': {}}\n\n@pytest.fixture\ndef exceptions():\n return {'verb': {}}\n\n@pytest.fixture\ndef rules():\n return {\"verb\": [[\"ed\", \"e\"]]}\n\n@pytest.fixture\ndef lemmatizer(index, exceptions, rules):\n return Lemmatizer(index, exceptions, rules)\n\n\n@pytest.fixture\ndef tag_map():\n return {'VB': {POS: VERB, 'morph': VerbForm_inf}}\n\n\n@pytest.fixture\ndef vocab(lemmatizer, tag_map):\n return Vocab(lemmatizer=lemmatizer, tag_map=tag_map)\n\n\ndef test_not_lemmatize_base_forms(vocab):\n doc = Doc(vocab, words=[\"Do\", \"n't\", \"feed\", \"the\", \"dog\"])\n feed = doc[2]\n feed.tag_ = u'VB'\n assert feed.text == u'feed'\n assert feed.lemma_ == u'feed'\n","subject":"Remove unnecessary argument in test","message":"Remove unnecessary argument in test\n","lang":"Python","license":"mit","repos":"oroszgy\/spaCy.hu,honnibal\/spaCy,banglakit\/spaCy,explosion\/spaCy,explosion\/spaCy,explosion\/spaCy,recognai\/spaCy,raphael0202\/spaCy,recognai\/spaCy,recognai\/spaCy,explosion\/spaCy,banglakit\/spaCy,Gregory-Howard\/spaCy,recognai\/spaCy,Gregory-Howard\/spaCy,raphael0202\/spaCy,honnibal\/spaCy,Gregory-Howard\/spaCy,oroszgy\/spaCy.hu,spacy-io\/spaCy,honnibal\/spaCy,banglakit\/spaCy,banglakit\/spaCy,spacy-io\/spaCy,raphael0202\/spaCy,raphael0202\/spaCy,Gregory-Howard\/spaCy,explosion\/spaCy,aikramer2\/spaCy,aikramer2\/spaCy,banglakit\/spaCy,aikramer2\/spaCy,aikramer2\/spaCy,Gregory-Howard\/spaCy,spacy-io\/spaCy,spacy-io\/spaCy,oroszgy\/spaCy.hu,spacy-io\/spaCy,explosion\/spaCy,oroszgy\/spaCy.hu,recognai\/spaCy,banglakit\/spaCy,raphael0202\/spaCy,recognai\/spaCy,aikramer2\/spaCy,spacy-io\/spaCy,oroszgy\/spaCy.hu,Gregory-Howard\/spaCy,aikramer2\/spaCy,honnibal\/spaCy,oroszgy\/spaCy.hu,raphael0202\/spaCy"} {"commit":"bca3c8f7b2c12b86e0d200009d23201bdc05d716","old_file":"make_spectra.py","new_file":"make_spectra.py","old_contents":"# -*- coding: utf-8 -*-\nimport randspectra as rs\nimport sys\nimport os.path as path\n\nsnapnum=sys.argv[1]\nsim=sys.argv[2]\n#base=\"\/n\/hernquistfs1\/mvogelsberger\/projects\/GFM\/Production\/Cosmo\/Cosmo\"+str(sim)+\"_V6\/L25n512\/output\/\"\n#savedir=\"\/n\/home11\/spb\/scratch\/Cosmo\/Cosmo\"+str(sim)+\"_V6_512\/snapdir_\"+str(snapnum).rjust(3,'0')\nbase=path.expanduser(\"~\/data\/Cosmo\/Cosmo\"+str(sim)+\"_V6\/L25n512\")\nhalo = rs.RandSpectra(snapnum, base)\nhalo.get_observer_tau(\"Si\",2)\nhalo.get_col_density(\"H\",1)\n#halo.get_tau(\"H\",1,1)\nhalo.get_col_density(\"Z\",-1)\nhalo.get_col_density(\"H\",-1)\nhalo.save_file()\n\n","new_contents":"# -*- coding: utf-8 -*-\nimport randspectra as rs\nimport sys\nimport os.path as path\n\nsnapnum=sys.argv[1]\nsim=sys.argv[2]\n#base=\"\/n\/hernquistfs1\/mvogelsberger\/projects\/GFM\/Production\/Cosmo\/Cosmo\"+str(sim)+\"_V6\/L25n512\/output\/\"\n#savedir=\"\/n\/home11\/spb\/scratch\/Cosmo\/Cosmo\"+str(sim)+\"_V6_512\/snapdir_\"+str(snapnum).rjust(3,'0')\nbase=path.expanduser(\"~\/data\/Cosmo\/Cosmo\"+str(sim)+\"_V6\/L25n512\")\nhalo = rs.RandSpectra(snapnum, base)\nhalo.save_file()\nhalo.get_observer_tau(\"Si\",2)\nhalo.get_col_density(\"H\",1)\n#halo.get_tau(\"H\",1,1)\nhalo.get_col_density(\"Z\",-1)\nhalo.get_col_density(\"H\",-1)\nhalo.save_file()\n\n","subject":"Handle the case where the savefile already exists by moving it out of the way","message":"Handle the case where the savefile already exists by moving it out of the way\n","lang":"Python","license":"mit","repos":"sbird\/vw_spectra"} {"commit":"bed7f5c80f6c5b5ce9b9a17aea5c9eadd047ee47","old_file":"mfr\/conftest.py","new_file":"mfr\/conftest.py","old_contents":"\"\"\"Project-wide test configuration, including fixutres that can be\nused by any module.\n\nExample test: ::\n\n def test_my_renderer(fakefile):\n assert my_renderer(fakefile) == '..expected result..'\n\n\"\"\"\nimport io\nimport pytest\n\n@pytest.fixture\ndef fakefile():\n return io.BytesIO(b'foo')\n","new_contents":"\"\"\"Project-wide test configuration, including fixutres that can be\nused by any module.\n\nExample test: ::\n\n def test_my_renderer(fakefile):\n assert my_renderer(fakefile) == '..expected result..'\n\n\"\"\"\nimport pytest\nimport mock\n\n@pytest.fixture\ndef fakefile():\n \"\"\"A simple file-like object.\"\"\"\n return mock.Mock(spec=file)\n","subject":"Make fakefile a mock instead of an io object","message":"Make fakefile a mock instead of an io object\n\nMakes it possible to mutate attributes, e.g. the name,\nfor tests\n","lang":"Python","license":"apache-2.0","repos":"felliott\/modular-file-renderer,CenterForOpenScience\/modular-file-renderer,mfraezz\/modular-file-renderer,CenterForOpenScience\/modular-file-renderer,Johnetordoff\/modular-file-renderer,mfraezz\/modular-file-renderer,icereval\/modular-file-renderer,icereval\/modular-file-renderer,AddisonSchiller\/modular-file-renderer,chrisseto\/modular-file-renderer,haoyuchen1992\/modular-file-renderer,erinspace\/modular-file-renderer,rdhyee\/modular-file-renderer,TomBaxter\/modular-file-renderer,rdhyee\/modular-file-renderer,CenterForOpenScience\/modular-file-renderer,Johnetordoff\/modular-file-renderer,felliott\/modular-file-renderer,erinspace\/modular-file-renderer,chrisseto\/modular-file-renderer,felliott\/modular-file-renderer,felliott\/modular-file-renderer,rdhyee\/modular-file-renderer,haoyuchen1992\/modular-file-renderer,chrisseto\/modular-file-renderer,CenterForOpenScience\/modular-file-renderer,mfraezz\/modular-file-renderer,AddisonSchiller\/modular-file-renderer,haoyuchen1992\/modular-file-renderer,haoyuchen1992\/modular-file-renderer,mfraezz\/modular-file-renderer,TomBaxter\/modular-file-renderer,rdhyee\/modular-file-renderer,erinspace\/modular-file-renderer,TomBaxter\/modular-file-renderer,icereval\/modular-file-renderer,AddisonSchiller\/modular-file-renderer,Johnetordoff\/modular-file-renderer,TomBaxter\/modular-file-renderer,AddisonSchiller\/modular-file-renderer,Johnetordoff\/modular-file-renderer"} {"commit":"3770095f087309efe901c2f22afd29ba6f3ddd18","old_file":"comrade\/core\/context_processors.py","new_file":"comrade\/core\/context_processors.py","old_contents":"from django.conf import settings\nfrom django.contrib.sites.models import Site\nfrom settings import DeploymentType\n\ndef default(request):\n context = {}\n context['DEPLOYMENT'] = settings.DEPLOYMENT\n context['current_site'] = Site.objects.get_current()\n if settings.DEPLOYMENT != DeploymentType.PRODUCTION:\n context['GIT_COMMIT'] = settings.GIT_COMMIT\n context['site_email'] = settings.CONTACT_EMAIL\n if request.is_secure():\n context['protocol'] = 'https:\/\/'\n else:\n context['protocol'] = 'http:\/\/'\n context['current_site_url'] = (context['protocol'] +\n context['current_site'].domain)\n return context\n\ndef ssl_media(request):\n if request.is_secure():\n ssl_media_url = settings.MEDIA_URL.replace('http:\/\/','https:\/\/')\n else:\n ssl_media_url = settings.MEDIA_URL\n return {'MEDIA_URL': ssl_media_url}\n","new_contents":"from django.conf import settings\nfrom django.contrib.sites.models import Site\nfrom settings import DeploymentType\n\ndef default(request):\n context = {}\n context['DEPLOYMENT'] = settings.DEPLOYMENT\n context['current_site'] = Site.objects.get_current()\n if settings.DEPLOYMENT != DeploymentType.PRODUCTION:\n context['GIT_COMMIT'] = settings.GIT_COMMIT\n context['site_email'] = settings.CONTACT_EMAIL\n if request.is_secure():\n context['protocol'] = 'https:\/\/'\n else:\n context['protocol'] = 'http:\/\/'\n context['current_site_url'] = (context['protocol'] +\n context['current_site'].domain)\n return context\n\ndef profile(request):\n context = {}\n if request.user.is_authenticated():\n context['profile'] = request.user.get_profile()\n return context\n\ndef ssl_media(request):\n if request.is_secure():\n ssl_media_url = settings.MEDIA_URL.replace('http:\/\/','https:\/\/')\n else:\n ssl_media_url = settings.MEDIA_URL\n return {'MEDIA_URL': ssl_media_url}\n","subject":"Add a context processor that adds the UserProfile to each context.","message":"Add a context processor that adds the UserProfile to each context.\n","lang":"Python","license":"mit","repos":"bueda\/django-comrade"} {"commit":"921e315e61355d80caea673ce09f8944388d86e2","old_file":"tests\/unit\/util\/test_cache.py","new_file":"tests\/unit\/util\/test_cache.py","old_contents":"\"\"\"Test praw.util.cache.\"\"\"\n\nfrom .. import UnitTest\n\nfrom praw.util.cache import cachedproperty\n\n\nclass TestCachedProperty(UnitTest):\n class Klass:\n @cachedproperty\n def nine(self):\n \"\"\"Return 9.\"\"\"\n return 9\n\n def ten(self):\n return 10\n\n ten = cachedproperty(ten, doc=\"Return 10.\")\n\n def test_get(self):\n klass = self.Klass()\n assert \"nine\" not in klass.__dict__\n assert klass.nine == 9\n assert \"nine\" in klass.__dict__\n\n def test_repr(self):\n klass = self.Klass()\n assert repr(klass.nine) == \"9\"\n\n property_repr = repr(self.Klass.nine)\n assert property_repr.startswith(\"BHB', len(chunk), addr & 0xffff, 0) + chunk\n checksum = (-(sum(data))) & 0xff\n hexline = ':%s%02X' % (str(binascii.hexlify(chunk), 'utf8').upper(), checksum)\n print(hexline)\n addr += 16\n","new_contents":"'''\nTurn a Python script into Intel HEX format to be concatenated at the\nend of the MicroPython firmware.hex. A simple header is added to the\nscript.\n'''\n\nimport sys\nimport struct\nimport binascii\n\n# read script body\nwith open(sys.argv[1], \"rb\") as f:\n data = f.read()\n\n# add header, pad to multiple of 16 bytes\ndata = b'MP' + struct.pack('BHB', len(chunk), addr & 0xffff, 0) + chunk\n checksum = (-(sum(chunk))) & 0xff\n hexline = ':%s%02X' % (str(binascii.hexlify(chunk), 'utf8').upper(), checksum)\n print(hexline)\n addr += 16\n","subject":"Fix bug in heexlifyscript.py when computing checksum.","message":"Fix bug in heexlifyscript.py when computing checksum.\n","lang":"Python","license":"mit","repos":"JoeGlancy\/micropython,JoeGlancy\/micropython,JoeGlancy\/micropython"} {"commit":"e40ef4cbe59c5c3d064e60f02f60f19b0bb202a4","old_file":"test_daily_parser.py","new_file":"test_daily_parser.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: latin-1 -*-\n\n\"\"\"Unit tests.\"\"\"\n\nimport unittest\nfrom daily_parser import url_from_args\n\n\nclass TestDailyParser(unittest.TestCase):\n\n \"\"\"Testing methods from daily_parser.\"\"\"\n\n def test_url_from_args(self):\n output = url_from_args(2014, 1)\n expected = 'https:\/\/dons.wikimedia.fr\/journal\/2014-01'\n self.assertEqual(output, expected)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: latin-1 -*-\n\n\"\"\"Unit tests.\"\"\"\n\nimport unittest\nfrom daily_parser import url_from_args, DonationsParser\n\n\nclass TestDailyParser(unittest.TestCase):\n\n \"\"\"Testing methods from daily_parser.\"\"\"\n\n def test_url_from_args(self):\n output = url_from_args(2014, 1)\n expected = 'https:\/\/dons.wikimedia.fr\/journal\/2014-01'\n self.assertEqual(output, expected)\n\n\nclass TestDonationsParser(unittest.TestCase):\n\n \"\"\"Testing DonationsParser class.\"\"\"\n\n def setUp(self):\n self.donations_parser = DonationsParser(2014, 01)\n donations_data = {\n '01': {'sum': 370, 'avg': 46.25, 'quantity': 8},\n '02': {'sum': 5682, 'avg': 132.14, 'quantity': 43}\n }\n self.donations_parser.donations = donations_data\n\n def test_get_csv(self):\n expected = \"\"\"'day', 'sum', 'quantity', 'avg'\n'2014-01-01', 370, 8, 46.25\n'2014-01-02', 5682, 43, 132.14\n\"\"\"\n output = self.donations_parser.get_csv()\n self.assertEqual(output, expected)\n","subject":"Add unit test for DonationsParser.get_csv","message":"Add unit test for DonationsParser.get_csv\n","lang":"Python","license":"mit","repos":"Commonists\/DonationsLogParser,Commonists\/DonationsLogParser"} {"commit":"2377f500d4667623da9a2921c62862b00d7f404c","old_file":"school\/frontend\/views.py","new_file":"school\/frontend\/views.py","old_contents":"from flask import Blueprint, render_template, url_for, redirect, flash\nfrom flask.ext.login import login_required, logout_user, current_user, login_user\nfrom .forms import LoginForm\nfrom school.config import FLASH_SUCCESS, FLASH_INFO, FLASH_WARNING\n\nfrontend = Blueprint('frontend', __name__)\n\n\n@frontend.route('\/login', methods=[\"GET\", \"POST\"])\ndef login():\n if current_user.is_authenticated(): # user is already logged in\n flash(\"You are already logged in\", FLASH_WARNING)\n return redirect(url_for('user.index'))\n\n form = LoginForm()\n if form.validate_on_submit():\n flash('Successfully logged in as %s' % form.user.username, FLASH_SUCCESS)\n login_user(form.user)\n\n return form.redirect(\"user.index\")\n\n return render_template('frontend\/index.html', form=form)\n\n\n@frontend.route('\/logout')\n@login_required\ndef logout():\n logout_user()\n flash('You have been logged out.', FLASH_INFO)\n\n return redirect(url_for(\"frontend.login\"))","new_contents":"from flask import Blueprint, render_template, url_for, redirect, flash\nfrom flask.ext.login import login_required, logout_user, current_user, login_user\nfrom .forms import LoginForm\nfrom school.config import FLASH_SUCCESS, FLASH_INFO, FLASH_WARNING\n\nfrontend = Blueprint('frontend', __name__)\n\n\n@frontend.route('\/login', methods=[\"GET\", \"POST\"])\ndef login():\n if current_user.is_authenticated(): # user is already logged in\n flash(\"You are already logged in\", FLASH_WARNING)\n return redirect(url_for('user.index'))\n\n form = LoginForm()\n if form.validate_on_submit():\n login_user(form.user)\n\n return form.redirect(\"user.index\")\n\n return render_template('frontend\/index.html', form=form)\n\n\n@frontend.route('\/logout')\n@login_required\ndef logout():\n logout_user()\n flash('You have been logged out.', FLASH_INFO)\n\n return redirect(url_for(\"frontend.login\"))","subject":"Remove flash success message when logging in.","message":"Remove flash success message when logging in.\n","lang":"Python","license":"mit","repos":"leyyin\/university-SE,leyyin\/university-SE,leyyin\/university-SE"} {"commit":"7aedc2151035174632a7f3e55be7563f71e65117","old_file":"tests\/audio\/test_loading.py","new_file":"tests\/audio\/test_loading.py","old_contents":"import pytest\n\n@pytest.mark.xfail\ndef test_missing_file(audiomgr):\n sound = audiomgr.get_sound('\/not\/a\/valid\/file.ogg')\n assert sound is None\n","new_contents":"import pytest\n\ndef test_missing_file(audiomgr):\n sound = audiomgr.get_sound('\/not\/a\/valid\/file.ogg')\n assert str(sound).startswith('NullAudioSound')\n","subject":"Update audio test to recognize missing sounds as NullAudioSound","message":"tests: Update audio test to recognize missing sounds as NullAudioSound\n","lang":"Python","license":"bsd-3-clause","repos":"chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d,chandler14362\/panda3d"} {"commit":"9c5349595dca8013f1353785bbd34fb3d7cd4a6a","old_file":"misc\/__init__.py","new_file":"misc\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\nimport logging\n\n__version__ = VERSION = '0.0.1'\n__project__ = PROJECT = 'django-misc'\n\nlog = logging.getLogger( __name__ )\n\n\n\n","new_contents":"# -*- coding: utf-8 -*-\nimport logging\n\n__version__ = VERSION = '0.0.2'\n__project__ = PROJECT = 'django-misc'\n\nlog = logging.getLogger( __name__ )\n\n\n\n","subject":"Change version to 0.0.2 to update pypi repo","message":"Change version to 0.0.2 to update pypi repo\n","lang":"Python","license":"mit","repos":"ilblackdragon\/django-misc"} {"commit":"66c50cdeda974f2159259b466995339244ffb694","old_file":"training\/level-2-command-line-interfaces\/dragon-warrior\/tmarsha1\/primes\/Tests\/PrimeFinderTests.py","new_file":"training\/level-2-command-line-interfaces\/dragon-warrior\/tmarsha1\/primes\/Tests\/PrimeFinderTests.py","old_contents":"\"\"\"\nTest the Prime Finder class\n\nStill working on getting dependency injection working.\n\"\"\"\n\nimport unittest\nfrom primes.Primes import PrimeFinder\n#from primes.Primes import PrimeGenerator\n\nclass PrimeFinderTests(unittest.TestCase):\n def test_find_prime(self):\n prime_finder = PrimeFinder.PrimeFinder(PrimeGenerator.PrimeGenerator())\n self.assertEqual(prime_finder.find_prime(6), 13)\n","new_contents":"\"\"\"\nTest the Prime Finder class\n\nStill working on getting dependency injection working.\nInjecting the Generator into the Finder allows for many possibilities.\nFrom the testing perspective this would allow me to inject a mock object\nfor the Generator that returns a set value speeding up the testing of the\nPrime Finder class.\n\"\"\"\n\nimport unittest\nfrom primes.Primes import PrimeFinder\n#from primes.Primes import PrimeGenerator\n\nclass PrimeFinderTests(unittest.TestCase):\n def test_find_prime(self):\n prime_finder = PrimeFinder.PrimeFinder(PrimeGenerator.PrimeGenerator())\n self.assertEqual(prime_finder.find_prime(6), 13)\n","subject":"Add additional comments regarding Dependency Injection","message":"Add additional comments regarding Dependency Injection\n","lang":"Python","license":"artistic-2.0","repos":"bigfatpanda-training\/pandas-practical-python-primer,bigfatpanda-training\/pandas-practical-python-primer"} {"commit":"7a7729e9af8e91411526525c19c5d434609e0f21","old_file":"logger.py","new_file":"logger.py","old_contents":"MSG_INFO = 0x01\nMSG_WARNING = 0x02\nMSG_ERROR = 0x04\nMSG_VERBOSE = 0x08\nMSG_ALL = MSG_INFO | MSG_WARNING | MSG_ERROR | MSG_VERBOSE\n\ndef logi(msg):\n print(\"[INFO] \" + msg)\ndef logv(msg):\n print(\"[VERBOSE] \" + msg)\ndef logw(msg):\n print(\"[WARNING] \" + msg)\ndef loge(msg):\n print(\"[ERROR] \" + msg)\n\nclass Logger(object):\n def __init__(self):\n self.logger_level = MSG_ALL\n def info(self, msg):\n if self.logger_level & MSG_INFO:\n logi(msg)\n def warning(self, msg):\n if self.logger_level & MSG_WARNING:\n logw(msg)\n def error(self, msg):\n if self.logger_level & MSG_ERROR:\n loge(msg)\n def verbose(self, msg):\n if self.logger_level & MSG_VERBOSE:\n logv(msg)\n","new_contents":"MSG_INFO = 0x01\nMSG_WARNING = 0x02\nMSG_ERROR = 0x04\nMSG_VERBOSE = 0x08\nMSG_ALL = MSG_INFO | MSG_WARNING | MSG_ERROR | MSG_VERBOSE\n\ndef logi(msg):\n print(\"[INFO] \" + msg)\ndef logv(msg):\n print(\"[VERBOSE] \" + msg)\ndef logw(msg):\n print(\"[WARNING] \" + msg)\ndef loge(msg):\n print(\"\\033[1;31m[ERROR] \" + msg + \"\\033[m\")\n\nclass Logger(object):\n def __init__(self):\n self.logger_level = MSG_ALL\n def info(self, msg):\n if self.logger_level & MSG_INFO:\n logi(msg)\n def warning(self, msg):\n if self.logger_level & MSG_WARNING:\n logw(msg)\n def error(self, msg):\n if self.logger_level & MSG_ERROR:\n loge(msg)\n def verbose(self, msg):\n if self.logger_level & MSG_VERBOSE:\n logv(msg)\n","subject":"Add color for error message.","message":"Add color for error message.\n","lang":"Python","license":"mit","repos":"PyOCL\/oclGA,PyOCL\/OpenCLGA,PyOCL\/OpenCLGA,PyOCL\/oclGA,PyOCL\/oclGA,PyOCL\/TSP,PyOCL\/TSP,PyOCL\/oclGA,PyOCL\/OpenCLGA"} {"commit":"d3b326e421ca482723cafcadd2442eebb8cf2ee6","old_file":"bot\/main.py","new_file":"bot\/main.py","old_contents":"import os\nimport json\nimport time\nimport socket\nfrom subprocess import call\n\nimport requests\n\nfrom test import test_activity\nfrom build import compile_bundle\n\n# Fixes a weird bug... it might create some though :P\nos.environ['http_proxy'] = ''\n\n# HOST = 'http:\/\/localhost:5001'\nHOST = 'http:\/\/aslo-bot-master.sugarlabs.org'\n\nprint 'Waiting for 1st task'\n\nwhile True:\n r = requests.get(HOST + '\/task')\n if r.status_code == 404:\n time.sleep(7)\n continue\n task = r.json()\n\n print 'Got new task'\n call(['git', 'clone', 'https:\/\/www.github.com\/' + task['gh'],\n 'dl'])\n result = test_activity(task['bundle_id'], task['gh'])\n\n data = {'result': result, 'file': compile_bundle(),\n 'bundle_id': task['bundle_id'], 'task_id': task['task_id']}\n\n headers = {'Content-type': 'application\/json', 'Accept': 'text\/plain'}\n r = requests.post(HOST + '\/done',\n data=json.dumps(data), headers=headers)\n\n call(['rm', '-rf', 'dl'])\n\n print 'Mined 1 activity:', task['bundle_id'], task['gh']\n","new_contents":"import os\nimport json\nimport time\nimport socket\nfrom subprocess import call\n\nimport requests\n\nfrom test import test_activity\nfrom build import compile_bundle\n\n# Fixes a weird bug... it might create some though :P\nos.environ['http_proxy'] = ''\n\n# HOST = 'http:\/\/localhost:5001'\nHOST = 'http:\/\/aslo-bot-master.sugarlabs.org'\n\nprint 'Waiting for 1st task'\n\nwhile True:\n try:\n r = requests.get(HOST + '\/task')\n except requests.exceptions.ConnectionError, e:\n continue\n if r.status_code == 404:\n time.sleep(7)\n continue\n task = r.json()\n\n print 'Got new task'\n call(['git', 'clone', 'https:\/\/www.github.com\/' + task['gh'],\n 'dl'])\n result = test_activity(task['bundle_id'], task['gh'])\n\n data = {'result': result, 'file': compile_bundle(),\n 'bundle_id': task['bundle_id'], 'task_id': task['task_id']}\n\n headers = {'Content-type': 'application\/json', 'Accept': 'text\/plain'}\n r = requests.post(HOST + '\/done',\n data=json.dumps(data), headers=headers)\n\n call(['rm', '-rf', 'dl'])\n\n print 'Mined 1 activity:', task['bundle_id'], task['gh']\n","subject":"Stop bots from randomly disconnecting","message":"Stop bots from randomly disconnecting\n","lang":"Python","license":"agpl-3.0","repos":"samdroid-apps\/aslo,samdroid-apps\/aslo,samdroid-apps\/aslo"} {"commit":"fc50a002e967f8e3b7de205a866e010dda717962","old_file":"logger.py","new_file":"logger.py","old_contents":"import os\nimport logging\n\ndef logger_setup(name):\n loglevel = ''\n try:\n if os.environ['PB_LOGLEVEL'] == 'DEBUG':\n loglevel = logging.DEBUG\n\n if os.environ['PB_LOGLEVEL'] == 'INFO':\n loglevel = logging.INFO\n\n if os.environ['PB_LOGLEVEL'] == 'WARN':\n loglevel = logging.WARN\n except:\n loglevel = logging.DEBUG\n\n # from https:\/\/docs.python.org\/2\/howto\/logging.html#configuring-logging\n # set up new logger for this file\n logger = logging.getLogger(name)\n logger.setLevel(loglevel)\n\n # console handler for logging\n conLog = logging.StreamHandler()\n conLog.setLevel(loglevel)\n\n # formatter\n formatter = logging.Formatter('%(asctime)s - %(name)s - %(levelname)s - %(message)s')\n\n # format console logs using formatter\n conLog.setFormatter(formatter)\n\n # add console logging transport to logger\n logger.addHandler(conLog)\n\n return logger\n","new_contents":"import os\nimport logging\n\ndef logger_setup(name):\n loglevel = ''\n try:\n if os.environ['PB_LOGLEVEL'] == 'DEBUG':\n loglevel = logging.DEBUG\n\n if os.environ['PB_LOGLEVEL'] == 'INFO':\n loglevel = logging.INFO\n\n if os.environ['PB_LOGLEVEL'] == 'WARN':\n loglevel = logging.WARN\n except:\n loglevel = logging.DEBUG\n\n # from https:\/\/docs.python.org\/2\/howto\/logging.html#configuring-logging\n # set up new logger for this file\n logger = logging.getLogger(name)\n logger.setLevel(loglevel)\n\n # formatter\n formatter = logging.Formatter('PID: %(process)d - %(asctime)s - %(name)s - %(levelname)s - %(funcName)s - %(message)s')\n\n # console handler for logging\n conLog = logging.StreamHandler()\n conLog.setLevel(loglevel)\n # format console logs using formatter\n conLog.setFormatter(formatter)\n\n # log to file handler\n fileLog = logging.FileHandler('pairing-bot.log', encoding='utf-8')\n fileLog.setLevel(logging.DEBUG)\n # format console logs using formatter\n fileLog.setFormatter(formatter)\n\n # add console logging transport to logger\n logger.addHandler(conLog)\n\n # add file transport to logger\n logger.addHandler(fileLog)\n\n return logger\n","subject":"Add file transport and update log msg formatting to include pid and fn name","message":"Add file transport and update log msg formatting to include pid and fn name\n","lang":"Python","license":"mit","repos":"zeusdeux\/zulip-pairing-bot"} {"commit":"4f71339cad35b2444ea295fd4b518e539f1088bb","old_file":"fluent_faq\/urls.py","new_file":"fluent_faq\/urls.py","old_contents":"from django.conf.urls import patterns, url\nfrom .views import FaqQuestionList, FaqCategoryDetail, FaqQuestionDetail\n\nurlpatterns = patterns('',\n url(r'^$', FaqQuestionList.as_view(), name='faqquestion_index'),\n url(r'^(?P[^\/]+)\/$', FaqCategoryDetail.as_view(), name='faqcategory_detail'),\n url(r'^(?P[^\/]+)\/(?P[^\/]+)\/$', FaqQuestionDetail.as_view(), name='faqquestion_detail'),\n)\n","new_contents":"from django.conf.urls import url\nfrom .views import FaqQuestionList, FaqCategoryDetail, FaqQuestionDetail\n\nurlpatterns = [\n url(r'^$', FaqQuestionList.as_view(), name='faqquestion_index'),\n url(r'^(?P[^\/]+)\/$', FaqCategoryDetail.as_view(), name='faqcategory_detail'),\n url(r'^(?P[^\/]+)\/(?P[^\/]+)\/$', FaqQuestionDetail.as_view(), name='faqquestion_detail'),\n]\n","subject":"Fix Django 1.9 warnings about patterns('', ..)","message":"Fix Django 1.9 warnings about patterns('', ..)\n","lang":"Python","license":"apache-2.0","repos":"edoburu\/django-fluent-faq,edoburu\/django-fluent-faq"} {"commit":"f113123a3f31e176ae7165f1ca11118dc00625a3","old_file":"tests\/test_backend_forms.py","new_file":"tests\/test_backend_forms.py","old_contents":"import floppyforms.__future__ as floppyforms\nfrom django.test import TestCase\n\nfrom django_backend.backend.base.forms import BaseBackendForm\n\nfrom .models import OneFieldModel\n\n\nclass OneFieldForm(BaseBackendForm):\n class Meta:\n model = OneFieldModel\n exclude = ()\n\n\nclass BaseBackendFormTests(TestCase):\n def test_has_superform_metaclass(self):\n from django_superform.forms import SuperModelFormMetaclass\n\n self.assertTrue(\n issubclass(BaseBackendForm.__metaclass__, SuperModelFormMetaclass))\n\n def test_has_floppyforms_metaclass(self):\n from floppyforms.__future__.models import ModelFormMetaclass\n\n self.assertTrue(\n issubclass(BaseBackendForm.__metaclass__, ModelFormMetaclass))\n\n def test_model_field_is_using_floppyforms_widget(self):\n form = OneFieldForm()\n self.assertTrue(\n isinstance(form.fields['chars'].widget, floppyforms.TextInput))\n","new_contents":"import floppyforms.__future__ as floppyforms\nfrom django.test import TestCase\n\nfrom django_backend.forms import BaseBackendForm\n\nfrom .models import OneFieldModel\n\n\nclass OneFieldForm(BaseBackendForm):\n class Meta:\n model = OneFieldModel\n exclude = ()\n\n\nclass BaseBackendFormTests(TestCase):\n def test_has_superform_metaclass(self):\n from django_superform.forms import SuperModelFormMetaclass\n\n self.assertTrue(\n issubclass(BaseBackendForm.__metaclass__, SuperModelFormMetaclass))\n\n def test_has_floppyforms_metaclass(self):\n from floppyforms.__future__.models import ModelFormMetaclass\n\n self.assertTrue(\n issubclass(BaseBackendForm.__metaclass__, ModelFormMetaclass))\n\n def test_model_field_is_using_floppyforms_widget(self):\n form = OneFieldForm()\n self.assertTrue(\n isinstance(form.fields['chars'].widget, floppyforms.TextInput))\n","subject":"Fix forms import in tests","message":"Fix forms import in tests\n","lang":"Python","license":"bsd-3-clause","repos":"team23\/django_backend,team23\/django_backend,team23\/django_backend,team23\/django_backend,team23\/django_backend"} {"commit":"e291ae29926a3cd05c9268c625e14d205638dfe8","old_file":"jarn\/mkrelease\/scp.py","new_file":"jarn\/mkrelease\/scp.py","old_contents":"import tempfile\nimport tee\n\nfrom process import Process\nfrom exit import err_exit\n\n\nclass SCP(object):\n \"\"\"Secure copy and FTP abstraction.\"\"\"\n\n def __init__(self, process=None):\n self.process = process or Process()\n\n def run_scp(self, distfile, location):\n if not self.process.quiet:\n print 'scp-ing to %(location)s' % locals()\n rc = self.process.os_system(\n 'scp \"%(distfile)s\" \"%(location)s\"' % locals())\n if rc != 0:\n err_exit('scp failed')\n return rc\n\n def run_sftp(self, distfile, location):\n if not self.process.quiet:\n print 'sftp-ing to %(location)s' % locals()\n with tempfile.NamedTemporaryFile(prefix='sftp-') as file:\n file.write('put \"%(distfile)s\"\\n' % locals())\n file.write('quit\\n')\n file.flush()\n cmdfile = file.name\n rc, lines = self.process.popen(\n 'sftp -b \"%(cmdfile)s\" \"%(location)s\"' % locals(),\n echo=tee.StartsWith('Uploading'))\n if rc != 0:\n err_exit('sftp failed')\n return rc\n\n","new_contents":"import tempfile\nimport tee\n\nfrom os.path import split\n\nfrom process import Process\nfrom chdir import ChdirStack\nfrom exit import err_exit\n\n\nclass SCP(object):\n \"\"\"Secure copy and FTP abstraction.\"\"\"\n\n def __init__(self, process=None):\n self.process = process or Process()\n self.dirstack = ChdirStack()\n\n def run_scp(self, distfile, location):\n if not self.process.quiet:\n print 'scp-ing to %(location)s' % locals()\n rc = self.process.os_system(\n 'scp \"%(distfile)s\" \"%(location)s\"' % locals())\n if rc != 0:\n err_exit('scp failed')\n return rc\n\n def run_sftp(self, distfile, location):\n if not self.process.quiet:\n print 'sftp-ing to %(location)s' % locals()\n dir, distfile = split(distfile)\n self.dirstack.push(dir)\n try:\n with tempfile.NamedTemporaryFile(prefix='sftp-') as file:\n file.write('put \"%(distfile)s\"\\n' % locals())\n file.write('bye\\n')\n file.flush()\n cmdfile = file.name\n rc, lines = self.process.popen(\n 'sftp -b \"%(cmdfile)s\" \"%(location)s\"' % locals(),\n echo=tee.StartsWith('Uploading'))\n if rc != 0:\n err_exit('sftp failed')\n return rc\n finally:\n self.dirstack.pop()\n\n","subject":"Change to dist dir before uploading to lose the absolute path.","message":"Change to dist dir before uploading to lose the absolute path.\n","lang":"Python","license":"bsd-2-clause","repos":"Jarn\/jarn.mkrelease"} {"commit":"6b9ccae880e9582f38e2a8aa3c451bc6f6a88d37","old_file":"thing\/tasks\/tablecleaner.py","new_file":"thing\/tasks\/tablecleaner.py","old_contents":"import datetime\n\nfrom celery import task\nfrom celery.utils.log import get_task_logger\nlogger = get_task_logger(__name__)\n\nfrom django.db.models import Q\n\nfrom thing.models import APIKey, TaskState\n\n# ---------------------------------------------------------------------------\n# Periodic task to perform database table cleanup\n@task(name='thing.table_cleaner')\ndef table_cleaner():\n utcnow = datetime.datetime.utcnow()\n\n queued_timeout = utcnow - datetime.timedelta(minutes=120)\n\n # Build a QuerySet to find broken tasks\n taskstates = TaskState.objects.filter(state=TaskState.QUEUED_STATE, mod_time__lte=queued_timeout)\n for ts in taskstates:\n logger.warn('[table_cleaner] Stuck task: %d | %d | %s | %s', ts.id, ts.keyid, ts.parameter, ts.url)\n \n count = taskstates.update(mod_time=utcnow, next_time=utcnow, state=TaskState.READY_STATE)\n if count > 0:\n logger.warn('[table_cleaner] Reset %d broken task(s)', count)\n\n # Build a QuerySet to find tasks that refer to no longer existent keys\n taskstates = TaskState.objects.exclude(\n Q(keyid=-1)\n |\n Q(keyid__in=APIKey.objects.values('keyid'))\n )\n taskstates.delete()\n\n# ---------------------------------------------------------------------------\n","new_contents":"import datetime\n\nfrom celery import task\nfrom celery.utils.log import get_task_logger\nlogger = get_task_logger(__name__)\n\nfrom django.db.models import Q\n\nfrom thing.models import APIKey, TaskState\n\n# ---------------------------------------------------------------------------\n# Periodic task to perform database table cleanup\n@task(name='thing.table_cleaner')\ndef table_cleaner():\n utcnow = datetime.datetime.utcnow()\n\n queued_timeout = utcnow - datetime.timedelta(minutes=120)\n\n # Build a QuerySet to find broken tasks\n taskstates = TaskState.objects.filter(state=TaskState.QUEUED_STATE, mod_time__lte=queued_timeout)\n for ts in taskstates:\n logger.warn('[table_cleaner] Stuck task: %d | %d | %s | %s', ts.id, ts.keyid, ts.parameter, ts.url)\n\n count = taskstates.update(mod_time=utcnow, next_time=utcnow, state=TaskState.READY_STATE)\n if count > 0:\n logger.warn('[table_cleaner] Reset %d broken task(s)', count)\n\n # Build a QuerySet to find tasks that refer to no longer existent keys\n taskstates = TaskState.objects.exclude(\n Q(keyid=-1)\n |\n Q(keyid__in=APIKey.objects.filter(valid=True).values('keyid'))\n )\n taskstates.delete()\n\n# ---------------------------------------------------------------------------\n","subject":"Change thing.tasks.table_cleaner to delete TaskState objects for any invalid APIKeys","message":"Change thing.tasks.table_cleaner to delete TaskState objects for any invalid APIKeys\n","lang":"Python","license":"bsd-2-clause","repos":"madcowfred\/evething,madcowfred\/evething,Gillingham\/evething,Gillingham\/evething,cmptrgeekken\/evething,madcowfred\/evething,Gillingham\/evething,cmptrgeekken\/evething,cmptrgeekken\/evething,Gillingham\/evething,madcowfred\/evething,cmptrgeekken\/evething,cmptrgeekken\/evething"} {"commit":"e1c970d76dbd0eb631e726e101b09c0f5e5599ec","old_file":"doc\/api_changes\/2015-04-27-core.py","new_file":"doc\/api_changes\/2015-04-27-core.py","old_contents":"Grid-building functions\n-----------------------\n\n:func:`pixels_to_radius` and :func:`pixels_to_phi` were renamed to\n:func:`radial_grid` and :func:`angle_grid` respectively. The name and order\nof their arguments was also changed: see their docstring or API docs for\ndetails.\n","new_contents":"Grid-building functions\n-----------------------\n\n:func:`pixels_to_radius` and :func:`pixels_to_phi` were renamed to\n:func:`radial_grid` and :func:`angle_grid` respectively. The name and order\nof their arguments was also changed: see their docstring or API docs for\ndetails. Importantly, the orientation of the output of angle grid has been\nchanged. Again, see the docstring for details.\n","subject":"Add change in angle_grid orientation to API changes.","message":"DOC: Add change in angle_grid orientation to API changes.\n","lang":"Python","license":"bsd-3-clause","repos":"licode\/scikit-xray,Nikea\/scikit-xray,hainm\/scikit-xray,celiafish\/scikit-xray,licode\/scikit-beam,CJ-Wright\/scikit-beam,CJ-Wright\/scikit-beam,tacaswell\/scikit-xray,giltis\/scikit-xray,ericdill\/scikit-xray,scikit-xray\/scikit-xray,yugangzhang\/scikit-beam,scikit-xray\/scikit-xray,giltis\/scikit-xray,danielballan\/scikit-xray,tacaswell\/scikit-beam,tacaswell\/scikit-xray,licode\/scikit-beam,danielballan\/scikit-xray,celiafish\/scikit-xray,Nikea\/scikit-xray,Nikea\/scikit-xray,celiafish\/scikit-xray,yugangzhang\/scikit-beam,licode\/scikit-xray,licode\/scikit-beam,danielballan\/scikit-xray,hainm\/scikit-xray,hainm\/scikit-xray,yugangzhang\/scikit-beam,tacaswell\/scikit-xray,giltis\/scikit-xray,licode\/scikit-xray,tacaswell\/scikit-beam,ericdill\/scikit-xray,tacaswell\/scikit-beam,scikit-xray\/scikit-xray,CJ-Wright\/scikit-beam,ericdill\/scikit-xray"} {"commit":"6974cba56413527c8b7cef9e4b6ad6ca9fe5049e","old_file":"tests\/test_memory.py","new_file":"tests\/test_memory.py","old_contents":"# coding: utf-8\nfrom unittest import TestCase\nfrom chipy8 import Memory\n\n\nclass TestMemory(TestCase):\n def setUp(self):\n self.memory = Memory()\n\n def test_write(self):\n 'Write a byte to memory then read it.'\n address = 0x200\n self.memory.write_byte(0x200, 0x01)\n self.assertEqual(0x01, self.memory.read_byte(0x200))\n\n def test_load(self):\n 'Load a stream of bytes to memory starting on an address.'\n address = 0x200\n self.memory.load(0x200, [0x01, 0x02, 0x03])\n self.assertEqual(0x01, self.memory.read_byte(address))\n self.assertEqual(0x02, self.memory.read_byte(address + 1))\n self.assertEqual(0x03, self.memory.read_byte(address + 2))\n","new_contents":"# coding: utf-8\nfrom unittest import TestCase\nfrom chipy8 import Memory\n\n\nclass TestMemory(TestCase):\n def setUp(self):\n self.memory = Memory()\n\n def test_write(self):\n 'Write a byte to memory then read it.'\n address = 0x200\n self.memory.write_byte(address, 0x01)\n self.assertEqual(0x01, self.memory.read_byte(address))\n\n def test_load(self):\n 'Load a stream of bytes to memory starting on an address.'\n address = 0x200\n self.memory.load(address, [0x01, 0x02, 0x03])\n self.assertEqual(0x01, self.memory.read_byte(address))\n self.assertEqual(0x02, self.memory.read_byte(address + 1))\n self.assertEqual(0x03, self.memory.read_byte(address + 2))\n","subject":"Clarify values used in tests.","message":"Clarify values used in tests.","lang":"Python","license":"bsd-3-clause","repos":"gutomaia\/chipy8"} {"commit":"ff489b1541f896025a0c630be6abe2d23843ec36","old_file":"examples\/05_alternative_language.py","new_file":"examples\/05_alternative_language.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom pyhmsa.datafile import DataFile\nfrom pyhmsa.type.language import langstr\ndatafile = DataFile()\n\nauthor = langstr('Fyodor Dostoyevsky', {'ru': u'Фёдор Миха́йлович Достое́вский'})\ndatafile.header.author = author\n\nprint(datafile.header.author.alternatives['ru']) # Returns ...","new_contents":"#!\/usr\/bin\/env python\n\nfrom pyhmsa.datafile import DataFile\nfrom pyhmsa.type.language import langstr\ndatafile = DataFile()\n\nauthor = langstr('Wilhelm Conrad Roentgen', {'de': u'Wilhelm Conrad Röntgen'})\ndatafile.header.author = author\n\nprint(datafile.header.author.alternatives['de']) # Returns ...","subject":"Replace name in alternative language to prevent compilation problems with LaTeX","message":"Replace name in alternative language to prevent compilation problems\nwith LaTeX","lang":"Python","license":"mit","repos":"pyhmsa\/pyhmsa"} {"commit":"8a95648d03453a4e612f54adf8c6142b1a9e7ad9","old_file":"src\/gh2md\/__init__.py","new_file":"src\/gh2md\/__init__.py","old_contents":"__version__ = \"0.1.3\"\n\n__title__ = \"gh2md\"\n__description__ = \"Export Github repository issues to markdown files: https:\/\/github.com\/mattduck\/gh2md.\"\n__doc__ = __description__\n__uri__ = \"https:\/\/github.com\/mattduck\/gh2md\"\n\n__author__ = \"Matthew Duck\"\n__email__ = \"matt@mattduck.com\"\n\n__license__ = \"MIT\"\n__copyright__ = \"Copyright (c) 2017 Matthew Duck\"\n","new_contents":"__version__ = \"0.1.4dev\"\n\n__title__ = \"gh2md\"\n__description__ = \"Export Github repository issues to markdown files: https:\/\/github.com\/mattduck\/gh2md.\"\n__doc__ = __description__\n__uri__ = \"https:\/\/github.com\/mattduck\/gh2md\"\n\n__author__ = \"Matthew Duck\"\n__email__ = \"matt@mattduck.com\"\n\n__license__ = \"MIT\"\n__copyright__ = \"Copyright (c) 2017 Matthew Duck\"\n","subject":"Bump to next dev version","message":"Bump to next dev version\n\nThis is really just to confirm that Travis is still building correctly.\n","lang":"Python","license":"mit","repos":"mattduck\/gh2md"} {"commit":"61fecbed71129228e7020a9e95dbcd2487bbdbb3","old_file":"turbustat\/tests\/test_scf.py","new_file":"turbustat\/tests\/test_scf.py","old_contents":"# Licensed under an MIT open source license - see LICENSE\n\n\n'''\nTest functions for SCF\n'''\n\nfrom unittest import TestCase\n\nimport numpy as np\nimport numpy.testing as npt\n\nfrom ..statistics import SCF, SCF_Distance\nfrom ._testing_data import \\\n dataset1, dataset2, computed_data, computed_distances\n\n\nclass testSCF(TestCase):\n\n def setUp(self):\n self.dataset1 = dataset1\n self.dataset2 = dataset2\n\n def test_SCF_method(self):\n self.tester = SCF(dataset1[\"cube\"], size=11)\n self.tester.run()\n\n assert np.allclose(self.tester.scf_surface, computed_data['scf_val'])\n\n def test_SCF_distance(self):\n self.tester_dist = \\\n SCF_Distance(dataset1[\"cube\"],\n dataset2[\"cube\"], size=11).distance_metric()\n npt.assert_almost_equal(self.tester_dist.distance,\n computed_distances['scf_distance'])\n","new_contents":"# Licensed under an MIT open source license - see LICENSE\n\n\n'''\nTest functions for SCF\n'''\n\nfrom unittest import TestCase\n\nimport numpy as np\nimport numpy.testing as npt\nfrom scipy.ndimage import zoom\n\nfrom ..statistics import SCF, SCF_Distance\nfrom ._testing_data import \\\n dataset1, dataset2, computed_data, computed_distances\n\n\nclass testSCF(TestCase):\n\n def test_SCF_method(self):\n self.tester = SCF(dataset1[\"cube\"], size=11)\n self.tester.run()\n\n assert np.allclose(self.tester.scf_surface, computed_data['scf_val'])\n\n def test_SCF_distance(self):\n self.tester_dist = \\\n SCF_Distance(dataset1[\"cube\"],\n dataset2[\"cube\"], size=11).distance_metric()\n npt.assert_almost_equal(self.tester_dist.distance,\n computed_distances['scf_distance'])\n\n def test_SCF_regrid_distance(self):\n hdr = dataset1[\"cube\"][1].copy()\n hdr[\"CDELT2\"] = 0.5 * hdr[\"CDELT2\"]\n hdr[\"CDELT1\"] = 0.5 * hdr[\"CDELT1\"]\n cube = zoom(dataset1[\"cube\"][0], (1, 2, 2))\n\n self.tester_dist_zoom = \\\n SCF_Distance([cube, hdr], dataset1[\"cube\"],\n size=11).distance_metric(verbose=True)\n\n fid_dist = 0.02\n\n assert self.tester_dist_zoom < fid_dist\n","subject":"Add simple SCF test for unequal grids","message":"Add simple SCF test for unequal grids\n","lang":"Python","license":"mit","repos":"e-koch\/TurbuStat,Astroua\/TurbuStat"} {"commit":"e5616c75e8218442634904c024158d3e69000733","old_file":"alabaster\/_version.py","new_file":"alabaster\/_version.py","old_contents":"__version_info__ = (0, 4, 1)\n__version__ = '.'.join(map(str, __version_info__))\n","new_contents":"__version_info__ = (0, 5, 0)\n__version__ = '.'.join(map(str, __version_info__))\n","subject":"Bump version to 0.5 cuz feature","message":"Bump version to 0.5 cuz feature\n","lang":"Python","license":"bsd-3-clause","repos":"ellisonbg\/jupyter-alabaster-theme,jupytercalpoly\/jupyter-alabaster-theme,nikolas\/alabaster,charnpreetsingh\/jupyter-alabaster-theme,bgeron\/alabaster,nikolas\/alabaster,bgeron\/alabaster,charnpreetsingh\/jupyter-alabaster-theme,jupytercalpoly\/jupyter-alabaster-theme,ellisonbg\/jupyter-alabaster-theme,ellisonbg\/jupyter-alabaster-theme,jupytercalpoly\/jupyter-alabaster-theme,charnpreetsingh\/jupyter-alabaster-theme"} {"commit":"698677a623722b63ec4cceb7690b62fa7e4ede37","old_file":"django_prices_openexchangerates\/templatetags\/prices_multicurrency_i18n.py","new_file":"django_prices_openexchangerates\/templatetags\/prices_multicurrency_i18n.py","old_contents":"from django.template import Library\nfrom django_prices.templatetags import prices_i18n\n\nfrom .. import exchange_currency\n\nregister = Library()\n\n\n@register.simple_tag # noqa\ndef gross_in_currency(price, currency): # noqa\n converted_price = exchange_currency(price, currency)\n return prices_i18n.gross(converted_price)\n\n\n@register.simple_tag # noqa\ndef net_in_currency(price, currency): # noqa\n converted_price = exchange_currency(price, currency)\n return prices_i18n.net(converted_price)\n\n\n@register.simple_tag # noqa\ndef tax_in_currency(price, currency): # noqa\n converted_price = exchange_currency(price, currency)\n return prices_i18n.tax(converted_price)\n\n","new_contents":"from django.template import Library\nfrom django_prices.templatetags import prices_i18n\n\nfrom .. import exchange_currency\n\nregister = Library()\n\n\n@register.simple_tag # noqa\ndef gross_in_currency(price, currency): # noqa\n converted_price = exchange_currency(price, currency)\n return prices_i18n.gross(converted_price)\n\n\n@register.simple_tag # noqa\ndef net_in_currency(price, currency): # noqa\n converted_price = exchange_currency(price, currency)\n return prices_i18n.net(converted_price)\n\n\n@register.simple_tag # noqa\ndef tax_in_currency(price, currency): # noqa\n converted_price = exchange_currency(price, currency)\n return prices_i18n.tax(converted_price)\n\n\n@register.simple_tag\ndef discount_amount_in_currency(discount, price, currency):\n price = exchange_currency(price, to_currency=currency)\n discount_amount = exchange_currency(discount.amount, to_currency=currency)\n discount.amount = discount_amount\n return (price | discount) - price\n","subject":"Add templatetag for converting discounts between currencies","message":"Add templatetag for converting discounts between currencies\n","lang":"Python","license":"bsd-3-clause","repos":"mirumee\/django-prices-openexchangerates,artursmet\/django-prices-openexchangerates"} {"commit":"32a54ff0588930efc5e0ee3c61f2efbf57e450e0","old_file":"inviter\/tests.py","new_file":"inviter\/tests.py","old_contents":"\"\"\"\nThis file demonstrates writing tests using the unittest module. These will pass\nwhen you run \"manage.py test\".\n\nReplace this with more appropriate tests for your application.\n\"\"\"\nfrom django.contrib.auth.models import User\nfrom django.core.mail import get_connection\nfrom django.test import TestCase\nfrom inviter.utils import invite\nimport shortuuid\n\n\nclass InviteTest(TestCase):\n def setUp(self):\n self.inviter = User.objects.create(username = shortuuid.uuid())\n self.existing = User.objects.create(username = shortuuid.uuid(),\n email = 'existing@example.com')\n self.backend = get_connection('django.core.mail.backends.locmem.EmailBackend')\n\n def test_inviting(self):\n user = invite(\"foo@example.com\", self.inviter) \n self.assertFalse(user.is_active)\n self.assertEqual(1, len(self.backend.emails))\n self.assertEqual(3, User.objects.count())\n \n user = invite(\"foo@example.com\", self.inviter)\n self.assertFalse(user.is_active)\n self.assertEqual(2, len(self.backend.emails))\n self.assertEqual(3, User.objects.count())\n \n user = invite(\"existing@example.com\", self.inviter)\n self.assertTrue(user.is_active)\n self.assertEqual(2, len(self.backend.emails))\n self.assertEqual(3, User.objects.count())\n \n\n","new_contents":"\"\"\"\nThis file demonstrates writing tests using the unittest module. These will pass\nwhen you run \"manage.py test\".\n\nReplace this with more appropriate tests for your application.\n\"\"\"\nfrom django.contrib.auth.models import User\nfrom django.core.mail import outbox\nfrom django.test import TestCase\nfrom inviter.utils import invite\nimport shortuuid\n\n\nclass InviteTest(TestCase):\n def setUp(self):\n self.inviter = User.objects.create(username = shortuuid.uuid())\n self.existing = User.objects.create(username = shortuuid.uuid(),\n email = 'existing@example.com')\n self.outbox = outbox\n\n def test_inviting(self):\n user = invite(\"foo@example.com\", self.inviter) \n self.assertFalse(user.is_active)\n self.assertEqual(1, len(self.outbox))\n self.assertEqual(3, User.objects.count())\n \n user = invite(\"foo@example.com\", self.inviter)\n self.assertFalse(user.is_active)\n self.assertEqual(2, len(self.outbox))\n self.assertEqual(3, User.objects.count())\n \n user = invite(\"existing@example.com\", self.inviter)\n self.assertTrue(user.is_active)\n self.assertEqual(2, len(self.outbox))\n self.assertEqual(3, User.objects.count())\n \n\n","subject":"Test fix to check django.core.mail.outbox","message":"Test fix to check django.core.mail.outbox\n","lang":"Python","license":"mit","repos":"caffeinehit\/django-inviter"} {"commit":"9d6c8eaa491d0988bf16633bbba9847350f57778","old_file":"spacy\/lang\/norm_exceptions.py","new_file":"spacy\/lang\/norm_exceptions.py","old_contents":"# coding: utf8\nfrom __future__ import unicode_literals\n\n\n# These exceptions are used to add NORM values based on a token's ORTH value.\n# Individual languages can also add their own exceptions and overwrite them -\n# for example, British vs. American spelling in English.\n\n# Norms are only set if no alternative is provided in the tokenizer exceptions.\n# Note that this does not change any other token attributes. Its main purpose\n# is to normalise the word representations so that equivalent tokens receive\n# similar representations. For example: $ and € are very different, but they're\n# both currency symbols. By normalising currency symbols to $, all symbols are\n# seen as similar, no matter how common they are in the training data.\n\n\nBASE_NORMS = {\n \"'s\": \"'s\",\n \"'S\": \"'s\",\n \"’s\": \"'s\",\n \"’S\": \"'s\",\n \"’\": \"'\",\n \"‘\": \"'\",\n \"´\": \"'\",\n \"`\": \"'\",\n \"”\": '\"',\n \"“\": '\"',\n \"''\": '\"',\n \"``\": '\"',\n \"´´\": '\"',\n \"„\": '\"',\n \"»\": '\"',\n \"«\": '\"',\n \"…\": \"...\",\n \"—\": \"-\",\n \"–\": \"-\",\n \"--\": \"-\",\n \"---\": \"-\",\n \"€\": \"$\",\n \"£\": \"$\",\n \"¥\": \"$\",\n \"฿\": \"$\",\n \"US$\": \"$\",\n \"C$\": \"$\",\n \"A$\": \"$\"\n}\n","new_contents":"# coding: utf8\nfrom __future__ import unicode_literals\n\n\n# These exceptions are used to add NORM values based on a token's ORTH value.\n# Individual languages can also add their own exceptions and overwrite them -\n# for example, British vs. American spelling in English.\n\n# Norms are only set if no alternative is provided in the tokenizer exceptions.\n# Note that this does not change any other token attributes. Its main purpose\n# is to normalise the word representations so that equivalent tokens receive\n# similar representations. For example: $ and € are very different, but they're\n# both currency symbols. By normalising currency symbols to $, all symbols are\n# seen as similar, no matter how common they are in the training data.\n\n\nBASE_NORMS = {\n \"'s\": \"'s\",\n \"'S\": \"'s\",\n \"’s\": \"'s\",\n \"’S\": \"'s\",\n \"’\": \"'\",\n \"‘\": \"'\",\n \"´\": \"'\",\n \"`\": \"'\",\n \"”\": '\"',\n \"“\": '\"',\n \"''\": '\"',\n \"``\": '\"',\n \"´´\": '\"',\n \"„\": '\"',\n \"»\": '\"',\n \"«\": '\"',\n \"‘‘\": '\"',\n \"’’\": '\"',\n \"?\": \"?\",\n \"!\": \"!\",\n \",\": \",\",\n \";\": \";\",\n \":\": \":\",\n \"。\": \".\",\n \"।\": \".\",\n \"…\": \"...\",\n \"—\": \"-\",\n \"–\": \"-\",\n \"--\": \"-\",\n \"---\": \"-\",\n \"——\": \"-\",\n \"€\": \"$\",\n \"£\": \"$\",\n \"¥\": \"$\",\n \"฿\": \"$\",\n \"US$\": \"$\",\n \"C$\": \"$\",\n \"A$\": \"$\"\n}\n","subject":"Update base norm exceptions with more unicode characters","message":"Update base norm exceptions with more unicode characters\n\ne.g. unicode variations of punctuation used in Chinese\n","lang":"Python","license":"mit","repos":"aikramer2\/spaCy,aikramer2\/spaCy,recognai\/spaCy,explosion\/spaCy,aikramer2\/spaCy,recognai\/spaCy,honnibal\/spaCy,explosion\/spaCy,recognai\/spaCy,aikramer2\/spaCy,explosion\/spaCy,recognai\/spaCy,aikramer2\/spaCy,recognai\/spaCy,spacy-io\/spaCy,spacy-io\/spaCy,honnibal\/spaCy,explosion\/spaCy,honnibal\/spaCy,explosion\/spaCy,spacy-io\/spaCy,recognai\/spaCy,spacy-io\/spaCy,aikramer2\/spaCy,spacy-io\/spaCy,explosion\/spaCy,spacy-io\/spaCy,honnibal\/spaCy"} {"commit":"56c4de583847be5fb16818dcd1ca855fc6007b50","old_file":"pylsdj\/__init__.py","new_file":"pylsdj\/__init__.py","old_contents":"","new_contents":"__title__ = 'pylsdj'\n__version__ = '1.1.0'\n__build__ = 0x010100\n__author__ = 'Alex Rasmussen'\n__license__ = 'MIT'\n__copyright__ = 'Copyright 2014 Alex Rasmussen'\n\nimport bread_spec\nimport chain\nimport clock\nimport consts\nimport filepack\nfrom instrument import Instrument\nfrom phrase import Phrase\nfrom project import load_lsdsng, load_srm, Project\nfrom savfile import SAVFile\nfrom song import Song, Sequence\nfrom speech_instrument import Word, SpeechInstrument\nfrom synth import Synth\nfrom table import Table\n","subject":"Make imports a little more intuitive.","message":"Make imports a little more intuitive.\n","lang":"Python","license":"mit","repos":"alexras\/pylsdj,alexras\/pylsdj"} {"commit":"770fd77e0fc7a0700b81e4418e5f97bd88d842d0","old_file":"pymoira\/filesys.py","new_file":"pymoira\/filesys.py","old_contents":"#\n## PyMoira client library\n##\n## This file contains the more abstract methods which allow user to work with\n## lists and list members.\n#\n\nimport protocol\nimport utils\nimport datetime\nfrom errors import *\n\nclass Filesys(object):\n info_query_description = (\n ('label', str),\n ('type', str),\n ('machine', str),\n ('name', str),\n ('mountpoint', str),\n ('access_mode', str),\n ('description', str),\n ('owner_user', str),\n ('owner_group', str),\n ('create', bool),\n ('type', str),\n ('lastmod_datetime', datetime.datetime),\n ('lastmod_by', str),\n ('lastmod_with', str),\n )\n\n def __init__(self, client, name):\n self.client = client\n self.name = name\n\n def loadInfo(self):\n \"\"\"Loads the information about the list from the server into the object.\"\"\"\n \n response, = self.client.query( 'get_filesys_by_label', (self.name, ), version = 14 )\n result = utils.responseToDict(self.info_query_description, response)\n self.__dict__.update(result)\n","new_contents":"#\n## PyMoira client library\n##\n## This file contains the more abstract methods which allow user to work with\n## lists and list members.\n#\n\nimport protocol\nimport utils\nimport datetime\nfrom errors import *\n\nclass Filesys(object):\n info_query_description = (\n ('label', str),\n ('type', str),\n ('machine', str),\n ('name', str),\n ('mountpoint', str),\n ('access_mode', str),\n ('description', str),\n ('owner_user', str),\n ('owner_group', str),\n ('create', bool),\n ('locker_type', str),\n ('lastmod_datetime', datetime.datetime),\n ('lastmod_by', str),\n ('lastmod_with', str),\n )\n\n def __init__(self, client, name):\n self.client = client\n self.name = name\n\n def loadInfo(self):\n \"\"\"Loads the information about the list from the server into the object.\"\"\"\n \n response, = self.client.query( 'get_filesys_by_label', (self.name, ), version = 14 )\n result = utils.responseToDict(self.info_query_description, response)\n self.__dict__.update(result)\n","subject":"Fix a name collision for two types of types.","message":"Fix a name collision for two types of types.\n","lang":"Python","license":"mit","repos":"vasilvv\/pymoira"} {"commit":"7394ba6eba50282bd7252e504a80e5d595dd12bc","old_file":"ci\/fix_paths.py","new_file":"ci\/fix_paths.py","old_contents":"import distutils.sysconfig\nfrom glob import glob\nimport os\nfrom os.path import join as pjoin, basename\nfrom shutil import copy\nfrom sys import platform\n\ndef main():\n \"\"\"\n Copy HDF5 DLLs into installed h5py package\n \"\"\"\n # This is the function Tox also uses to locate site-packages (Apr 2019)\n sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True)\n print(\"site packages dir:\", sitepackagesdir)\n\n hdf5_path = os.environ.get(\"HDF5_DIR\")\n print(\"HDF5_DIR\", hdf5_path)\n\n # HDF5_DIR is not set when we're testing wheels; these should already have\n # the necessary libraries bundled in.\n if platform.startswith('win') and hdf5_path is not None:\n for f in glob(pjoin(hdf5_path, 'lib\/*.dll')):\n copy(f, pjoin(sitepackagesdir, 'h5py', basename(f)))\n print(\"Copied\", f)\n\n print(\"In installed h5py:\", os.listdir(pjoin(sitepackagesdir, 'h5py')))\n\nif __name__ == '__main__':\n main()\n","new_contents":"import distutils.sysconfig\nfrom glob import glob\nimport os\nfrom os.path import join as pjoin, basename\nfrom shutil import copy\nfrom sys import platform\n\ndef main():\n \"\"\"\n Copy HDF5 DLLs into installed h5py package\n \"\"\"\n # This is the function Tox also uses to locate site-packages (Apr 2019)\n sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True)\n print(\"site packages dir:\", sitepackagesdir)\n\n hdf5_path = os.environ.get(\"HDF5_DIR\")\n print(\"HDF5_DIR\", hdf5_path)\n\n # HDF5_DIR is not set when we're testing wheels; these should already have\n # the necessary libraries bundled in.\n if platform.startswith('win') and hdf5_path is not None:\n for f in glob(pjoin(hdf5_path, 'lib\/*.dll')):\n copy(f, pjoin(sitepackagesdir, 'h5py', basename(f)))\n print(\"Copied\", f)\n\n zlib_root = os.environ.get(\"ZLIB_ROOT\")\n if zlib_root:\n f = pjoin(zlib_root, 'bin_release', 'zlib.dll')\n copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll'))\n print(\"Copied\", f)\n\n print(\"In installed h5py:\", os.listdir(pjoin(sitepackagesdir, 'h5py')))\n\nif __name__ == '__main__':\n main()\n","subject":"Copy zlib.dll into Windows h5py installed from source","message":"Copy zlib.dll into Windows h5py installed from source\n","lang":"Python","license":"bsd-3-clause","repos":"h5py\/h5py,h5py\/h5py,h5py\/h5py"} {"commit":"6db2bb9b1634a7b37790207e5b8d420de643a9cb","old_file":"turbasen\/__init__.py","new_file":"turbasen\/__init__.py","old_contents":"VERSION = '1.0.0'\n\ndef configure(**settings):\n from .settings import Settings\n for key, value in settings.items():\n Settings.setattr(key, value)\n","new_contents":"VERSION = '1.0.0'\n\nfrom .models import \\\n Omrade, \\\n Sted\n\ndef configure(**settings):\n from .settings import Settings\n for key, value in settings.items():\n Settings.setattr(key, value)\n","subject":"Add relevant models to turbasen module","message":"Add relevant models to turbasen module\n","lang":"Python","license":"mit","repos":"Turbasen\/turbasen.py"} {"commit":"e8b1cee54f679cbd6a2d158b3c2789f3f6a3d9c0","old_file":"uppercase.py","new_file":"uppercase.py","old_contents":"from twisted.internet import protocol, reactor\n\nfactory = protocol.ServerFactory()\nfactory.protocol = protocol.Protocol\n\nreactor.listenTCP(8000, factory)\nreactor.run()\n","new_contents":"from twisted.internet import endpoints, protocol, reactor\n\nclass UpperProtocol(protocol.Protocol):\n pass\n\nfactory = protocol.ServerFactory()\nfactory.protocol = UpperProtocol\n\nendpoints.serverFromString(reactor, 'tcp:8000').listen(factory)\nreactor.run()\n","subject":"Convert to endpoints API and use custom protocol class","message":"Convert to endpoints API and use custom protocol class\n","lang":"Python","license":"mit","repos":"cataliniacob\/ep2012-tutorial-twisted"} {"commit":"6b179dc4fb95f4db380b9156381b6210adeef2e5","old_file":"conftest.py","new_file":"conftest.py","old_contents":"# Copyright 2016 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\n\nimport mock\nimport pytest\n\nPROJECT = os.environ['GCLOUD_PROJECT']\n\n\n@pytest.fixture\ndef api_client_inject_project_id():\n \"\"\"Patches all googleapiclient requests to replace 'YOUR_PROJECT_ID' with\n the project ID.\"\"\"\n import googleapiclient.http\n\n old_execute = googleapiclient.http.HttpRequest.execute\n\n def new_execute(self, http=None, num_retries=0):\n self.uri = self.uri.replace('YOUR_PROJECT_ID', PROJECT)\n return old_execute(self, http=http, num_retries=num_retries)\n\n with mock.patch(\n 'googleapiclient.http.HttpRequest.execute',\n new=new_execute):\n yield\n","new_contents":"# Copyright 2016 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport os\n\nimport mock\nimport pytest\n\nPROJECT = 'python-docs-samples'\n\n\n@pytest.fixture\ndef api_client_inject_project_id():\n \"\"\"Patches all googleapiclient requests to replace 'YOUR_PROJECT_ID' with\n the project ID.\"\"\"\n import googleapiclient.http\n\n old_execute = googleapiclient.http.HttpRequest.execute\n\n def new_execute(self, http=None, num_retries=0):\n self.uri = self.uri.replace('YOUR_PROJECT_ID', PROJECT)\n return old_execute(self, http=http, num_retries=num_retries)\n\n with mock.patch(\n 'googleapiclient.http.HttpRequest.execute',\n new=new_execute):\n yield\n","subject":"Set the Project in code","message":"Set the Project in code\n","lang":"Python","license":"apache-2.0","repos":"GoogleCloudPlatform\/getting-started-python,GoogleCloudPlatform\/getting-started-python,GoogleCloudPlatform\/getting-started-python"} {"commit":"378f3bf0bb2e05260b7cbeeb4a4637d7d3a7ca7c","old_file":"workflows\/consumers.py","new_file":"workflows\/consumers.py","old_contents":"from urllib.parse import parse_qs\nfrom channels import Group\nfrom channels.sessions import channel_session\n\n\n@channel_session\ndef ws_add(message):\n message.reply_channel.send({\"accept\": True})\n qs = parse_qs(message['query_string'])\n workflow_pk = qs['workflow_pk'][0]\n message.channel_session['workflow_pk'] = workflow_pk\n Group(\"workflow-{}\".format(workflow_pk)).add(message.reply_channel)\n\n\n@channel_session\ndef ws_disconnect(message):\n workflow_pk = message.channel_session['workflow_pk']\n Group(\"workflow-{}\".format(workflow_pk)).discard(message.reply_channel)\n","new_contents":"from urllib.parse import parse_qs\nfrom channels import Group\nfrom channels.sessions import channel_session\n\n\n@channel_session\ndef ws_add(message):\n message.reply_channel.send({\"accept\": True})\n qs = parse_qs(message['query_string'])\n workflow_pk = qs[b'workflow_pk'][0].decode('utf-8')\n message.channel_session['workflow_pk'] = workflow_pk\n Group(\"workflow-{}\".format(workflow_pk)).add(message.reply_channel)\n\n\n@channel_session\ndef ws_disconnect(message):\n workflow_pk = message.channel_session['workflow_pk']\n Group(\"workflow-{}\".format(workflow_pk)).discard(message.reply_channel)\n","subject":"Fix query string problem of comparing byte strings and unicode strings in py3","message":"Fix query string problem of comparing byte strings and unicode strings in py3\n","lang":"Python","license":"mit","repos":"xflows\/clowdflows-backend,xflows\/clowdflows-backend,xflows\/clowdflows-backend,xflows\/clowdflows-backend"} {"commit":"9a7c80744bc1e57fe0ec5fc7cf149dada2d05121","old_file":"neuroimaging\/utils\/tests\/data\/__init__.py","new_file":"neuroimaging\/utils\/tests\/data\/__init__.py","old_contents":"\"\"\"Information used for locating nipy test data.\n\nNipy uses a set of test data that is installed separately. The test\ndata should be located in the directory ``~\/.nipy\/tests\/data``.\n\nInstall the data in your home directory from the data repository::\n $ mkdir -p .nipy\/tests\/data\n $ svn co http:\/\/neuroimaging.scipy.org\/svn\/ni\/data\/trunk\/fmri .nipy\/tests\/data\n\n\"\"\"\nfrom os.path import expanduser, exists, join\n\nfrom neuroimaging.io.datasource import Repository\n\n# data directory should be: $HOME\/.nipy\/tests\/data\ndatapath = expanduser(join('~', '.nipy', 'tests', 'data'))\n\nif not exists(datapath):\n raise IOError, 'Nipy data directory is not found!'\n\nrepository = Repository(datapath)\n\n","new_contents":"\"\"\"\nNipy uses a set of test data that is installed separately. The test\ndata should be located in the directory ``~\/.nipy\/tests\/data``.\n\nInstall the data in your home directory from the data repository::\n $ mkdir -p .nipy\/tests\/data\n $ svn co http:\/\/neuroimaging.scipy.org\/svn\/ni\/data\/trunk\/fmri .nipy\/tests\/data\n\n\"\"\"\n\n# Fernando pointed out that we should wrap the test data into a\n# tarball and write a pure python function to grab the data for people\n# instead of using svn. Users may not have svn and the mkdir may not\n# work on Windows.\n\nfrom os.path import expanduser, exists, join\n\nfrom neuroimaging.io.datasource import Repository\n\n# data directory should be: $HOME\/.nipy\/tests\/data\ndatapath = expanduser(join('~', '.nipy', 'tests', 'data'))\n\nif not exists(datapath):\n msg = 'Nipy data directory is not found!\\n%s' % __doc__\n raise IOError(msg)\n\nrepository = Repository(datapath)\n","subject":"Extend error message regarding missing test data.","message":"Extend error message regarding missing test data.","lang":"Python","license":"bsd-3-clause","repos":"alexis-roche\/register,bthirion\/nipy,bthirion\/nipy,nipy\/nipy-labs,alexis-roche\/nipy,bthirion\/nipy,alexis-roche\/register,alexis-roche\/nireg,bthirion\/nipy,arokem\/nipy,alexis-roche\/register,alexis-roche\/niseg,alexis-roche\/nipy,arokem\/nipy,nipy\/nipy-labs,alexis-roche\/nipy,alexis-roche\/nipy,alexis-roche\/nireg,arokem\/nipy,arokem\/nipy,nipy\/nireg,nipy\/nireg,alexis-roche\/niseg"} {"commit":"446400fa4e40ca7e47e48dd00209d80858094552","old_file":"buffer\/managers\/profiles.py","new_file":"buffer\/managers\/profiles.py","old_contents":"import json\n\nfrom buffer.models.profile import PATHS, Profile\n\nclass Profiles(list):\n\n def __init__(self, api, *args, **kwargs):\n super(Profiles, self).__init__(*args, **kwargs)\n\n self.api = api\n\n def all(self):\n response = self.api.get(url=PATHS['GET_PROFILES'], parser=json.loads)\n\n for raw_profile in response:\n self.append(Profile(self.api, raw_profile))\n\n return self\n\n def filter(self, **kwargs):\n\n if not len(self):\n self.all()\n\n new_list = filter(lambda item: [True for arg in kwargs if item[arg] == kwargs[arg]] != [], self)\n\n return Profiles(self.api, new_list)\n","new_contents":"import json\n\nfrom buffer.models.profile import PATHS, Profile\n\nclass Profiles(list):\n '''\n Manage profiles\n + all -> get all the profiles from buffer\n + filter -> wrapper for list filtering\n '''\n\n def __init__(self, api, *args, **kwargs):\n super(Profiles, self).__init__(*args, **kwargs)\n\n self.api = api\n\n def all(self):\n '''\n Get all network profiles\n '''\n\n response = self.api.get(url=PATHS['GET_PROFILES'], parser=json.loads)\n\n for raw_profile in response:\n self.append(Profile(self.api, raw_profile))\n\n return self\n\n def filter(self, **kwargs):\n '''\n Based on some criteria, filter the profiles and return a new Profiles\n Manager containing only the chosen items\n\n If the manager doen't have any items, get all the profiles from Buffer\n '''\n\n if not len(self):\n self.all()\n\n new_list = filter(lambda item: [True for arg in kwargs if item[arg] == kwargs[arg]] != [], self)\n\n return Profiles(self.api, new_list)\n","subject":"Write documentation for profile mananger","message":"Write documentation for profile mananger\n","lang":"Python","license":"mit","repos":"bufferapp\/buffer-python,vtemian\/buffpy"} {"commit":"1c7bbeabe1c1f3eea053c8fd8b6649ba388c1d2e","old_file":"waliki\/slides\/views.py","new_file":"waliki\/slides\/views.py","old_contents":"from os import path\nimport shutil\nimport tempfile\nfrom sh import hovercraft\nfrom django.shortcuts import get_object_or_404\nfrom django.http import HttpResponse\nfrom waliki.models import Page\n\n\ndef slides(request, slug):\n page = get_object_or_404(Page, slug=slug)\n outpath = tempfile.mkdtemp()\n try:\n infile = page.abspath\n template = path.join(path.dirname(path.realpath(__file__)), 'template')\n hovercraft('-t', template, infile, outpath)\n with open(path.join(outpath, 'index.html')) as f:\n content = f.read()\n finally:\n shutil.rmtree(outpath)\n return HttpResponse(content)\n","new_contents":"from os import path\nimport shutil\nimport tempfile\nfrom sh import hovercraft\nfrom django.shortcuts import get_object_or_404\nfrom django.http import HttpResponse\nfrom waliki.models import Page\nfrom waliki.acl import permission_required\n\n\n@permission_required('view_page')\ndef slides(request, slug):\n page = get_object_or_404(Page, slug=slug)\n outpath = tempfile.mkdtemp()\n try:\n infile = page.abspath\n template = path.join(path.dirname(path.realpath(__file__)), 'template')\n hovercraft('-t', template, infile, outpath)\n with open(path.join(outpath, 'index.html')) as f:\n content = f.read()\n finally:\n shutil.rmtree(outpath)\n return HttpResponse(content)\n","subject":"Add permission check to slide urls.","message":"Add permission check to slide urls.\n","lang":"Python","license":"bsd-3-clause","repos":"RobertoMaurizzi\/waliki,aszepieniec\/waliki,aszepieniec\/waliki,OlegGirko\/waliki,RobertoMaurizzi\/waliki,rizotas\/waliki,beres\/waliki,beres\/waliki,santiavenda2\/waliki,mgaitan\/waliki,fpytloun\/waliki,OlegGirko\/waliki,OlegGirko\/waliki,santiavenda2\/waliki,santiavenda2\/waliki,beres\/waliki,mgaitan\/waliki,RobertoMaurizzi\/waliki,aszepieniec\/waliki,rizotas\/waliki,mgaitan\/waliki,fpytloun\/waliki,fpytloun\/waliki,rizotas\/waliki"} {"commit":"ceac9c401f80a279e7291e7ba2a9e06757d4dd1d","old_file":"buildtools\/wrapper\/cmake.py","new_file":"buildtools\/wrapper\/cmake.py","old_contents":"import os\n\nfrom buildtools.bt_logging import log\nfrom buildtools.os_utils import cmd, ENV\n\nclass CMake(object):\n def __init__(self):\n self.flags = {}\n self.generator = None\n \n def setFlag(self, key, val):\n log.info('CMake: {} = {}'.format(key, val))\n self.flags[key] = val\n \n def build(self, CMAKE, dir='.', env=None, target=None, moreflags=[]):\n moreflags += ['--build']\n if target is not None:\n moreflags += ['--target',target]\n self.run(CMAKE,dir,env,moreflags)\n \n def run(self, CMAKE, env=None, dir='.', moreflags=[]):\n if env is None:\n env = ENV.env\n flags = []\n \n if self.generator is not None:\n flags += ['-G',self.generator]\n \n for key, value in self.flags.items():\n flags += ['-D{0}={1}'.format(key, value)]\n \n flags += moreflags\n \n with log.info('Running CMake:'):\n for key, value in env.items():\n log.info('+{0}=\"{1}\"'.format(key, value))\n return cmd([CMAKE] + flags + [dir], env=env, critical=True, echo=True)\n return False\n","new_contents":"import os\n\nfrom buildtools.bt_logging import log\nfrom buildtools.os_utils import cmd, ENV\n\nclass CMake(object):\n def __init__(self):\n self.flags = {}\n self.generator = None\n \n def setFlag(self, key, val):\n log.info('CMake: {} = {}'.format(key, val))\n self.flags[key] = val\n \n def build(self, CMAKE, dir='.', env=None, target=None, moreflags=[]):\n moreflags += ['--build']\n if target is not None:\n moreflags += ['--target', target]\n self.run(CMAKE, env, dir, moreflags)\n \n def run(self, CMAKE, env=None, dir='.', moreflags=[]):\n if env is None:\n env = ENV.env\n flags = []\n \n if self.generator is not None:\n flags += ['-G', self.generator]\n \n for key, value in self.flags.items():\n flags += ['-D{0}={1}'.format(key, value)]\n \n flags += moreflags\n \n with log.info('Running CMake:'):\n for key, value in env.items():\n log.info('+{0}=\"{1}\"'.format(key, value))\n return cmd([CMAKE] + flags + [dir], env=env, critical=True, echo=True)\n return False\n","subject":"Fix mismatched args in CMake.build","message":"Fix mismatched args in CMake.build\n","lang":"Python","license":"mit","repos":"N3X15\/python-build-tools,N3X15\/python-build-tools,N3X15\/python-build-tools"} {"commit":"f04c451de83b66b733dc28eb13bc16ade2675b3a","old_file":"changes\/api\/stream.py","new_file":"changes\/api\/stream.py","old_contents":"from __future__ import absolute_import\n\nimport gevent\n\nfrom gevent.queue import Queue\n\nfrom changes.config import pubsub\n\n\nclass EventStream(object):\n def __init__(self, channels, pubsub=pubsub):\n self.pubsub = pubsub\n self.pending = Queue()\n self.channels = channels\n self.active = True\n\n for channel in channels:\n self.pubsub.subscribe(channel, self.push)\n\n def __iter__(self):\n while self.active:\n # TODO(dcramer): figure out why we have to send this to ensure\n # the connection is opened\n yield \"\\n\"\n event = self.pending.get()\n yield \"event: {}\\n\".format(event['event'])\n for line in event['data'].splitlines():\n yield \"data: {}\\n\".format(line)\n yield \"\\n\"\n gevent.sleep(0)\n\n def __del__(self):\n self.close()\n\n def push(self, message):\n self.pending.put_nowait(message)\n\n def close(self):\n for channel in self.channels:\n self.pubsub.unsubscribe(channel, self.push)\n","new_contents":"from __future__ import absolute_import\n\nimport gevent\n\nfrom collections import deque\n\nfrom changes.config import pubsub\n\n\nclass EventStream(object):\n def __init__(self, channels, pubsub=pubsub):\n self.pubsub = pubsub\n self.pending = deque()\n self.channels = channels\n self.active = True\n\n for channel in channels:\n self.pubsub.subscribe(channel, self.push)\n\n def __iter__(self):\n while self.active:\n # TODO(dcramer): figure out why we have to send this to ensure\n # the connection is opened\n yield \"\\n\"\n while self.pending:\n event = self.pending.pop()\n yield \"event: {}\\n\".format(event['event'])\n for line in event['data'].splitlines():\n yield \"data: {}\\n\".format(line)\n yield \"\\n\"\n gevent.sleep(0)\n gevent.sleep(0.3)\n\n def __del__(self):\n self.close()\n\n def push(self, message):\n self.pending.append(message)\n\n def close(self):\n for channel in self.channels:\n self.pubsub.unsubscribe(channel, self.push)\n","subject":"Revert \"Switch to gevent queues for EventStream\"","message":"Revert \"Switch to gevent queues for EventStream\"\n\nThis reverts commit 4c102945fe46bb463e6a641324d26384c77fbae8.\n","lang":"Python","license":"apache-2.0","repos":"bowlofstew\/changes,wfxiang08\/changes,dropbox\/changes,wfxiang08\/changes,dropbox\/changes,bowlofstew\/changes,dropbox\/changes,dropbox\/changes,bowlofstew\/changes,wfxiang08\/changes,wfxiang08\/changes,bowlofstew\/changes"} {"commit":"24a90b0aa38a21c9b116d4b8b9c4878678fda9cc","old_file":"suddendev\/tasks.py","new_file":"suddendev\/tasks.py","old_contents":"from . import celery, celery_socketio\nfrom .game_instance import GameInstance\nimport time\n\n@celery.task(time_limit=5, max_retries=3)\ndef play_game(game_id, player_names, scripts, player_ids, colors, namespace, room, wave=1):\n\n cleared = True\n current_wave = wave - 1\n while cleared:\n current_wave += 1\n game = GameInstance(game_id, player_names, scripts, player_ids, colors, wave=current_wave)\n for batch in game.run():\n celery_socketio.emit('result', '{\\\"result\\\": [ ' + ','.join(batch) + ']}', room=room, namespace=namespace)\n cleared = game.was_cleared()\n\n # return highest reached wave\n return current_wave - 1\n\n@celery.task(time_limit=5, max_retries=3)\ndef test_round(game_id, player_names, scripts, player_ids, colors, namespace, room, wave=1):\n game = GameInstance(game_id, player_names, scripts, player_ids, colors, wave=wave)\n for batch in game.run():\n celery_socketio.emit('result', '{\\\"result\\\": [ ' + ','.join(batch) + ']}', room=room, namespace=namespace)\n\n # return if the test round was cleared\n return game.was_cleared()\n","new_contents":"from . import celery, celery_socketio\nfrom .game_instance import GameInstance\nimport time\n\n@celery.task(time_limit=15, max_retries=3)\ndef play_game(game_id, player_names, scripts, player_ids, colors, namespace, room, wave=1):\n\n cleared = True\n current_wave = wave - 1\n while cleared:\n current_wave += 1\n game = GameInstance(game_id, player_names, scripts, player_ids, colors, wave=current_wave)\n for batch in game.run():\n celery_socketio.emit('result', '{\\\"result\\\": [ ' + ','.join(batch) + ']}', room=room, namespace=namespace)\n cleared = game.was_cleared()\n\n # return highest reached wave\n return current_wave - 1\n\n@celery.task(time_limit=5, max_retries=3)\ndef test_round(game_id, player_names, scripts, player_ids, colors, namespace, room, wave=1):\n game = GameInstance(game_id, player_names, scripts, player_ids, colors, wave=wave)\n for batch in game.run():\n celery_socketio.emit('result', '{\\\"result\\\": [ ' + ','.join(batch) + ']}', room=room, namespace=namespace)\n\n # return if the test round was cleared\n return game.was_cleared()\n","subject":"Bump up max game time to 15.","message":"[NG] Bump up max game time to 15.\n","lang":"Python","license":"mit","repos":"SuddenDevs\/SuddenDev,SuddenDevs\/SuddenDev,SuddenDevs\/SuddenDev,SuddenDevs\/SuddenDev"} {"commit":"1b0428aaf77f1c6eadfb6b20611a2e2e6f30fbce","old_file":"poller.py","new_file":"poller.py","old_contents":"#!\/usr\/bin\/env python\nimport urllib2\nimport ssl\n\n\ndef poll(sites, timeout):\n for site in sites:\n print 'polling ' + site\n\n try:\n response = urllib2.urlopen(site, timeout=timeout)\n response.read()\n except urllib2.URLError as e:\n print e.code\n except ssl.SSLError as e:\n print e.message\n else:\n print 'ok'\n\nif __name__ == '__main__':\n poll(sites=(\n 'https:\/\/redmine.codegrove.org',\n 'http:\/\/koodilehto.fi',\n 'http:\/\/vakiopaine.net',\n ), timeout=5)\n","new_contents":"#!\/usr\/bin\/env python\nimport urllib2\nimport ssl\n\ntry:\n import gntp.notifier as notify\nexcept ImportError:\n notify = None\n\ndef poll(sites, timeout, ok, error):\n for site in sites:\n ok('polling ' + site)\n\n try:\n response = urllib2.urlopen(site, timeout=timeout)\n response.read()\n except urllib2.URLError as e:\n error(site + ' ' + e.code)\n except ssl.SSLError as e:\n error(site + ' ' + e.message)\n else:\n ok('ok')\n\n\ndef empty(data):\n pass\n\n\ndef output(data):\n if notify:\n notify.mini(data)\n else:\n print data\n\n\nif __name__ == '__main__':\n poll(sites=(\n 'https:\/\/redmine.codegrove.org',\n 'http:\/\/koodilehto.fi',\n 'http:\/\/vakiopaine.net',\n ), timeout=5, ok=empty, error=output)\n","subject":"Add initial support for growl","message":"Add initial support for growl\n\nIf growl lib isn't available, prints to console instead.\n","lang":"Python","license":"mit","repos":"koodilehto\/website-poller,koodilehto\/website-poller"} {"commit":"13f1f2643c6f00a3f890d735e1ae2c8ac757fe75","old_file":"setup.py","new_file":"setup.py","old_contents":"import sys\nimport codecs\n\nfrom setuptools import setup, find_packages\nfrom setuptools.command.test import test as TestCommand\n\nimport dmp\n\nsetup(\n name='dmp',\n version=dmp.__version__,\n description='MuG DMP API',\n url='http:\/\/www.multiscalegenomics.eu',\n download_url='https:\/\/github.com\/Multiscale-Genomics\/mg-dm-api',\n author=dmp.__author__,\n author_email='mcdowall@ebi.ac.uk',\n license=dmp.__license__,\n #packages=find_packages(),\n include_package_data=True,\n install_requires = [\n 'pymongo>=3.3', 'monogomock>=3.7.0'\n ]\n)\n","new_contents":"import sys\nimport codecs\n\nfrom setuptools import setup, find_packages\nfrom setuptools.command.test import test as TestCommand\n\nimport dmp\n\nsetup(\n name='dmp',\n version=dmp.__version__,\n description='MuG DMP API',\n url='http:\/\/www.multiscalegenomics.eu',\n download_url='https:\/\/github.com\/Multiscale-Genomics\/mg-dm-api',\n author=dmp.__author__,\n author_email='mcdowall@ebi.ac.uk',\n license=dmp.__license__,\n #packages=find_packages(),\n include_package_data=True,\n install_requires = [\n 'pymongo>=3.3', 'mongomock>=3.7'\n ]\n)\n","subject":"Fix to the name of the required module","message":"Fix to the name of the required module\n","lang":"Python","license":"apache-2.0","repos":"Multiscale-Genomics\/mg-dm-api,Multiscale-Genomics\/mg-dm-api"} {"commit":"1479e89d65c3d3c3b3ce3da9f81dae73e91b5ae3","old_file":"rororo\/__init__.py","new_file":"rororo\/__init__.py","old_contents":"\"\"\"\n======\nrororo\n======\n\nCollection of utilities, helpers, and principles for building Python backend\napplications. Supports `aiohttp.web `_,\n`Flask `_, and your web-framework.\n\n\"\"\"\n\n__author__ = 'Igor Davydenko'\n__license__ = 'BSD'\n__version__ = '1.0.dev0'\n","new_contents":"\"\"\"\n======\nrororo\n======\n\nCollection of utilities, helpers, and principles for building Python backend\napplications. Supports `aiohttp.web `_,\n`Flask `_, and your web-framework.\n\n\"\"\"\n\n__author__ = 'Igor Davydenko'\n__license__ = 'BSD'\n__version__ = '1.0.0.dev0'\n","subject":"Use semantic versioning for rororo.","message":"Use semantic versioning for rororo.\n","lang":"Python","license":"bsd-3-clause","repos":"playpauseandstop\/rororo,playpauseandstop\/rororo"} {"commit":"9967d14a10c0c53094b65e0946beb2d1c1200916","old_file":"python\/helpers\/pydev\/pydevd_attach_to_process\/linux\/lldb_threads_settrace.py","new_file":"python\/helpers\/pydev\/pydevd_attach_to_process\/linux\/lldb_threads_settrace.py","old_contents":"# This file is meant to be run inside lldb as a command after\n# the attach_linux.dylib dll has already been loaded to settrace for all threads.\ndef __lldb_init_module(debugger, internal_dict):\n # Command Initialization code goes here\n # print('Startup LLDB in Python!')\n import lldb\n\n try:\n show_debug_info = 0\n is_debug = 0\n target = debugger.GetSelectedTarget()\n if target:\n process = target.GetProcess()\n if process:\n for thread in process:\n # Get the first frame\n # print('Thread %s, suspended %s\\n'%(thread, thread.IsStopped()))\n\n process.SetSelectedThread(thread)\n\n if internal_dict.get('_thread_%d' % thread.GetThreadID(), False):\n frame = thread.GetSelectedFrame()\n if frame:\n print('Will settrace in: %s' % (frame,))\n res = frame.EvaluateExpression(\"(int) SetSysTraceFunc(%s, %s)\" % (\n show_debug_info, is_debug), lldb.eDynamicCanRunTarget)\n error = res.GetError()\n if error:\n print(error)\n thread.Resume()\n except:\n import traceback;traceback.print_exc()\n","new_contents":"# This file is meant to be run inside lldb as a command after\n# the attach_linux.dylib dll has already been loaded to settrace for all threads.\ndef __lldb_init_module(debugger, internal_dict):\n # Command Initialization code goes here\n # print('Startup LLDB in Python!')\n import lldb\n\n try:\n show_debug_info = 1\n is_debug = 0\n\n options = lldb.SBExpressionOptions()\n options.SetFetchDynamicValue()\n options.SetTryAllThreads(run_others=False)\n options.SetTimeoutInMicroSeconds(timeout=10000000)\n\n target = debugger.GetSelectedTarget()\n if target:\n process = target.GetProcess()\n if process:\n for thread in process:\n # Get the first frame\n # print('Thread %s, suspended %s\\n'%(thread, thread.IsStopped()))\n\n if internal_dict.get('_thread_%d' % thread.GetThreadID(), False):\n process.SetSelectedThread(thread)\n if not thread.IsStopped():\n # thread.Suspend()\n error = process.Stop()\n frame = thread.GetSelectedFrame()\n if frame:\n print('Will settrace in: %s' % (frame,))\n\n res = frame.EvaluateExpression(\"(int) SetSysTraceFunc(%s, %s)\" % (\n show_debug_info, is_debug), options)\n error = res.GetError()\n if error:\n print(error)\n\n thread.Resume()\n except:\n import traceback;traceback.print_exc()\n","subject":"Use timeout for lldb expr eval (PY-14252).","message":"Use timeout for lldb expr eval (PY-14252).\n","lang":"Python","license":"apache-2.0","repos":"fnouama\/intellij-community,samthor\/intellij-community,ftomassetti\/intellij-community,idea4bsd\/idea4bsd,dslomov\/intellij-community,ThiagoGarciaAlves\/intellij-community,izonder\/intellij-community,fnouama\/intellij-community,holmes\/intellij-community,MER-GROUP\/intellij-community,idea4bsd\/idea4bsd,ivan-fedorov\/intellij-community,SerCeMan\/intellij-community,ivan-fedorov\/intellij-community,kdwink\/intellij-community,dslomov\/intellij-community,petteyg\/intellij-community,allotria\/intellij-community,hurricup\/intellij-community,gnuhub\/intellij-community,jagguli\/intellij-community,mglukhikh\/intellij-community,ol-loginov\/intellij-community,blademainer\/intellij-community,ibinti\/intellij-community,vladmm\/intellij-community,asedunov\/intellij-community,orekyuu\/intellij-community,muntasirsyed\/intellij-community,dslomov\/intellij-community,ThiagoGarciaAlves\/intellij-community,ol-loginov\/intellij-community,dslomov\/intellij-community,muntasirsyed\/intellij-community,apixandru\/intellij-community,kool79\/intellij-community,vvv1559\/intellij-community,Lekanich\/intellij-community,youdonghai\/intellij-community,Lekanich\/intellij-community,MER-GROUP\/intellij-community,muntasirsyed\/intellij-community,da1z\/intellij-community,amith01994\/intellij-community,akosyakov\/intellij-community,michaelgallacher\/intellij-community,retomerz\/intellij-community,dslomov\/intellij-community,orekyuu\/intellij-community,salguarnieri\/intellij-community,allotria\/intellij-community,da1z\/intellij-community,Distrotech\/intellij-community,youdonghai\/intellij-community,retomerz\/intellij-community,dslomov\/intellij-community,izonder\/intellij-community,Distrotech\/intellij-community,samthor\/intellij-community,kool79\/intellij-community,adedayo\/intellij-community,akosyakov\/intellij-community,caot\/intellij-community,retomerz\/intellij-community,xfournet\/intellij-community,nicolargo\/intellij-community,ibinti\/intellij-community,supersven\/intellij-community,idea4bsd\/idea4bsd,wreckJ\/intellij-community,hurricup\/intellij-community,TangHao1987\/intellij-community,vvv1559\/intellij-community,lucafavatella\/intellij-community,slisson\/intellij-community,ibinti\/intellij-community,gnuhub\/intellij-community,retomerz\/intellij-community,akosyakov\/intellij-community,fengbaicanhe\/intellij-community,kdwink\/intellij-community,clumsy\/intellij-community,MER-GROUP\/intellij-community,adedayo\/intellij-community,ryano144\/intellij-community,pwoodworth\/intellij-community,ahb0327\/intellij-community,fitermay\/intellij-community,ftomassetti\/intellij-community,ivan-fedorov\/intellij-community,adedayo\/intellij-community,apixandru\/intellij-community,xfournet\/intellij-community,izonder\/intellij-community,jagguli\/intellij-community,kdwink\/intellij-community,wreckJ\/intellij-community,vvv1559\/intellij-community,blademainer\/intellij-community,michaelgallacher\/intellij-community,MER-GROUP\/intellij-community,ryano144\/intellij-community,orekyuu\/intellij-community,izonder\/intellij-community,fengbaicanhe\/intellij-community,muntasirsyed\/intellij-community,ibinti\/intellij-community,orekyuu\/intellij-community,wreckJ\/intellij-community,orekyuu\/intellij-community,TangHao1987\/intellij-community,MichaelNedzelsky\/intellij-community,vvv1559\/intellij-community,lucafavatella\/intellij-community,allotria\/intellij-community,ivan-fedorov\/intellij-community,allotria\/intellij-community,lucafavatella\/intellij-community,kdwink\/intellij-community,MichaelNedzelsky\/intellij-community,ol-loginov\/intellij-community,samthor\/intellij-community,ftomassetti\/intellij-community,fnouama\/intellij-community,izonder\/intellij-community,lucafavatella\/intellij-community,youdonghai\/intellij-community,vvv1559\/intellij-community,vladmm\/intellij-community,ftomassetti\/intellij-community,fitermay\/intellij-community,retomerz\/intellij-community,petteyg\/intellij-community,ahb0327\/intellij-community,pwoodworth\/intellij-community,fnouama\/intellij-community,youdonghai\/intellij-community,akosyakov\/intellij-community,fitermay\/intellij-community,tmpgit\/intellij-community,ftomassetti\/intellij-community,TangHao1987\/intellij-community,tmpgit\/intellij-community,jagguli\/intellij-community,diorcety\/intellij-community,holmes\/intellij-community,ibinti\/intellij-community,diorcety\/intellij-community,nicolargo\/intellij-community,suncycheng\/intellij-community,vladmm\/intellij-community,TangHao1987\/intellij-community,blademainer\/intellij-community,jagguli\/intellij-community,fitermay\/intellij-community,wreckJ\/intellij-community,robovm\/robovm-studio,wreckJ\/intellij-community,petteyg\/intellij-community,mglukhikh\/intellij-community,youdonghai\/intellij-community,ftomassetti\/intellij-community,muntasirsyed\/intellij-community,da1z\/intellij-community,izonder\/intellij-community,fitermay\/intellij-community,nicolargo\/intellij-community,michaelgallacher\/intellij-community,MichaelNedzelsky\/intellij-community,vvv1559\/intellij-community,semonte\/intellij-community,orekyuu\/intellij-community,FHannes\/intellij-community,holmes\/intellij-community,supersven\/intellij-community,fengbaicanhe\/intellij-community,Distrotech\/intellij-community,ahb0327\/intellij-community,signed\/intellij-community,ibinti\/intellij-community,supersven\/intellij-community,apixandru\/intellij-community,adedayo\/intellij-community,clumsy\/intellij-community,alphafoobar\/intellij-community,youdonghai\/intellij-community,suncycheng\/intellij-community,fitermay\/intellij-community,salguarnieri\/intellij-community,petteyg\/intellij-community,gnuhub\/intellij-community,apixandru\/intellij-community,ibinti\/intellij-community,ryano144\/intellij-community,apixandru\/intellij-community,pwoodworth\/intellij-community,nicolargo\/intellij-community,diorcety\/intellij-community,slisson\/intellij-community,xfournet\/intellij-community,robovm\/robovm-studio,fengbaicanhe\/intellij-community,ahb0327\/intellij-community,TangHao1987\/intellij-community,ftomassetti\/intellij-community,youdonghai\/intellij-community,akosyakov\/intellij-community,fitermay\/intellij-community,retomerz\/intellij-community,caot\/intellij-community,ThiagoGarciaAlves\/intellij-community,MER-GROUP\/intellij-community,gnuhub\/intellij-community,kdwink\/intellij-community,kdwink\/intellij-community,ivan-fedorov\/intellij-community,dslomov\/intellij-community,petteyg\/intellij-community,mglukhikh\/intellij-community,hurricup\/intellij-community,samthor\/intellij-community,SerCeMan\/intellij-community,mglukhikh\/intellij-community,apixandru\/intellij-community,youdonghai\/intellij-community,akosyakov\/intellij-community,signed\/intellij-community,petteyg\/intellij-community,semonte\/intellij-community,clumsy\/intellij-community,alphafoobar\/intellij-community,salguarnieri\/intellij-community,idea4bsd\/idea4bsd,semonte\/intellij-community,samthor\/intellij-community,caot\/intellij-community,Lekanich\/intellij-community,samthor\/intellij-community,kdwink\/intellij-community,MER-GROUP\/intellij-community,gnuhub\/intellij-community,blademainer\/intellij-community,SerCeMan\/intellij-community,signed\/intellij-community,supersven\/intellij-community,izonder\/intellij-community,samthor\/intellij-community,pwoodworth\/intellij-community,blademainer\/intellij-community,semonte\/intellij-community,ryano144\/intellij-community,xfournet\/intellij-community,Lekanich\/intellij-community,xfournet\/intellij-community,MichaelNedzelsky\/intellij-community,allotria\/intellij-community,petteyg\/intellij-community,wreckJ\/intellij-community,michaelgallacher\/intellij-community,caot\/intellij-community,ibinti\/intellij-community,adedayo\/intellij-community,tmpgit\/intellij-community,hurricup\/intellij-community,mglukhikh\/intellij-community,jagguli\/intellij-community,MER-GROUP\/intellij-community,amith01994\/intellij-community,semonte\/intellij-community,da1z\/intellij-community,clumsy\/intellij-community,robovm\/robovm-studio,kool79\/intellij-community,ol-loginov\/intellij-community,signed\/intellij-community,ryano144\/intellij-community,amith01994\/intellij-community,ryano144\/intellij-community,salguarnieri\/intellij-community,fitermay\/intellij-community,hurricup\/intellij-community,muntasirsyed\/intellij-community,jagguli\/intellij-community,idea4bsd\/idea4bsd,apixandru\/intellij-community,kool79\/intellij-community,supersven\/intellij-community,ryano144\/intellij-community,robovm\/robovm-studio,orekyuu\/intellij-community,ryano144\/intellij-community,hurricup\/intellij-community,adedayo\/intellij-community,SerCeMan\/intellij-community,kdwink\/intellij-community,asedunov\/intellij-community,vladmm\/intellij-community,pwoodworth\/intellij-community,adedayo\/intellij-community,mglukhikh\/intellij-community,pwoodworth\/intellij-community,fitermay\/intellij-community,muntasirsyed\/intellij-community,ThiagoGarciaAlves\/intellij-community,nicolargo\/intellij-community,da1z\/intellij-community,holmes\/intellij-community,robovm\/robovm-studio,vladmm\/intellij-community,samthor\/intellij-community,MichaelNedzelsky\/intellij-community,tmpgit\/intellij-community,MichaelNedzelsky\/intellij-community,idea4bsd\/idea4bsd,petteyg\/intellij-community,semonte\/intellij-community,salguarnieri\/intellij-community,slisson\/intellij-community,blademainer\/intellij-community,muntasirsyed\/intellij-community,amith01994\/intellij-community,Distrotech\/intellij-community,vladmm\/intellij-community,petteyg\/intellij-community,pwoodworth\/intellij-community,ryano144\/intellij-community,Distrotech\/intellij-community,petteyg\/intellij-community,mglukhikh\/intellij-community,supersven\/intellij-community,nicolargo\/intellij-community,holmes\/intellij-community,salguarnieri\/intellij-community,muntasirsyed\/intellij-community,akosyakov\/intellij-community,tmpgit\/intellij-community,ftomassetti\/intellij-community,youdonghai\/intellij-community,petteyg\/intellij-community,dslomov\/intellij-community,xfournet\/intellij-community,suncycheng\/intellij-community,asedunov\/intellij-community,ThiagoGarciaAlves\/intellij-community,lucafavatella\/intellij-community,samthor\/intellij-community,slisson\/intellij-community,ibinti\/intellij-community,xfournet\/intellij-community,dslomov\/intellij-community,izonder\/intellij-community,kool79\/intellij-community,Lekanich\/intellij-community,amith01994\/intellij-community,holmes\/intellij-community,caot\/intellij-community,signed\/intellij-community,apixandru\/intellij-community,ivan-fedorov\/intellij-community,ThiagoGarciaAlves\/intellij-community,michaelgallacher\/intellij-community,hurricup\/intellij-community,ThiagoGarciaAlves\/intellij-community,fnouama\/intellij-community,MER-GROUP\/intellij-community,gnuhub\/intellij-community,Distrotech\/intellij-community,holmes\/intellij-community,youdonghai\/intellij-community,tmpgit\/intellij-community,vladmm\/intellij-community,blademainer\/intellij-community,ibinti\/intellij-community,suncycheng\/intellij-community,wreckJ\/intellij-community,idea4bsd\/idea4bsd,diorcety\/intellij-community,mglukhikh\/intellij-community,apixandru\/intellij-community,ryano144\/intellij-community,salguarnieri\/intellij-community,Distrotech\/intellij-community,wreckJ\/intellij-community,robovm\/robovm-studio,FHannes\/intellij-community,FHannes\/intellij-community,alphafoobar\/intellij-community,ol-loginov\/intellij-community,vvv1559\/intellij-community,allotria\/intellij-community,MichaelNedzelsky\/intellij-community,slisson\/intellij-community,TangHao1987\/intellij-community,Lekanich\/intellij-community,supersven\/intellij-community,ivan-fedorov\/intellij-community,signed\/intellij-community,michaelgallacher\/intellij-community,wreckJ\/intellij-community,gnuhub\/intellij-community,idea4bsd\/idea4bsd,gnuhub\/intellij-community,salguarnieri\/intellij-community,youdonghai\/intellij-community,ivan-fedorov\/intellij-community,idea4bsd\/idea4bsd,fengbaicanhe\/intellij-community,wreckJ\/intellij-community,diorcety\/intellij-community,suncycheng\/intellij-community,xfournet\/intellij-community,muntasirsyed\/intellij-community,ftomassetti\/intellij-community,holmes\/intellij-community,signed\/intellij-community,alphafoobar\/intellij-community,kool79\/intellij-community,diorcety\/intellij-community,xfournet\/intellij-community,suncycheng\/intellij-community,ivan-fedorov\/intellij-community,hurricup\/intellij-community,xfournet\/intellij-community,jagguli\/intellij-community,dslomov\/intellij-community,TangHao1987\/intellij-community,pwoodworth\/intellij-community,fitermay\/intellij-community,da1z\/intellij-community,alphafoobar\/intellij-community,amith01994\/intellij-community,vladmm\/intellij-community,asedunov\/intellij-community,semonte\/intellij-community,ol-loginov\/intellij-community,FHannes\/intellij-community,retomerz\/intellij-community,clumsy\/intellij-community,jagguli\/intellij-community,michaelgallacher\/intellij-community,diorcety\/intellij-community,adedayo\/intellij-community,nicolargo\/intellij-community,fengbaicanhe\/intellij-community,diorcety\/intellij-community,asedunov\/intellij-community,ol-loginov\/intellij-community,MER-GROUP\/intellij-community,MER-GROUP\/intellij-community,suncycheng\/intellij-community,akosyakov\/intellij-community,asedunov\/intellij-community,holmes\/intellij-community,ThiagoGarciaAlves\/intellij-community,ibinti\/intellij-community,signed\/intellij-community,Lekanich\/intellij-community,lucafavatella\/intellij-community,vladmm\/intellij-community,Distrotech\/intellij-community,xfournet\/intellij-community,robovm\/robovm-studio,caot\/intellij-community,caot\/intellij-community,youdonghai\/intellij-community,alphafoobar\/intellij-community,clumsy\/intellij-community,michaelgallacher\/intellij-community,lucafavatella\/intellij-community,signed\/intellij-community,alphafoobar\/intellij-community,lucafavatella\/intellij-community,da1z\/intellij-community,SerCeMan\/intellij-community,akosyakov\/intellij-community,SerCeMan\/intellij-community,nicolargo\/intellij-community,alphafoobar\/intellij-community,asedunov\/intellij-community,slisson\/intellij-community,clumsy\/intellij-community,kool79\/intellij-community,samthor\/intellij-community,gnuhub\/intellij-community,lucafavatella\/intellij-community,ThiagoGarciaAlves\/intellij-community,FHannes\/intellij-community,nicolargo\/intellij-community,asedunov\/intellij-community,idea4bsd\/idea4bsd,Distrotech\/intellij-community,salguarnieri\/intellij-community,fnouama\/intellij-community,slisson\/intellij-community,caot\/intellij-community,fitermay\/intellij-community,ahb0327\/intellij-community,apixandru\/intellij-community,amith01994\/intellij-community,TangHao1987\/intellij-community,blademainer\/intellij-community,caot\/intellij-community,signed\/intellij-community,ahb0327\/intellij-community,Distrotech\/intellij-community,lucafavatella\/intellij-community,alphafoobar\/intellij-community,gnuhub\/intellij-community,vvv1559\/intellij-community,dslomov\/intellij-community,ThiagoGarciaAlves\/intellij-community,holmes\/intellij-community,suncycheng\/intellij-community,ftomassetti\/intellij-community,orekyuu\/intellij-community,ftomassetti\/intellij-community,samthor\/intellij-community,da1z\/intellij-community,idea4bsd\/idea4bsd,ryano144\/intellij-community,asedunov\/intellij-community,asedunov\/intellij-community,Lekanich\/intellij-community,alphafoobar\/intellij-community,ibinti\/intellij-community,TangHao1987\/intellij-community,gnuhub\/intellij-community,pwoodworth\/intellij-community,pwoodworth\/intellij-community,wreckJ\/intellij-community,kool79\/intellij-community,supersven\/intellij-community,suncycheng\/intellij-community,xfournet\/intellij-community,robovm\/robovm-studio,FHannes\/intellij-community,retomerz\/intellij-community,SerCeMan\/intellij-community,clumsy\/intellij-community,TangHao1987\/intellij-community,ahb0327\/intellij-community,apixandru\/intellij-community,fengbaicanhe\/intellij-community,tmpgit\/intellij-community,SerCeMan\/intellij-community,da1z\/intellij-community,vvv1559\/intellij-community,allotria\/intellij-community,suncycheng\/intellij-community,jagguli\/intellij-community,fnouama\/intellij-community,holmes\/intellij-community,lucafavatella\/intellij-community,jagguli\/intellij-community,fitermay\/intellij-community,adedayo\/intellij-community,fnouama\/intellij-community,ol-loginov\/intellij-community,FHannes\/intellij-community,vvv1559\/intellij-community,MichaelNedzelsky\/intellij-community,ftomassetti\/intellij-community,supersven\/intellij-community,clumsy\/intellij-community,Lekanich\/intellij-community,MichaelNedzelsky\/intellij-community,hurricup\/intellij-community,slisson\/intellij-community,fengbaicanhe\/intellij-community,kool79\/intellij-community,tmpgit\/intellij-community,diorcety\/intellij-community,diorcety\/intellij-community,mglukhikh\/intellij-community,adedayo\/intellij-community,slisson\/intellij-community,tmpgit\/intellij-community,robovm\/robovm-studio,Lekanich\/intellij-community,fitermay\/intellij-community,blademainer\/intellij-community,FHannes\/intellij-community,muntasirsyed\/intellij-community,hurricup\/intellij-community,caot\/intellij-community,orekyuu\/intellij-community,allotria\/intellij-community,retomerz\/intellij-community,adedayo\/intellij-community,supersven\/intellij-community,adedayo\/intellij-community,fnouama\/intellij-community,ibinti\/intellij-community,asedunov\/intellij-community,ol-loginov\/intellij-community,mglukhikh\/intellij-community,apixandru\/intellij-community,semonte\/intellij-community,ahb0327\/intellij-community,muntasirsyed\/intellij-community,akosyakov\/intellij-community,apixandru\/intellij-community,caot\/intellij-community,SerCeMan\/intellij-community,kdwink\/intellij-community,ol-loginov\/intellij-community,gnuhub\/intellij-community,hurricup\/intellij-community,asedunov\/intellij-community,diorcety\/intellij-community,Distrotech\/intellij-community,allotria\/intellij-community,FHannes\/intellij-community,mglukhikh\/intellij-community,retomerz\/intellij-community,suncycheng\/intellij-community,izonder\/intellij-community,dslomov\/intellij-community,vvv1559\/intellij-community,amith01994\/intellij-community,supersven\/intellij-community,MER-GROUP\/intellij-community,amith01994\/intellij-community,nicolargo\/intellij-community,kdwink\/intellij-community,xfournet\/intellij-community,blademainer\/intellij-community,fengbaicanhe\/intellij-community,jagguli\/intellij-community,lucafavatella\/intellij-community,slisson\/intellij-community,mglukhikh\/intellij-community,petteyg\/intellij-community,Lekanich\/intellij-community,semonte\/intellij-community,nicolargo\/intellij-community,izonder\/intellij-community,ivan-fedorov\/intellij-community,ivan-fedorov\/intellij-community,kdwink\/intellij-community,youdonghai\/intellij-community,michaelgallacher\/intellij-community,clumsy\/intellij-community,SerCeMan\/intellij-community,holmes\/intellij-community,semonte\/intellij-community,FHannes\/intellij-community,michaelgallacher\/intellij-community,fnouama\/intellij-community,semonte\/intellij-community,FHannes\/intellij-community,akosyakov\/intellij-community,hurricup\/intellij-community,robovm\/robovm-studio,vladmm\/intellij-community,amith01994\/intellij-community,kool79\/intellij-community,pwoodworth\/intellij-community,allotria\/intellij-community,semonte\/intellij-community,asedunov\/intellij-community,hurricup\/intellij-community,michaelgallacher\/intellij-community,Distrotech\/intellij-community,retomerz\/intellij-community,ivan-fedorov\/intellij-community,blademainer\/intellij-community,da1z\/intellij-community,idea4bsd\/idea4bsd,da1z\/intellij-community,orekyuu\/intellij-community,orekyuu\/intellij-community,vvv1559\/intellij-community,alphafoobar\/intellij-community,semonte\/intellij-community,fengbaicanhe\/intellij-community,Lekanich\/intellij-community,allotria\/intellij-community,ol-loginov\/intellij-community,wreckJ\/intellij-community,lucafavatella\/intellij-community,ol-loginov\/intellij-community,signed\/intellij-community,tmpgit\/intellij-community,supersven\/intellij-community,vvv1559\/intellij-community,samthor\/intellij-community,allotria\/intellij-community,vladmm\/intellij-community,fengbaicanhe\/intellij-community,akosyakov\/intellij-community,FHannes\/intellij-community,amith01994\/intellij-community,suncycheng\/intellij-community,izonder\/intellij-community,kdwink\/intellij-community,jagguli\/intellij-community,retomerz\/intellij-community,ahb0327\/intellij-community,MichaelNedzelsky\/intellij-community,kool79\/intellij-community,pwoodworth\/intellij-community,allotria\/intellij-community,fengbaicanhe\/intellij-community,caot\/intellij-community,amith01994\/intellij-community,ryano144\/intellij-community,ahb0327\/intellij-community,MichaelNedzelsky\/intellij-community,slisson\/intellij-community,ThiagoGarciaAlves\/intellij-community,MichaelNedzelsky\/intellij-community,diorcety\/intellij-community,tmpgit\/intellij-community,signed\/intellij-community,retomerz\/intellij-community,ahb0327\/intellij-community,MER-GROUP\/intellij-community,TangHao1987\/intellij-community,TangHao1987\/intellij-community,FHannes\/intellij-community,apixandru\/intellij-community,slisson\/intellij-community,mglukhikh\/intellij-community,tmpgit\/intellij-community,nicolargo\/intellij-community,robovm\/robovm-studio,da1z\/intellij-community,clumsy\/intellij-community,michaelgallacher\/intellij-community,izonder\/intellij-community,da1z\/intellij-community,salguarnieri\/intellij-community,kool79\/intellij-community,alphafoobar\/intellij-community,fnouama\/intellij-community,signed\/intellij-community,blademainer\/intellij-community,idea4bsd\/idea4bsd,salguarnieri\/intellij-community,vladmm\/intellij-community,robovm\/robovm-studio,SerCeMan\/intellij-community,fnouama\/intellij-community,salguarnieri\/intellij-community,ThiagoGarciaAlves\/intellij-community,orekyuu\/intellij-community,clumsy\/intellij-community,ahb0327\/intellij-community,SerCeMan\/intellij-community"} {"commit":"90012f9fb9a256e6086a0b421661fd74cd8ef880","old_file":"sedlex\/AddCocoricoVoteVisitor.py","new_file":"sedlex\/AddCocoricoVoteVisitor.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom AbstractVisitor import AbstractVisitor\n\nfrom duralex.alinea_parser import *\n\nimport requests\n\nclass AddCocoricoVoteVisitor(AbstractVisitor):\n def __init__(self, args):\n self.url = 'https:\/\/local.cocorico.cc'\n\n r = requests.post(\n self.url + '\/api\/oauth\/token',\n auth=(args.cocorico_app_id, args.cocorico_secret),\n data={ 'grant_type': 'client_credentials' },\n verify=self.url != 'https:\/\/local.cocorico.cc'\n )\n self.access_token = r.json()['access_token']\n\n super(AddCocoricoVoteVisitor, self).__init__()\n\n def visit_node(self, node):\n if not self.access_token:\n return\n\n # if on root node\n if 'parent' not in node and 'type' not in node:\n r = requests.post(\n self.url + '\/api\/vote',\n headers={'Authorization': 'Bearer ' + self.access_token},\n data={\n 'title': 'test de vote',\n 'description': 'ceci est un test',\n 'url': 'https:\/\/legilibre.fr\/?test=49'\n },\n verify=self.url != 'https:\/\/local.cocorico.cc'\n )\n node['cocoricoVote'] = r.json()['vote']['id']\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom AbstractVisitor import AbstractVisitor\n\nfrom duralex.alinea_parser import *\n\nimport requests\n\nclass AddCocoricoVoteVisitor(AbstractVisitor):\n def __init__(self, args):\n self.url = args.cocorico_url\n if not self.url:\n self.url = 'https:\/\/cocorico.cc'\n\n r = requests.post(\n self.url + '\/api\/oauth\/token',\n auth=(args.cocorico_app_id, args.cocorico_secret),\n data={ 'grant_type': 'client_credentials' },\n verify=self.url != 'https:\/\/local.cocorico.cc'\n )\n self.access_token = r.json()['access_token']\n\n super(AddCocoricoVoteVisitor, self).__init__()\n\n def visit_node(self, node):\n if not self.access_token:\n return\n\n # if on root node\n if 'parent' not in node and 'type' not in node:\n r = requests.post(\n self.url + '\/api\/vote',\n headers={'Authorization': 'Bearer ' + self.access_token},\n data={\n 'title': 'test de vote',\n 'description': 'ceci est un test',\n 'url': 'https:\/\/legilibre.fr\/?test=49'\n },\n verify=self.url != 'https:\/\/local.cocorico.cc'\n )\n node['cocoricoVote'] = r.json()['vote']['id']\n","subject":"Handle the --cocorico-url command line option.","message":"Handle the --cocorico-url command line option.\n","lang":"Python","license":"agpl-3.0","repos":"Legilibre\/SedLex"} {"commit":"4c5cf98be65ee2564062cce2a43b7833eef1a6c9","old_file":"AFQ\/utils\/volume.py","new_file":"AFQ\/utils\/volume.py","old_contents":"import scipy.ndimage as ndim\nfrom skimage.filters import gaussian\n\n\ndef patch_up_roi(roi, sigma=0.5, truncate=2):\n \"\"\"\n After being non-linearly transformed, ROIs tend to have holes in them.\n We perform a couple of computational geometry operations on the ROI to\n fix that up.\n\n Parameters\n ----------\n roi : 3D binary array\n The ROI after it has been transformed.\n\n sigma : float\n The sigma for initial Gaussian smoothing.\n\n truncate : float\n The truncation for the Gaussian\n\n Returns\n -------\n ROI after dilation and hole-filling\n \"\"\"\n\n return (ndim.binary_fill_holes(\n gaussian(roi, sigma=sigma, truncate=truncate)).astype(float) > 0.1)\n","new_contents":"import scipy.ndimage as ndim\nfrom skimage.filters import gaussian\nfrom skimage.morphology import convex_hull_image\n\ndef patch_up_roi(roi, sigma=0.5, truncate=2):\n \"\"\"\n After being non-linearly transformed, ROIs tend to have holes in them.\n We perform a couple of computational geometry operations on the ROI to\n fix that up.\n\n Parameters\n ----------\n roi : 3D binary array\n The ROI after it has been transformed.\n\n sigma : float\n The sigma for initial Gaussian smoothing.\n\n truncate : float\n The truncation for the Gaussian\n\n Returns\n -------\n ROI after dilation and hole-filling\n \"\"\"\n\n return convex_hull_image(gaussian(ndim.binary_fill_holes(roi),\n sigma=sigma, truncate=truncate) > 0.1)\n","subject":"Add a convex hull operation to really close this up.","message":"Add a convex hull operation to really close this up.\n","lang":"Python","license":"bsd-2-clause","repos":"yeatmanlab\/pyAFQ,arokem\/pyAFQ,arokem\/pyAFQ,yeatmanlab\/pyAFQ"} {"commit":"788229f43eab992d6f4d79681604336e4d721b0c","old_file":"gameserver\/api\/endpoints\/players.py","new_file":"gameserver\/api\/endpoints\/players.py","old_contents":"import logging\n\nfrom flask import request\nfrom flask_restplus import Resource\n\nfrom gameserver.game import Game\nfrom gameserver.models import Player\n\nfrom gameserver.api.restplus import api\nfrom gameserver.api.serializers import player_get, player_post\n\nfrom gameserver.database import db\n\ndb_session = db.session\n\nlog = logging.getLogger(__name__)\n\nns = api.namespace('players', description='Operations related to players')\n\ngame = Game()\n\n@ns.route('\/')\nclass PlayerCollection(Resource):\n\n @api.response(200, 'Success')\n @api.marshal_list_with(player_get)\n def get(self):\n \"\"\"\n Returns list of players.\n \"\"\"\n players = game.get_players()\n return players\n\n @api.response(201, 'Player successfully created.')\n @api.expect(player_post)\n def post(self):\n \"\"\"\n Creates a new game player.\n \"\"\"\n data = request.json\n player = game.create_player(data['name'])\n db_session.commit()\n return player.id, 201\n\n\n@ns.route('\/')\n@ns.param('id', 'The player id')\nclass Player(Resource):\n\n @api.response(404, 'Player not found')\n @api.response(200, 'Success')\n @api.marshal_with(player_get)\n def get(self, id):\n \"\"\" \n Returns the specified player. \n \"\"\"\n player = game.get_player(id)\n if not player:\n api.abort(404)\n return player\n\n\n\n","new_contents":"import logging\n\nfrom flask import request\nfrom flask_restplus import Resource\n\nfrom gameserver.game import Game\nfrom gameserver.models import Player\n\nfrom gameserver.api.restplus import api\nfrom gameserver.api.serializers import player_get, player_post\n\nfrom gameserver.database import db\n\ndb_session = db.session\n\nlog = logging.getLogger(__name__)\n\nns = api.namespace('players', description='Operations related to players')\n\ngame = Game()\n\n@ns.route('\/')\nclass PlayerCollection(Resource):\n\n @api.response(200, 'Success')\n @api.marshal_list_with(player_get)\n def get(self):\n \"\"\"\n Returns list of players.\n \"\"\"\n players = game.get_players()\n return players\n\n @api.response(201, 'Player successfully created.')\n @api.expect(player_post)\n def post(self):\n \"\"\"\n Creates a new game player.\n \"\"\"\n data = request.json\n player = game.create_player(data['name'])\n db_session.commit()\n return dict(id=player.id), 201\n\n\n@ns.route('\/')\n@ns.param('id', 'The player id')\nclass Player(Resource):\n\n @api.response(404, 'Player not found')\n @api.response(200, 'Success')\n @api.marshal_with(player_get)\n def get(self, id):\n \"\"\"\n Returns the specified player.\n \"\"\"\n player = game.get_player(id)\n if not player:\n api.abort(404)\n return player\n\n\n\n","subject":"Return json, not just a string id","message":"Return json, not just a string id\n","lang":"Python","license":"apache-2.0","repos":"hammertoe\/didactic-spork,hammertoe\/didactic-spork,hammertoe\/didactic-spork,hammertoe\/didactic-spork"} {"commit":"99e164e2e4cf732485e6692a67c732cb657a1c11","old_file":"test\/util.py","new_file":"test\/util.py","old_contents":"'''Helper code for theanets unit tests.'''\n\nimport numpy as np\nimport skdata.mnist\n\n\nclass MNIST:\n NUM_DIGITS = 100\n DIGIT_SIZE = 784\n\n def setUp(self):\n np.random.seed(3)\n\n mnist = skdata.mnist.dataset.MNIST()\n mnist.meta # trigger download if needed.\n def arr(n, dtype):\n arr = mnist.arrays[n]\n return arr.reshape((len(arr), -1)).astype(dtype)\n self.images = arr('train_images', 'f')[:MNIST.NUM_DIGITS] \/ 255.\n self.labels = arr('train_labels', 'b')[:MNIST.NUM_DIGITS]\n","new_contents":"'''Helper code for theanets unit tests.'''\n\nimport numpy as np\nimport skdata.mnist\n\n\nclass MNIST(object):\n NUM_DIGITS = 100\n DIGIT_SIZE = 784\n\n def setUp(self):\n np.random.seed(3)\n\n mnist = skdata.mnist.dataset.MNIST()\n mnist.meta # trigger download if needed.\n def arr(n, dtype):\n arr = mnist.arrays[n]\n return arr.reshape((len(arr), -1)).astype(dtype)\n self.images = arr('train_images', 'f')[:MNIST.NUM_DIGITS] \/ 255.\n self.labels = arr('train_labels', 'b')[:MNIST.NUM_DIGITS]\n","subject":"Make MNIST inherit from object for py2k.","message":"Make MNIST inherit from object for py2k.\n","lang":"Python","license":"mit","repos":"devdoer\/theanets,chrinide\/theanets,lmjohns3\/theanets"} {"commit":"9e110d998532600a351ae35f79b0f46bce576c9f","old_file":"humfrey\/desc\/rdf_processors.py","new_file":"humfrey\/desc\/rdf_processors.py","old_contents":"import rdflib\n\nfrom humfrey.utils.namespaces import NS\nfrom humfrey.linkeddata.uri import doc_forwards\n\ndef formats(request, context):\n graph, subject, doc_uri = context['graph'], context['subject'], context['doc_uri']\n formats_for_context = []\n for renderer in context['renderers']:\n url = rdflib.URIRef(renderer['url'])\n graph += [\n (doc_uri, NS['dcterms'].hasFormat, url),\n (url, NS['dcterms']['title'], rdflib.Literal('%s description of %s' % (renderer['name'], subject.label))),\n ]\n graph += [(url, NS['dc']['format'], rdflib.Literal(mimetype)) for mimetype in renderer['mimetypes']]\n\n formats_for_context.sort(key=lambda f:f['name'])\n return {\n 'formats': formats_for_context,\n }\n\ndef doc_meta(request, context):\n doc_uri = context['doc_uri']\n context['graph'] += [\n (doc_uri, NS['foaf'].primaryTopic, context['subject']._identifier),\n (doc_uri, NS['rdf'].type, NS['foaf'].Document),\n (doc_uri, NS['dcterms']['title'], rdflib.Literal('Description of {0}'.format(context['subject'].label)),)\n ]\n","new_contents":"import rdflib\n\nfrom humfrey.utils.namespaces import NS\nfrom humfrey.linkeddata.uri import doc_forwards\n\ndef formats(request, context):\n graph, subject, doc_uri = context['graph'], context['subject'], context['doc_uri']\n formats_for_context = []\n for renderer in context['renderers']:\n url = rdflib.URIRef(renderer['url'])\n graph += [\n (doc_uri, NS['dcterms'].hasFormat, url),\n (url, NS['dcterms']['title'], rdflib.Literal('%s description of %s' % (renderer['name'], subject.label))),\n ]\n graph += [(url, NS['dc']['format'], rdflib.Literal(mimetype)) for mimetype in renderer['mimetypes']]\n\n formats_for_context.sort(key=lambda f:f['name'])\n return {\n 'formats': formats_for_context,\n }\n\ndef doc_meta(request, context):\n doc_uri = context['doc_uri']\n context['graph'] += [\n (doc_uri, NS['foaf'].primaryTopic, context['subject']._identifier),\n (doc_uri, NS['rdf'].type, NS['foaf'].Document),\n (doc_uri, NS['dcterms']['title'], rdflib.Literal(u'Description of {0}'.format(context['subject'].label)),)\n ]\n","subject":"Fix ascii-only bug when something's label contains non-ascii characters.","message":"Fix ascii-only bug when something's label contains non-ascii characters.\n","lang":"Python","license":"bsd-3-clause","repos":"ox-it\/humfrey,ox-it\/humfrey,ox-it\/humfrey"} {"commit":"a7e64e1594ffc1dba32e55f99779c79d85fd05b9","old_file":"app.py","new_file":"app.py","old_contents":"from flask import Flask, render_template, flash\r\nfrom flask_wtf import Form\r\nfrom flask_wtf.file import FileField\r\nfrom tools import s3_upload, store_locally\r\nimport json \r\n\r\napp = Flask(__name__)\r\napp.config.from_object('config')\r\n\r\n\r\nclass UploadForm(Form):\r\n example = FileField('Example File')\r\n\r\n\r\n@app.route('\/', methods=['POST', 'GET'])\r\ndef upload_page():\r\n form = UploadForm(csrf_enabled=False)\r\n upload_file = form.example\r\n if form.validate_on_submit():\r\n output = store_locally(upload_file)\r\n response = {}\r\n if output is not None:\r\n response['url'] = output\r\n return json.dumps(response, indent=4)\r\n else:\r\n response['url'] = None\r\n return json.dumps(response, indent=4), app.config[\"INVALID_DATA\"] \r\n return render_template('example.html', form=form) \r\n\r\nif __name__ == '__main__':\r\n app.run(host=app.config[\"HOST\"], port=app.config[\"PORT\"], debug=app.config[\"DEBUG\"])\r\n","new_contents":"from os import environ\r\nfrom flask import Flask, render_template, flash\r\nfrom flask_wtf import Form\r\nfrom flask_wtf.file import FileField\r\nfrom tools import s3_upload, store_locally\r\nimport json \r\n\r\napp = Flask(__name__)\r\napp.config.from_object('config')\r\n\r\n\r\nclass UploadForm(Form):\r\n example = FileField('Example File')\r\n\r\n\r\n@app.route('\/', methods=['POST', 'GET'])\r\ndef upload_page():\r\n form = UploadForm(csrf_enabled=False)\r\n upload_file = form.example\r\n if form.validate_on_submit():\r\n output = store_locally(upload_file)\r\n response = {}\r\n if output is not None:\r\n response['url'] = output\r\n return json.dumps(response, indent=4)\r\n else:\r\n response['url'] = None\r\n return json.dumps(response, indent=4), app.config[\"INVALID_DATA\"] \r\n return render_template('example.html', form=form) \r\n\r\nif __name__ == '__main__':\r\n if environ.has_key('FLASK_ENV') and environ['FLASK_ENV'] == 'production':\r\n from gevent.wsgi import WSGIServer\r\n WSGIServer((app.config[\"HOST\"], app.config[\"PORT\"]), app).serve_forever()\r\n else:\r\n app.run(host=app.config[\"HOST\"], port=app.config[\"PORT\"], debug=app.config[\"DEBUG\"])\r\n","subject":"Use a production ready server; no disk polling","message":"Use a production ready server; no disk polling\n","lang":"Python","license":"mit","repos":"Mouleshwar\/Flask-S3-Uploader,themouli\/Flask-S3-Uploader,themouli\/Flask-S3-Uploader,Mouleshwar\/Flask-S3-Uploader"} {"commit":"0dc84650b2929d31c054882ad67570fda6f1ffb9","old_file":"incuna_test_utils\/testcases\/urls.py","new_file":"incuna_test_utils\/testcases\/urls.py","old_contents":"from django.core.urlresolvers import resolve, reverse\nfrom django.test import TestCase\n\n\nclass URLsTestCase(TestCase):\n \"\"\"A TestCase with a check_url helper method for testing urls\"\"\"\n\n def check_url(self, view_class, url, url_name, url_args=None, url_kwargs=None):\n \"\"\"\n Assert a view's url is correctly configured\n\n Check the url_name reverses to give a correctly formated url.\n Check the url resolves to the correct view.\n \"\"\"\n\n reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs)\n self.assertEqual(reversed_url, url)\n\n resolved_view_class = resolve(url).func.cls\n self.assertEqual(resolved_view_class, view_class)\n","new_contents":"from django.core.urlresolvers import resolve, reverse\nfrom django.test import TestCase\n\n\nclass URLsMixin(object):\n \"\"\"A TestCase Mixin with a check_url helper method for testing urls\"\"\"\n\n def check_url(self, view_class, expected_url, url_name,\n url_args=None, url_kwargs=None):\n \"\"\"\n Assert a view's url is correctly configured\n\n Check the url_name reverses to give a correctly formated expected_url.\n Check the expected_url resolves to the correct view.\n \"\"\"\n\n reversed_url = reverse(url_name, args=url_args, kwargs=url_kwargs)\n self.assertEqual(reversed_url, expected_url)\n\n resolved_view_class = resolve(expected_url).func.cls\n self.assertEqual(resolved_view_class, view_class)\n\n\nclass URLsTestCase(URLsMixin, TestCase):\n pass\n","subject":"Rename url -> expected_url; Add URLsMixin","message":"Rename url -> expected_url; Add URLsMixin\n","lang":"Python","license":"bsd-2-clause","repos":"incuna\/incuna-test-utils,incuna\/incuna-test-utils"} {"commit":"a0348f21ce7abb577b93913c6f1c805cc6ccc75f","old_file":"knowit2019\/13.py","new_file":"knowit2019\/13.py","old_contents":"import json\n\n\ndef navigate_maze_struct(strategy, f='input\/MAZE.txt'):\n rooms = json.load(open(f))\n\n for row in rooms:\n for room in row:\n room['visited'] = False\n\n queue = [(0, 0)]\n\n while queue:\n y, x = queue.pop()\n room = rooms[y][x]\n\n if room['visited']:\n continue\n\n room['visited'] = True\n\n if room['y'] == 499 and room['x'] == 499:\n return sum_visited(rooms)\n\n for d in strategy:\n if d == 'D' and room['y'] < 499 and not room['syd']:\n queue.append((y + 1, x), )\n elif d == 'U' and y > 0 and not room['nord']:\n queue.append((y - 1, x), )\n elif d == 'R' and x < 499 and not room['aust']:\n queue.append((y, x + 1), )\n elif d == 'L' and x > 0 and not room['vest']:\n queue.append((y, x - 1), )\n\n return None\n\n\ndef sum_visited(rooms):\n visited = 0\n\n for row in rooms:\n for room in row:\n visited += 1 if room['visited'] else 0\n\n return visited\n\n\nprint(str(navigate_maze_struct('DRLU')) + \" woop\")\nprint(navigate_maze_struct('RDLU'))","new_contents":"import json\n\n\ndef navigate_maze_struct(strategy, f='input\/MAZE.txt'):\n rooms = json.load(open(f))\n\n for row in rooms:\n for room in row:\n room['visited'] = False\n\n queue = [(0, 0)]\n\n while queue:\n y, x = queue.pop()\n room = rooms[y][x]\n\n if room['visited']:\n continue\n\n room['visited'] = True\n\n if room['y'] == 499 and room['x'] == 499:\n return sum_visited(rooms)\n\n for d in strategy:\n if d == 'D' and room['y'] < 499 and not room['syd']:\n queue.append((y + 1, x), )\n elif d == 'U' and y > 0 and not room['nord']:\n queue.append((y - 1, x), )\n elif d == 'R' and x < 499 and not room['aust']:\n queue.append((y, x + 1), )\n elif d == 'L' and x > 0 and not room['vest']:\n queue.append((y, x - 1), )\n\n return None\n\n\ndef sum_visited(rooms):\n visited = 0\n\n for row in rooms:\n for room in row:\n visited += 1 if room['visited'] else 0\n\n return visited\n\n\nprint(abs(navigate_maze_struct('ULRD') - navigate_maze_struct('ULDR')))\n","subject":"Update strategy (was wrong way around before)","message":"Update strategy (was wrong way around before)\n","lang":"Python","license":"mit","repos":"matslindh\/codingchallenges,matslindh\/codingchallenges"} {"commit":"a6a95afca2964756a7777ea43839da1709187a27","old_file":"planetstack\/openstack_observer\/backend.py","new_file":"planetstack\/openstack_observer\/backend.py","old_contents":"import threading\nimport time\nfrom observer.event_loop import PlanetStackObserver\nfrom observer.event_manager import EventListener\nfrom util.logger import Logger, logging\n\nlogger = Logger(level=logging.INFO)\n\nclass Backend:\n \n def run(self):\n try:\n # start the openstack observer\n observer = PlanetStackObserver()\n observer_thread = threading.Thread(target=observer.run)\n observer_thread.start()\n \n # start event listene\n event_manager = EventListener(wake_up=observer.wake_up)\n event_manager_thread = threading.Thread(target=event_manager.run)\n event_manager_thread.start()\n except:\n logger.log_exc(\"Exception in child thread\")\n\n","new_contents":"import threading\nimport time\nfrom observer.event_loop import PlanetStackObserver\nfrom observer.event_manager import EventListener\nfrom util.logger import Logger, logging\n\nlogger = Logger(level=logging.INFO)\n\nclass Backend:\n \n def run(self):\n # start the openstack observer\n observer = PlanetStackObserver()\n observer_thread = threading.Thread(target=observer.run)\n observer_thread.start()\n \n # start event listene\n event_manager = EventListener(wake_up=observer.wake_up)\n event_manager_thread = threading.Thread(target=event_manager.run)\n event_manager_thread.start()\n\n","subject":"Drop try\/catch that causes uncaught errors in the Observer to be silently ignored","message":"Drop try\/catch that causes uncaught errors in the Observer to be silently ignored\n","lang":"Python","license":"apache-2.0","repos":"opencord\/xos,opencord\/xos,zdw\/xos,open-cloud\/xos,cboling\/xos,zdw\/xos,cboling\/xos,opencord\/xos,open-cloud\/xos,zdw\/xos,cboling\/xos,cboling\/xos,cboling\/xos,open-cloud\/xos,zdw\/xos"} {"commit":"131a6d6a60b975b45cd551c1b52c059c857cf1e5","old_file":"user\/views.py","new_file":"user\/views.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import render_to_response\nfrom django.template import RequestContext\nfrom django.contrib.auth.models import User\nfrom cronos.announcements.models import Id\n\t\nif request.user.email[-21:] == 'notapplicablemail.com':\n\tmail = 'unset'\nelif request.user.get_profile().webmail_username:\n\tmail = webmail_username + '@teilar.gr'\nelse:\n\t''\nfor item in Id.objects.filter(urlid__exact = request.user.get_profile().school):\n\tschool = str(item.name)\n\n@login_required\ndef user(request):\n\treturn render_to_response('user.html', {\n\t\t\t'mail': mail,\n\t\t}, context_instance = RequestContext(request))\n\n@login_required\ndef user_settings(request):\n\treturn render_to_response('settings.html', {\n\t\t\t'school': school,\n\t\t\t'mail': mail,\n\t\t}, context_instance = RequestContext(request))\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom django.contrib.auth.decorators import login_required\nfrom django.shortcuts import render_to_response\nfrom django.template import RequestContext\nfrom django.contrib.auth.models import User\nfrom cronos.announcements.models import Id\n\ndef getmail(request):\n\tif request.user.email[-21:] == 'notapplicablemail.com':\n\t\tmail = 'unset'\n\telif request.user.get_profile().webmail_username:\n\t\tmail = webmail_username + '@teilar.gr'\n\telse:\n\t\t''\n\treturn mail\n\ndef getschool(request):\n\tfor item in Id.objects.filter(urlid__exact = request.user.get_profile().school):\n\t\tschool = str(item.name)\n\treturn school\n\n@login_required\ndef user(request):\n\treturn render_to_response('user.html', {\n\t\t\t'mail': getmail(request),\n\t\t}, context_instance = RequestContext(request))\n\n@login_required\ndef user_settings(request):\n\treturn render_to_response('settings.html', {\n\t\t\t'school': getschool(request),\n\t\t\t'mail': getmail(request),\n\t\t}, context_instance = RequestContext(request))\n","subject":"Create functions getmail and getschool so they can be used in both sites user and settings","message":"Create functions getmail and getschool so they can be used in both sites user and settings\n\n","lang":"Python","license":"agpl-3.0","repos":"LinuxTeam-teilar\/cronos.teilar.gr,LinuxTeam-teilar\/cronos.teilar.gr,LinuxTeam-teilar\/cronos.teilar.gr"} {"commit":"283ba7e4a08aeac07b030700b58e672f3f54ed12","old_file":"utils\/migrate.py","new_file":"utils\/migrate.py","old_contents":"import settings\nimport os\n\nimport yoyo\nimport yoyo.connections\n\ndef path():\n return os.path.join(os.path.dirname(__file__), 'migrations')\n\nif __name__ == '__main__':\n conn, paramstyle = yoyo.connections.connect(settings.DATABASE_PATH)\n\n migrations = yoyo.read_migrations(conn, paramstyle, path())\n migrations.to_apply().apply()\n\n conn.commit()\n","new_contents":"import os\n\nimport psycopg2\nimport yoyo\nimport yoyo.connections\n\nimport settings\n\ndef path():\n return os.path.join(os.path.dirname(__file__), '..', 'migrations')\n\ndef run_migrations(dbconn=None, names=[]):\n if dbconn is None:\n dbconn, paramstyle = yoyo.connections.connect(settings.DATABASE_PATH)\n else:\n paramstyle = psycopg2.paramstyle\n\n migrations = yoyo.read_migrations(dbconn, paramstyle, path(), names=names)\n migrations.to_apply().apply()\n\n dbconn.commit()\n\nif __name__ == '__main__':\n run_migrations()\n","subject":"Make migrations runnable from external modules","message":"Make migrations runnable from external modules\n\nThese will be used in tests to setup the database.\n","lang":"Python","license":"mit","repos":"Storj\/accounts"} {"commit":"52c0b5d678e062384b6d4682b85b632bdc0ab093","old_file":"ktbs_bench\/utils\/decorators.py","new_file":"ktbs_bench\/utils\/decorators.py","old_contents":"from functools import wraps\nfrom inspect import getcallargs\n\nfrom timer import Timer\n\n\ndef bench(f):\n \"\"\"Times a function given specific arguments.\"\"\"\n\n timer = Timer(tick_now=False)\n\n @wraps(f)\n def wrapped(*args, **kwargs):\n timer.start()\n f(*args, **kwargs)\n timer.stop()\n\n res = {call_signature(f, *args, **kwargs): timer.get_times()['real']} # TODO penser a quel temps garder\n return res\n\n return wrapped\n\n\ndef call_signature(f, *args, **kwargs):\n \"\"\"Return a string representation of a function call.\"\"\"\n call_args = getcallargs(f, *args, **kwargs)\n return ';'.join([\"%s=%s\" % (k, v) for k, v in call_args.items()])\n\n\n@bench\ndef lala(a, b, c=\"default c\", d=\"default d\"):\n print(\"lala est appelee\")\n\n\nif __name__ == '__main__':\n print(lala(\"cest a\", \"cest b\", d=\"change d\"))\n","new_contents":"from functools import wraps\nfrom inspect import getcallargs\n\nfrom timer import Timer\n\n\ndef bench(f):\n \"\"\"Times a function given specific arguments.\"\"\"\n\n timer = Timer(tick_now=False)\n\n @wraps(f)\n def wrapped(*args, **kwargs):\n timer.start()\n f(*args, **kwargs)\n timer.stop()\n\n res = [call_signature(f, *args, **kwargs),\n timer.get_times()['real']] # TODO penser a quel temps garder\n return res\n\n return wrapped\n\n\ndef call_signature(f, *args, **kwargs):\n \"\"\"Return a string representation of a function call.\"\"\"\n call_args = getcallargs(f, *args, **kwargs)\n return ';'.join([\"%s=%s\" % (k, v) for k, v in call_args.items()])\n\n\n@bench\ndef lala(a, b, c=\"default c\", d=\"default d\"):\n print(\"lala est appelee\")\n\n\nif __name__ == '__main__':\n print(lala(\"cest a\", \"cest b\", d=\"change d\"))\n","subject":"Change @bench to return a list, because there will never be more than 1 key in the dict","message":"Change @bench to return a list, because there will never be more than 1 key in the dict\n","lang":"Python","license":"mit","repos":"ktbs\/ktbs-bench,ktbs\/ktbs-bench"} {"commit":"6520fde5be81eb3d1a91662edeef8bd2a1f6389c","old_file":"stonemason\/service\/tileserver\/helper.py","new_file":"stonemason\/service\/tileserver\/helper.py","old_contents":"# -*- encoding: utf-8 -*-\n\n__author__ = 'ray'\n__date__ = '4\/4\/15'\n\nfrom stonemason.mason import Portrayal\nfrom stonemason.mason.theme import Theme\n\n\ndef jsonify_portrayal(portrayal):\n assert isinstance(portrayal, Portrayal)\n template = {\n 'name': portrayal.name,\n 'metadata': {\n 'version': portrayal.metadata.version,\n 'abstract': portrayal.metadata.abstract,\n 'attribution': portrayal.metadata.attribution,\n 'center': portrayal.metadata.center,\n 'center_zoom': portrayal.metadata.center_zoom\n },\n 'maptype': portrayal.bundle.map_type,\n 'tileformat': portrayal.bundle.tile_format,\n 'pyramid': portrayal.pyramid,\n 'schemas': []\n }\n\n for tag in portrayal.iter_schema():\n template['schemas'].append(tag)\n\n return template\n\n\ndef jsonify_map_theme(map_theme):\n assert isinstance(map_theme, Theme)\n return repr(map_theme)\n","new_contents":"# -*- encoding: utf-8 -*-\n\n__author__ = 'ray'\n__date__ = '4\/4\/15'\n\nfrom stonemason.mason import Portrayal\nfrom stonemason.mason.theme import Theme\n\n\ndef jsonify_portrayal(portrayal):\n assert isinstance(portrayal, Portrayal)\n template = {\n 'name': portrayal.name,\n 'metadata': {\n 'title': portrayal.metadata.title,\n 'version': portrayal.metadata.version,\n 'abstract': portrayal.metadata.abstract,\n 'attribution': portrayal.metadata.attribution,\n 'center': portrayal.metadata.center,\n 'center_zoom': portrayal.metadata.center_zoom\n },\n 'maptype': portrayal.bundle.map_type,\n 'tileformat': portrayal.bundle.tile_format,\n 'pyramid': portrayal.pyramid,\n 'schemas': []\n }\n\n for tag in portrayal.iter_schema():\n template['schemas'].append(tag)\n\n return template\n\n\ndef jsonify_map_theme(map_theme):\n assert isinstance(map_theme, Theme)\n return repr(map_theme)\n","subject":"Add metadata title in portrayal view","message":"FEATURE: Add metadata title in portrayal view\n","lang":"Python","license":"mit","repos":"Kotaimen\/stonemason,Kotaimen\/stonemason"} {"commit":"3785b2804c88215114e0bb21f1aab6dc0554b30c","old_file":"django_react_templatetags\/ssr\/hypernova.py","new_file":"django_react_templatetags\/ssr\/hypernova.py","old_contents":"import logging\nimport json\n\nfrom django.conf import settings\nimport hypernova\nfrom hypernova.plugins.dev_mode import DevModePlugin\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass HypernovaService():\n def load_or_empty(self, component, headers={}, ssr_context=None):\n renderer = hypernova.Renderer(\n settings.REACT_RENDER_HOST,\n [DevModePlugin(logger)] if settings.DEBUG else [],\n timeout=get_request_timeout(),\n headers=headers,\n )\n\n inner_html = \"\"\n try:\n inner_html = renderer.render({component['name']: component['json']})\n except Exception as e:\n msg = \"SSR request to '{}' failed: {}\".format(\n settings.REACT_RENDER_HOST,\n e.__class__.__name__\n )\n logger.exception(msg)\n\n return inner_html\n\n\ndef get_request_timeout():\n if not hasattr(settings, 'REACT_RENDER_TIMEOUT'):\n return 20\n\n return settings.REACT_RENDER_TIMEOUT\n","new_contents":"import logging\nimport json\n\nfrom django.conf import settings\nimport hypernova\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass HypernovaService():\n def load_or_empty(self, component, headers={}, ssr_context=None):\n # from hypernova.plugins.dev_mode import DevModePlugin\n\n renderer = hypernova.Renderer(\n settings.REACT_RENDER_HOST,\n # [DevModePlugin(logger)] if settings.DEBUG else [],\n [],\n timeout=get_request_timeout(),\n headers=headers,\n )\n\n inner_html = \"\"\n try:\n inner_html = renderer.render({component['name']: component['json']})\n except Exception as e:\n msg = \"SSR request to '{}' failed: {}\".format(\n settings.REACT_RENDER_HOST,\n e.__class__.__name__\n )\n logger.exception(msg)\n\n return inner_html\n\n\ndef get_request_timeout():\n if not hasattr(settings, 'REACT_RENDER_TIMEOUT'):\n return 20\n\n return settings.REACT_RENDER_TIMEOUT\n","subject":"Disable DevModePlugin until py3 fix is fixed upstream","message":"Disable DevModePlugin until py3 fix is fixed upstream\n","lang":"Python","license":"mit","repos":"Frojd\/django-react-templatetags,Frojd\/django-react-templatetags,Frojd\/django-react-templatetags"} {"commit":"d0c775dd7f7964db608dd56d1899aa4e3697cd1e","old_file":"life\/__main__.py","new_file":"life\/__main__.py","old_contents":"import pyglet\n\nfrom life import WIDTH, HEIGHT, CELL_SIZE, DISPLAY_FPS, FULLSCREEN\nfrom life.creator import Creator\nfrom life.view import Field\n\n\ncreator = Creator(width=WIDTH, height=HEIGHT)\n\nif FULLSCREEN:\n window = pyglet.window.Window(fullscreen=True)\n cell_size = min(window.width \/\/ WIDTH, window.height \/\/ HEIGHT)\n field = Field(\n field_creator=creator,\n cell_size=cell_size,\n dx=(window.width - WIDTH * cell_size) \/\/ 2,\n dy=(window.height - HEIGHT * cell_size) \/\/ 2)\nelse:\n field = Field(field_creator=creator, cell_size=CELL_SIZE)\n window = pyglet.window.Window(width=field.width, height=field.height)\n\nif DISPLAY_FPS:\n fps_display = pyglet.clock.ClockDisplay()\nelse:\n fps_display = None\n\n\n@window.event\ndef on_draw():\n window.clear()\n field.draw()\n if fps_display:\n fps_display.draw()\n\n\ncreator.start()\n\npyglet.app.run()\n","new_contents":"import pyglet\n\nfrom life import WIDTH, HEIGHT, CELL_SIZE, DISPLAY_FPS, FULLSCREEN\nfrom life.creator import Creator\nfrom life.view import Field\n\n\ncreator = Creator(width=WIDTH, height=HEIGHT)\n\nif FULLSCREEN:\n window = pyglet.window.Window(fullscreen=True)\n cell_size = min(window.width \/\/ WIDTH, window.height \/\/ HEIGHT)\n field = Field(\n field_creator=creator,\n cell_size=cell_size,\n dx=(window.width - WIDTH * cell_size) \/\/ 2,\n dy=(window.height - HEIGHT * cell_size) \/\/ 2)\nelse:\n field = Field(field_creator=creator, cell_size=CELL_SIZE)\n window = pyglet.window.Window(width=field.width, height=field.height)\n\nif DISPLAY_FPS:\n fps_display = pyglet.window.FPSDisplay(window)\n fps_display.update_period = 1.\n\nelse:\n fps_display = None\n\n\n@window.event\ndef on_draw():\n window.clear()\n field.draw()\n if fps_display:\n fps_display.draw()\n\n\ncreator.start()\n\npyglet.app.run()\n","subject":"Use correct FPS display implementation.","message":"Use correct FPS display implementation.\n","lang":"Python","license":"bsd-2-clause","repos":"lig\/life"} {"commit":"69d22e9e7ff574d4f510269e589dafa45132047f","old_file":"stdnum\/br\/__init__.py","new_file":"stdnum\/br\/__init__.py","old_contents":"# __init__.py - collection of Brazilian numbers\n# coding: utf-8\n#\n# Copyright (C) 2012 Arthur de Jong\n#\n# This library is free software; you can redistribute it and\/or\n# modify it under the terms of the GNU Lesser General Public\n# License as published by the Free Software Foundation; either\n# version 2.1 of the License, or (at your option) any later version.\n#\n# This library is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# Lesser General Public License for more details.\n#\n# You should have received a copy of the GNU Lesser General Public\n# License along with this library; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA\n# 02110-1301 USA\n\n\"\"\"Collection of Brazilian numbers.\"\"\"\n","new_contents":"# __init__.py - collection of Brazilian numbers\n# coding: utf-8\n#\n# Copyright (C) 2012 Arthur de Jong\n#\n# This library is free software; you can redistribute it and\/or\n# modify it under the terms of the GNU Lesser General Public\n# License as published by the Free Software Foundation; either\n# version 2.1 of the License, or (at your option) any later version.\n#\n# This library is distributed in the hope that it will be useful,\n# but WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# Lesser General Public License for more details.\n#\n# You should have received a copy of the GNU Lesser General Public\n# License along with this library; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA\n# 02110-1301 USA\n\n\"\"\"Collection of Brazilian numbers.\"\"\"\nfrom stdnum.br import cnpj as vat # noqa: F401\n","subject":"Add missing vat alias for Brazil","message":"Add missing vat alias for Brazil\n","lang":"Python","license":"lgpl-2.1","repos":"arthurdejong\/python-stdnum,arthurdejong\/python-stdnum,arthurdejong\/python-stdnum"} {"commit":"c22894e1ac7071e19515321df8eaa639045c9ae5","old_file":"dlux\/api.py","new_file":"dlux\/api.py","old_contents":"# Copyright 2014 Hewlett-Packard Development Company, L.P.\n#\n# Author: Endre Karlson \n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nimport requests\n\nfrom django.conf import settings\nfrom odlclient.v2 import client as odl_client\n\n\ndef get_client(request):\n session = requests.Session()\n session.cookies.update({\n 'JSESSIONID': request.user.jsessionid,\n 'JSESSIONIDSSO': request.user.jsessionidsso\n })\n url = request.user.controller + '\/controller\/nb\/v2'\n http = odl_client.HTTPClient(url, http=session, debug=settings.DEBUG)\n client = odl_client.Client(http)\n return client\n","new_contents":"# Copyright 2014 Hewlett-Packard Development Company, L.P.\n#\n# Author: Endre Karlson \n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nimport requests\n\nfrom django.conf import settings\nfrom odlclient.v2 import client as odl_client\n\n\ndef get_client(request):\n session = requests.Session()\n session.cookies.update({\n 'JSESSIONID': request.user.jsessionid,\n 'JSESSIONIDSSO': request.user.jsessionidsso\n })\n http = odl_client.HTTPClient(request.user.controller, http=session,\n debug=settings.DEBUG)\n client = odl_client.Client(http)\n return client\n","subject":"Update according to latest odlclient","message":"Update according to latest odlclient\n","lang":"Python","license":"apache-2.0","repos":"ekarlso\/dlux-horizon,ekarlso\/dlux-horizon,ekarlso\/dlux-horizon"} {"commit":"ba499556cf3a1f09c55ba2631c1dbb988e95fb82","old_file":"web\/test\/test_web.py","new_file":"web\/test\/test_web.py","old_contents":"def test_web(app):\n client = app.test_client()\n response = client.post(\n '\/user\/sign-in?next=\/',\n follow_redirects=True,\n data=dict(\n username='test@test.com',\n password='Password1'\n )\n )\n response = client.get('\/')\n assert response.status_code == 200\n assert b\"Training stats<\/title>\" in response.data\n","new_contents":"def login(client):\n response = client.post(\n '\/user\/sign-in?next=\/',\n follow_redirects=True,\n data=dict(\n username='test@test.com',\n password='Password1'\n )\n )\n return response\n\ndef test_navigating_to_startpage(app):\n client = app.test_client()\n login(client)\n response = client.get('\/')\n assert response.status_code == 200\n assert b\"<title>Training stats<\/title>\" in response.data\n","subject":"Refactor test to be more clear","message":"Refactor test to be more clear\n","lang":"Python","license":"apache-2.0","repos":"vinntreus\/training_stats,vinntreus\/training_stats"} {"commit":"d8168185aa0153fac55e3c59761a5e561a5b0137","old_file":"src\/ocspdash\/__init__.py","new_file":"src\/ocspdash\/__init__.py","old_contents":"\"\"\"A dashboard for the status of the top certificate authorities' OCSP responders.\"\"\"\n# metadata\n__version__ = '0.1.0-dev'\n\n__title__ = 'OCSPdash'\n# keep the __description__ synchronized with the package docstring\n__description__ = \"A dashboard for the status of the top certificate authorities' OCSP responders.\"\n__url__ = 'https:\/\/github.com\/scolby33\/OCSPdash'\n\n__author__ = 'Scott Colby'\n__email__ = 'scolby33@gmail.com'\n\n__license__ = 'MIT'\n__copyright__ = 'Copyright (c) 2017 Scott Colby'\n\nfrom .util import install_custom_json_encoder\ninstall_custom_json_encoder()\n","new_contents":"\"\"\"A dashboard for the status of the top certificate authorities' OCSP responders.\"\"\"\n# metadata\n__version__ = '0.1.0-dev'\n\n__title__ = 'OCSPdash'\n# keep the __description__ synchronized with the package docstring\n__description__ = \"A dashboard for the status of the top certificate authorities' OCSP responders.\"\n__url__ = 'https:\/\/github.com\/scolby33\/OCSPdash'\n\n__author__ = 'Scott Colby'\n__email__ = 'scolby33@gmail.com'\n\n__license__ = 'MIT'\n__copyright__ = 'Copyright (c) 2017 Scott Colby and Charles Tapley Hoyt'\n\nfrom .util import install_custom_json_encoder\ninstall_custom_json_encoder()\n","subject":"Add @cthoyt to package __copyright__","message":"Add @cthoyt to package __copyright__\n","lang":"Python","license":"mit","repos":"scolby33\/OCSPdash,scolby33\/OCSPdash,scolby33\/OCSPdash"} {"commit":"9bc3c6ef95bd3b229b019dcd0d39aa9940528cf4","old_file":"config\/development.py","new_file":"config\/development.py","old_contents":"\n# Secret passphrase\n# FIXME: Temporarily SECRET must have the same value as SECRET_KEYS\n# due to the current spring boot implementation\nSECRET = '\/etc\/keyczar\/keys'\n# Secret keyczar keys\nSECRET_KEYS = '\/etc\/keyczar\/keys'\n# Fill as needed\nDEBUG = True\nIGNORE_AUTH = True\nMESOS_FRAMEWORK_ENDPOINT = 'http:\/\/127.0.0.1:5004\/bigdata\/mesos_framework\/v1\/clusters'\nORQUESTRATOR_ENDPOINT = 'http:\/\/127.0.0.1:5005\/orquestrator\/v1\/clusters'\n#MESOS_FRAMEWORK_ENDPOINT = 'http:\/\/mesos_framework:5000\/bigdata\/mesos_framework\/v1\/clusters'\n#ORQUESTRATOR_ENDPOINT = 'http:\/\/orquestrator:5005\/orquestrator\/v1\/clusters'\nCONSUL_ENDPOINT = 'http:\/\/consul:8500\/v1\/kv'\n","new_contents":"\n# Secret passphrase\n# FIXME: Temporarily SECRET must have the same value as SECRET_KEYS\n# due to the current spring boot implementation\nSECRET = '\/etc\/keyczar\/keys'\n# Secret keyczar keys\nSECRET_KEYS = '\/etc\/keyczar\/keys'\n# Fill as needed\nDEBUG = True\nIGNORE_AUTH = True\n#MESOS_FRAMEWORK_ENDPOINT = 'http:\/\/127.0.0.1:5004\/bigdata\/mesos_framework\/v1\/clusters'\n#ORQUESTRATOR_ENDPOINT = 'http:\/\/127.0.0.1:5005\/orquestrator\/v1\/clusters'\nMESOS_FRAMEWORK_ENDPOINT = 'http:\/\/mesos_framework:5000\/bigdata\/mesos_framework\/v1\/clusters'\nORQUESTRATOR_ENDPOINT = 'http:\/\/orquestrator:5005\/orquestrator\/v1\/clusters'\nCONSUL_ENDPOINT = 'http:\/\/consul:8500\/v1\/kv'\n","subject":"Fix de mis endpoints que uso para desarrollo, disculpas T_T","message":"Fix de mis endpoints que uso para desarrollo, disculpas T_T\n","lang":"Python","license":"apache-2.0","repos":"bigdatacesga\/paas-service,bigdatacesga\/paas-service,bigdatacesga\/paas-service"} {"commit":"00d835c3b4512b407033af280600d9428a155b22","old_file":"noah\/noah.py","new_file":"noah\/noah.py","old_contents":"import json\nimport random\nimport pprint\n\nclass Noah(object):\n def __init__(self, dictionary_file):\n self.dictionary = json.load(dictionary_file)\n\n def list(self):\n return '\\n'.join([entry['word'] for entry in self.dictionary])\n\n def define(self, word):\n return self.output(filter(lambda x: x['word'] == word, self.dictionary))\n\n if not entry is None:\n return self.output(entry)\n\n def random(self):\n return self.output(random.choice(self.dictionary))\n\n def output(self, data):\n return json.dumps(data, indent=4)\n\ndef main():\n with open('..\/dictionaries\/english.json') as dictionary:\n n = Noah(dictionary)\n\n print n.list()\n print n.define('run')\n print n.random()\n\nif __name__ == '__main__':\n main()","new_contents":"import json\nimport random\nimport pprint\n\nclass Noah(object):\n def __init__(self, dictionary_file):\n self.dictionary = json.load(dictionary_file)\n\n def list(self):\n return '\\n'.join([entry['word'] for entry in self.dictionary])\n\n def define(self, word):\n return self.output(filter(lambda x: x['word'] == word, self.dictionary))\n\n def random(self):\n return self.output(random.choice(self.dictionary))\n\n def output(self, data):\n return json.dumps(data, indent=4)\n\ndef main():\n with open('..\/dictionaries\/english.json') as dictionary:\n n = Noah(dictionary)\n\n print n.list()\n print n.define('run')\n print n.random()\n\nif __name__ == '__main__':\n main()","subject":"Remove unneeded block in define.","message":"Remove unneeded block in define.\n","lang":"Python","license":"mit","repos":"maxdeviant\/noah"} {"commit":"4bcf7f83351bc64ed47c5531cb66ccb20f762dd0","old_file":"pyMKL\/__init__.py","new_file":"pyMKL\/__init__.py","old_contents":"from __future__ import absolute_import\nfrom __future__ import unicode_literals\nfrom __future__ import print_function\nfrom __future__ import division\nfrom future import standard_library\nstandard_library.install_aliases()\n\nimport numpy as np\nimport scipy.sparse as sp\nfrom ctypes import CDLL, cdll, RTLD_GLOBAL\nfrom ctypes import POINTER, byref, c_int, c_longlong\n\npath = 'libmkl_intel_lp64.dylib'\nMKLlib = CDLL(path, RTLD_GLOBAL)\n\nfrom .pardisoInterface import pardisoinit, pardiso\nfrom .pardisoSolver import pardisoSolver\n","new_contents":"from __future__ import absolute_import\nfrom __future__ import unicode_literals\nfrom __future__ import print_function\nfrom __future__ import division\nfrom future import standard_library\nstandard_library.install_aliases()\n\nimport platform\nimport numpy as np\nimport scipy.sparse as sp\nfrom ctypes import CDLL, cdll, RTLD_GLOBAL\nfrom ctypes import POINTER, byref, c_int, c_longlong\n\nsystype = platform.system()\nif systype == 'Darwin':\n path = 'libmkl_intel_lp64.dylib'\nelif systype == 'Linux':\n path = 'libmkl_intel_lp64.so'\nelif systype == 'Windows':\n path = 'mkl_intel_lp64_dll.lib'\nelse:\n raise ImportError('Cannot determine MKL library to import')\nMKLlib = CDLL(path, RTLD_GLOBAL)\n\nfrom .pardisoInterface import pardisoinit, pardiso\nfrom .pardisoSolver import pardisoSolver\n","subject":"Add experimental support for other OS versions.","message":"Add experimental support for other OS versions.\n","lang":"Python","license":"mit","repos":"dwfmarchant\/pyMKL"} {"commit":"6f1dc606b4c4f2702e0a5b48338488ac2eec197c","old_file":"scripts\/utils.py","new_file":"scripts\/utils.py","old_contents":"#!\/usr\/bin\/env python3\n\n# Touhou Community Reliant Automatic Patcher\n# Scripts\n#\n# ----\n#\n\"\"\"Utility functions shared among all the scripts.\"\"\"\n\nfrom collections import OrderedDict\nimport json\nimport os\n\njson_load_params = {\n 'object_pairs_hook': OrderedDict\n}\n\ndef patch_files_filter(files):\n \"\"\"Filters all file names that can not be among the content of a patch.\"\"\"\n for i in files:\n if i != 'files.js':\n yield i\n\n\njson_dump_params = {\n 'ensure_ascii': False,\n 'indent': '\\t',\n 'separators': (',', ': '),\n 'sort_keys': True\n}\n\n# Default parameters for JSON input and output\ndef json_load(fn, json_kwargs=json_load_params):\n with open(fn, 'r', encoding='utf-8') as file:\n return json.load(file, **json_kwargs)\n\n\ndef json_store(fn, obj, dirs=[''], json_kwargs=json_dump_params):\n \"\"\"Saves the JSON object [obj] to [fn], creating all necessary\n directories in the process. If [dirs] is given, the function is\n executed for every root directory in the array.\"\"\"\n for i in dirs:\n full_fn = os.path.join(i, fn)\n os.makedirs(os.path.dirname(full_fn), exist_ok=True)\n with open(full_fn, 'w', encoding='utf-8') as file:\n json.dump(obj, file, **json_kwargs)\n file.write('\\n')\n","new_contents":"#!\/usr\/bin\/env python3\n\n# Touhou Community Reliant Automatic Patcher\n# Scripts\n#\n# ----\n#\n\"\"\"Utility functions shared among all the scripts.\"\"\"\n\nfrom collections import OrderedDict\nimport json\nimport os\n\njson_load_params = {\n 'object_pairs_hook': OrderedDict\n}\n\ndef patch_files_filter(files):\n \"\"\"Filters all file names that can not be among the content of a patch.\"\"\"\n for i in files:\n if i != 'files.js':\n yield i\n\n\njson_dump_params = {\n 'ensure_ascii': False,\n 'indent': '\\t',\n 'separators': (',', ': '),\n 'sort_keys': True\n}\n\n# Default parameters for JSON input and output\ndef json_load(fn, json_kwargs=json_load_params):\n with open(fn, 'r', encoding='utf-8') as file:\n return json.load(file, **json_kwargs)\n\n\ndef json_store(fn, obj, dirs=[''], json_kwargs=json_dump_params):\n \"\"\"Saves the JSON object [obj] to [fn], creating all necessary\n directories in the process. If [dirs] is given, the function is\n executed for every root directory in the array.\"\"\"\n for i in dirs:\n full_fn = os.path.join(i, fn)\n os.makedirs(os.path.dirname(full_fn), exist_ok=True)\n with open(full_fn, 'w', newline='\\n', encoding='utf-8') as file:\n json.dump(obj, file, **json_kwargs)\n file.write('\\n')\n","subject":"Enforce Unix newlines when writing JSON files.","message":"scripts: Enforce Unix newlines when writing JSON files.\n","lang":"Python","license":"unlicense","repos":"VBChunguk\/thcrap,thpatch\/thcrap,thpatch\/thcrap,thpatch\/thcrap,VBChunguk\/thcrap,VBChunguk\/thcrap,thpatch\/thcrap,thpatch\/thcrap"} {"commit":"89929acbb2ee3c5617758966d8916139726d7b74","old_file":"app\/state.py","new_file":"app\/state.py","old_contents":"import multiprocessing\nimport unicornhathd as unicornhat\nimport importlib\nimport sys\nimport os\n\nimport app.programs.hd\n\n\nclass State:\n ''' Handles the Unicorn HAT state'''\n\n def __init__(self):\n self._process = None\n\n def start_program(self, name, params={}):\n try:\n program = getattr(app.programs.hd, name)\n except AttributeError:\n raise ProgramNotFound(name)\n\n self.stop_program()\n\n if params.get(\"brightness\"):\n unicornhat.brightness(float(params[\"brightness\"]))\n\n if params.get(\"rotation\"):\n unicornhat.rotation(int(params[\"rotation\"]))\n\n self._process = multiprocessing.Process(target=program.run, args=(params,))\n self._process.start()\n\n def stop_program(self):\n if self._process is not None:\n self._process.terminate()\n unicornhat.show()\n\n\nclass ProgramNotFound(Exception):\n pass","new_contents":"import multiprocessing\nimport unicornhathd as unicornhat\nimport importlib\nimport sys\nimport os\n\nimport app.programs.hd\n\n\nclass State:\n ''' Handles the Unicorn HAT state'''\n\n def __init__(self):\n self._process = None\n\n def start_program(self, name, params={}):\n try:\n program = getattr(app.programs.hd, name)\n except AttributeError:\n raise ProgramNotFound(name)\n\n self.stop_program()\n\n if params.get(\"brightness\") is not None:\n unicornhat.brightness(float(params[\"brightness\"]))\n\n if params.get(\"rotation\") is not None:\n unicornhat.rotation(int(params[\"rotation\"]))\n\n self._process = multiprocessing.Process(target=program.run, args=(params,))\n self._process.start()\n\n def stop_program(self):\n if self._process is not None:\n self._process.terminate()\n unicornhat.show()\n\n\nclass ProgramNotFound(Exception):\n pass","subject":"Fix setting rotation to 0","message":"Fix setting rotation to 0\n","lang":"Python","license":"mit","repos":"njbbaer\/unicorn-remote,njbbaer\/unicorn-remote,njbbaer\/unicorn-remote"} {"commit":"013ed651c3e8e7cfa4b8babefc2664644b928852","old_file":"pybtex\/bibtex\/exceptions.py","new_file":"pybtex\/bibtex\/exceptions.py","old_contents":"# Copyright (C) 2006, 2007, 2008 Andrey Golovizin\n#\n# This file is part of pybtex.\n#\n# pybtex is free software; you can redistribute it and\/or modify\n# under the terms of the GNU General Public License as published by the\n# Free Software Foundation; either version 2 of the License, or (at your\n# option) any later version.\n#\n# pybtex is distributed in the hope that it will be useful, but\n# WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with pybtex; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301\n# USA\n\nclass BibTeXError(Exception):\n pass\n","new_contents":"# Copyright (C) 2006, 2007, 2008 Andrey Golovizin\n#\n# This file is part of pybtex.\n#\n# pybtex is free software; you can redistribute it and\/or modify\n# under the terms of the GNU General Public License as published by the\n# Free Software Foundation; either version 2 of the License, or (at your\n# option) any later version.\n#\n# pybtex is distributed in the hope that it will be useful, but\n# WITHOUT ANY WARRANTY; without even the implied warranty of\n# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU\n# General Public License for more details.\n#\n# You should have received a copy of the GNU General Public License\n# along with pybtex; if not, write to the Free Software\n# Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301\n# USA\n\nfrom pybtex.exceptions import PybtexError\n\nclass BibTeXError(PybtexError):\n pass\n","subject":"Make BibTeXError a subclass of PybtexError.","message":"Make BibTeXError a subclass of PybtexError.\n","lang":"Python","license":"mit","repos":"andreas-h\/pybtex,chbrown\/pybtex,andreas-h\/pybtex,chbrown\/pybtex"} {"commit":"e75e6ec300e1127f7010d36ef63343e522318f90","old_file":"sunpy\/instr\/iris\/iris.py","new_file":"sunpy\/instr\/iris\/iris.py","old_contents":"\"\"\"\nSome very beta tools for IRIS\n\"\"\"\n\nimport sunpy.io\nimport sunpy.time\nimport sunpy.map\n\n__all__ = ['SJI_to_cube']\n\ndef SJI_to_cube(filename, start=0, stop=None):\n \"\"\"\n Read a SJI file and return a MapCube\n \n ..warning::\n This function is a very early beta and is not stable. Further work is \n on going to improve SunPy IRIS support.\n \n Parameters\n ----------\n filename: string\n File to read\n \n start:\n Temporal axis index to create MapCube from\n \n stop:\n Temporal index to stop MapCube at\n \n Returns\n -------\n \n iris_cube: sunpy.map.MapCube\n A map cube of the SJI sequence\n \"\"\"\n \n hdus = sunpy.io.read_file(filename)\n #Get the time delta\n time_range = sunpy.time.TimeRange(hdus[0][1]['STARTOBS'], hdus[0][1]['ENDOBS'])\n splits = time_range.split(hdus[0][0].shape[0])\n\n if not stop:\n stop = len(splits)\n\n headers = [hdus[0][1]]*(stop-start)\n datas = hdus[0][0][start:stop]\n \n #Make the cube:\n iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)\n #Set the date\/time\n for i,m in enumerate(iris_cube):\n m.meta['DATE-OBS'] = splits[i].center().isoformat()\n \n return iris_cube","new_contents":"\"\"\"\nSome very beta tools for IRIS\n\"\"\"\n\nimport sunpy.io\nimport sunpy.time\nimport sunpy.map\n\n__all__ = ['SJI_to_cube']\n\ndef SJI_to_cube(filename, start=0, stop=None, hdu=0):\n \"\"\"\n Read a SJI file and return a MapCube\n \n ..warning::\n This function is a very early beta and is not stable. Further work is \n on going to improve SunPy IRIS support.\n \n Parameters\n ----------\n filename: string\n File to read\n \n start:\n Temporal axis index to create MapCube from\n \n stop:\n Temporal index to stop MapCube at\n \n hdu:\n Choose hdu index\n\n Returns\n -------\n \n iris_cube: sunpy.map.MapCube\n A map cube of the SJI sequence\n \"\"\"\n \n hdus = sunpy.io.read_file(filename)\n #Get the time delta\n time_range = sunpy.time.TimeRange(hdus[hdu][1]['STARTOBS'], hdus[hdu][1]['ENDOBS'])\n splits = time_range.split(hdus[hdu][0].shape[0])\n\n if not stop:\n stop = len(splits)\n\n headers = [hdus[hdu][1]]*(stop-start)\n datas = hdus[hdu][0][start:stop]\n \n #Make the cube:\n iris_cube = sunpy.map.Map(zip(datas,headers),cube=True)\n #Set the date\/time\n for i,m in enumerate(iris_cube):\n m.meta['DATE-OBS'] = splits[i].center().isoformat()\n \n return iris_cube\n","subject":"Change hdu[0] to hdu for optional indexing","message":"Change hdu[0] to hdu for optional indexing\n","lang":"Python","license":"bsd-2-clause","repos":"Alex-Ian-Hamilton\/sunpy,dpshelio\/sunpy,dpshelio\/sunpy,dpshelio\/sunpy,Alex-Ian-Hamilton\/sunpy,Alex-Ian-Hamilton\/sunpy"} {"commit":"230bb0a09146cd0b696b528b3ad6dd9ccf057113","old_file":"tests\/test_checker.py","new_file":"tests\/test_checker.py","old_contents":"import pytest\nimport os, stat\n\nfrom botbot import checker, problems\n\ndef test_fastq_checker():\n assert checker.is_fastq(\"bad.fastq\") == problems.PROB_FILE_IS_FASTQ\n assert checker.is_fastq(\"good.py\") == problems.PROB_NO_PROBLEM\n assert checker.is_fastq(\"fastq.actually_ok_too\") == problems.PROB_NO_PROBLEM\n\ndef test_permission_checker(tmpdir):\n # Create a test file\n p = tmpdir.join(\"bad_permissions.txt\")\n p.write('')\n prev = tmpdir.chdir()\n\n # Change its permissions a bunch... maybe this is too expensive?\n for m in range(0o300, 0o700, 0o010):\n p.chmod(m)\n prob = checker.has_permission_issues(os.path.abspath(p.basename))\n if not bool(0o040 & m): # octal Unix permission for 'group readable'\n assert prob == problems.PROB_FILE_NOT_GRPRD\n else:\n assert prob == problems.PROB_NO_PROBLEM\n\n prev.chdir()\n","new_contents":"import pytest\nimport os, stat\n\nfrom botbot import checker, problems\n\ndef test_fastq_checker_path_names():\n assert checker.is_fastq(\"bad.fastq\") == problems.PROB_FILE_IS_FASTQ\n assert checker.is_fastq(\"good.py\") == problems.PROB_NO_PROBLEM\n assert checker.is_fastq(\"fastq.actually_ok_too\") == problems.PROB_NO_PROBLEM\n\ndef test_fastq_checker_symlinks(tmpdir):\n prev = tmpdir.chdir()\n\n # Make a test file\n p = tmpdir.join(\"bad.fastq\")\n p.write('')\n os.symlink(p.basename, \"good.fastq\")\n\n assert checker.is_fastq(\"bad.fastq\") == problems.PROB_FILE_IS_FASTQ\n assert checker.is_fastq(\"good.fastq\") == problems.PROB_NO_PROBLEM\n prev.chdir()\n\ndef test_permission_checker(tmpdir):\n # Create a test file\n p = tmpdir.join(\"bad_permissions.txt\")\n p.write('')\n prev = tmpdir.chdir()\n\n # Change its permissions a bunch... maybe this is too expensive?\n for m in range(0o300, 0o700, 0o010):\n p.chmod(m)\n prob = checker.has_permission_issues(os.path.abspath(p.basename))\n if not bool(0o040 & m): # octal Unix permission for 'group readable'\n assert prob == problems.PROB_FILE_NOT_GRPRD\n else:\n assert prob == problems.PROB_NO_PROBLEM\n\n prev.chdir()\n","subject":"Add test for symlink detection","message":"Add test for symlink detection\n","lang":"Python","license":"mit","repos":"jackstanek\/BotBot,jackstanek\/BotBot"} {"commit":"faed82947209b34ccb4063e8244a9da019fa52a2","old_file":"bills\/urls.py","new_file":"bills\/urls.py","old_contents":"from . import views\nfrom django.conf.urls import url\n\n\nurlpatterns = [\n url(r'^by_topic\/', views.bill_list_by_topic),\n url(r'^by_location', views.bill_list_by_location),\n url(r'^latest_activity\/', views.latest_bill_activity),\n url(r'^latest\/', views.latest_bill_actions),\n url(r'^detail\/(?P<bill_session>(.*))\/(?P<bill_identifier>(.*))\/$', views.bill_detail, name='bill_detail'),\n]\n","new_contents":"from . import views\nfrom django.conf.urls import url\n\n\nurlpatterns = [\n url(r'^by_topic\/', views.bill_list_by_topic),\n url(r'^by_location', views.bill_list_by_location),\n url(r'^current_session\/', views.bill_list_current_session),\n url(r'^latest_activity\/', views.bill_list_latest),\n url(r'^detail\/(?P<bill_session>(.*))\/(?P<bill_identifier>(.*))\/$', views.bill_detail, name='bill_detail'),\n]\n","subject":"Update bills added by current session","message":"Update bills added by current session\n","lang":"Python","license":"mit","repos":"jamesturk\/tot,jamesturk\/tot,jamesturk\/tot,jamesturk\/tot,jamesturk\/tot"} {"commit":"b631dadb54f90e4abb251f7680f883f2e3e0e914","old_file":"radar\/radar\/validation\/patient_numbers.py","new_file":"radar\/radar\/validation\/patient_numbers.py","old_contents":"from radar.groups import is_radar_group\nfrom radar.validation.core import Validation, pass_call, ValidationError, Field\nfrom radar.validation.sources import RadarSourceValidationMixin\nfrom radar.validation.meta import MetaValidationMixin\nfrom radar.validation.patients import PatientValidationMixin\nfrom radar.validation.validators import required, max_length, not_empty, normalise_whitespace\nfrom radar.validation.number_validators import NUMBER_VALIDATORS\n\n\nclass PatientNumberValidation(PatientValidationMixin, RadarSourceValidationMixin, MetaValidationMixin, Validation):\n number = Field([not_empty(), normalise_whitespace(), max_length(50)])\n number_group = Field([required()])\n\n def validate_number_group(self, number_group):\n if is_radar_group(number_group):\n raise ValidationError(\"Can't add RaDaR numbers.\")\n\n return number_group\n\n @pass_call\n def validate(self, call, obj):\n number_group = obj.number_group\n\n number_validators = NUMBER_VALIDATORS.get((number_group.type, number_group.code))\n\n if number_validators is not None:\n call.validators_for_field(number_validators, obj, self.number)\n\n return obj\n","new_contents":"from radar.groups import is_radar_group, get_radar_group\nfrom radar.validation.core import Validation, pass_call, ValidationError, Field\nfrom radar.validation.sources import RadarSourceValidationMixin\nfrom radar.validation.meta import MetaValidationMixin\nfrom radar.validation.patients import PatientValidationMixin\nfrom radar.validation.validators import required, max_length, not_empty, normalise_whitespace\nfrom radar.validation.number_validators import NUMBER_VALIDATORS\nfrom radar.models.patient_numbers import PatientNumber\nfrom radar.database import db\n\n\nclass PatientNumberValidation(PatientValidationMixin, RadarSourceValidationMixin, MetaValidationMixin, Validation):\n number = Field([not_empty(), normalise_whitespace(), max_length(50)])\n number_group = Field([required()])\n\n def validate_number_group(self, number_group):\n if is_radar_group(number_group):\n raise ValidationError(\"Can't add RaDaR numbers.\")\n\n return number_group\n\n @classmethod\n def is_duplicate(cls, obj):\n q = PatientNumber.query\n q = q.filter(PatientNumber.source_group == get_radar_group())\n q = q.filter(PatientNumber.number_group == obj.number_group)\n q = q.filter(PatientNumber.number == obj.number)\n\n if obj.id is not None:\n q = q.filter(PatientNumber.id != obj.id)\n\n q = q.exists()\n\n duplicate = db.session.query(q).scalar()\n\n return duplicate\n\n @pass_call\n def validate(self, call, obj):\n number_group = obj.number_group\n\n number_validators = NUMBER_VALIDATORS.get((number_group.type, number_group.code))\n\n if number_validators is not None:\n call.validators_for_field(number_validators, obj, self.number)\n\n if self.is_duplicate(obj):\n raise ValidationError({'number': 'A patient already exists with this number.'})\n\n return obj\n","subject":"Check for duplicate patient numbers","message":"Check for duplicate patient numbers\n\nFixes #286\n","lang":"Python","license":"agpl-3.0","repos":"renalreg\/radar,renalreg\/radar,renalreg\/radar,renalreg\/radar"} {"commit":"6a01e99585db3ea38a8d8325dd4f826e78fc0f1d","old_file":"test_project\/settings.py","new_file":"test_project\/settings.py","old_contents":"import os\nimport sys\n\nPROJECT_ROOT = os.path.abspath(os.path.dirname(__file__))\nsys.path.insert(0, os.path.join(PROJECT_ROOT, '..'))\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n }\n}\n\n\nINSTALLED_APPS = (\n 'djcelery_email',\n 'appconf',\n 'tester',\n)\n\nSECRET_KEY = 'unique snowflake'\n\n\nTEST_RUNNER = \"test_runner.DJCETestSuiteRunner\"\n\n# Not set here - see 'test_runner.py'\n# EMAIL_BACKEND = 'djcelery_email.backends.CeleryEmailBackend'\n\nCELERY_EMAIL_BACKEND = 'django.core.mail.backends.locmem.EmailBackend'\nCELERY_EMAIL_TASK_CONFIG = {\n 'queue' : 'django_email',\n 'delivery_mode' : 1, # non persistent\n 'rate_limit' : '50\/m', # 50 chunks per minute\n}\n","new_contents":"import os\nimport sys\n\nPROJECT_ROOT = os.path.abspath(os.path.dirname(__file__))\nsys.path.insert(0, os.path.join(PROJECT_ROOT, '..'))\n\nDATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n }\n}\n\n\nINSTALLED_APPS = (\n 'djcelery_email',\n 'appconf',\n 'tester',\n)\n\nSECRET_KEY = 'unique snowflake'\n\n# Django 1.7 throws dire warnings if this is not set.\n# We don't actually use any middleware, given that there are no views.\nMIDDLEWARE_CLASSES = ()\n\nTEST_RUNNER = \"test_runner.DJCETestSuiteRunner\"\n\n# Not set here - see 'test_runner.py'\n# EMAIL_BACKEND = 'djcelery_email.backends.CeleryEmailBackend'\n\nCELERY_EMAIL_BACKEND = 'django.core.mail.backends.locmem.EmailBackend'\nCELERY_EMAIL_TASK_CONFIG = {\n 'queue' : 'django_email',\n 'delivery_mode' : 1, # non persistent\n 'rate_limit' : '50\/m', # 50 chunks per minute\n}\n","subject":"Set MIDDLEWARE_CLASSES to stop Django 1.7 warnings.","message":"Set MIDDLEWARE_CLASSES to stop Django 1.7 warnings.\n","lang":"Python","license":"bsd-3-clause","repos":"pmclanahan\/django-celery-email,pmclanahan\/django-celery-email"} {"commit":"f59b249cf2b149f96833d9e1025a98819bf5f62a","old_file":"sharepa\/search.py","new_file":"sharepa\/search.py","old_contents":"import json\nimport requests\n\nfrom elasticsearch_dsl import Search\nfrom elasticsearch_dsl.result import Response\n\n\nclass ShareSearch(Search):\n BASE_URL = 'https:\/\/osf.io\/api\/v1\/share\/search\/'\n HEADERS = {'content-type': 'application\/json'}\n PARAMS = dict(raw=True)\n\n def execute(self):\n return Response(\n self._query(self.to_dict()),\n callbacks=self._doc_type_map\n )\n\n def count(self):\n d = self.to_dict()\n if d.get('aggs'):\n del d['aggs']\n self = ShareSearch.from_dict(d)\n return self._query(self.to_dict(), params=dict(count=True))['count']\n\n def scan(self, size=100):\n count = 0\n page = list(self[0:size].execute())\n while(page):\n for hit in page:\n count += 1\n yield hit\n page = list(self[count:count + size].execute())\n\n def _query(self, data, params=None):\n return requests.post(\n self.BASE_URL,\n headers=self.HEADERS,\n data=json.dumps(self.to_dict()),\n params=params or self.PARAMS\n ).json()\n\nbasic_search = ShareSearch()\nbasic_search.aggs.bucket(\n 'sourceAgg',\n 'terms',\n field='_type',\n size=0,\n min_doc_count=0\n)\n","new_contents":"import json\nimport requests\n\nfrom elasticsearch_dsl import Search\nfrom elasticsearch_dsl.result import Response\n\n\nclass ShareSearch(Search):\n BASE_URL = 'http:\/\/localhost:8000\/api\/search\/abstractcreativework\/_search'\n HEADERS = {'content-type': 'application\/json'}\n PARAMS = dict(raw=True)\n\n def execute(self):\n return Response(\n self._query(self.to_dict()),\n callbacks=self._doc_type_map\n )\n\n def count(self):\n d = self.to_dict()\n if d.get('aggs'):\n del d['aggs']\n self = ShareSearch.from_dict(d)\n return self._query(self.to_dict(), params=dict(size=0))['hits']['total']\n\n def scan(self, size=100):\n count = 0\n page = list(self[0:size].execute())\n while(page):\n for hit in page:\n count += 1\n yield hit\n page = list(self[count:count + size].execute())\n\n def _query(self, data, params=None):\n return requests.post(\n self.BASE_URL,\n headers=self.HEADERS,\n data=json.dumps(self.to_dict()),\n params=params or self.PARAMS\n ).json()\n\nbasic_search = ShareSearch()\nbasic_search.aggs.bucket(\n 'sourceAgg',\n 'terms',\n field='_type',\n size=0,\n min_doc_count=0\n)\n","subject":"Fix count param, use local es for now","message":"Fix count param, use local es for now\n","lang":"Python","license":"mit","repos":"CenterForOpenScience\/sharepa,fabianvf\/sharepa"} {"commit":"7e883fcfc539f18cd29c2babaf083583495f46d3","old_file":"migrations\/versions\/1f9c61031fa_.py","new_file":"migrations\/versions\/1f9c61031fa_.py","old_contents":"\"\"\"empty message\n\nRevision ID: 1f9c61031fa\nRevises: 1f872d11bbf\nCreate Date: 2016-01-24 17:46:54.879784\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '1f9c61031fa'\ndown_revision = '1f872d11bbf'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('setting', sa.Column('key', sa.String(length=100)))\n op.drop_column('setting', 'id')\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('setting', sa.Column('id', sa.INTEGER(), nullable=False))\n op.drop_column('setting', 'key')\n ### end Alembic commands ###\n","new_contents":"\"\"\"empty message\n\nRevision ID: 1f9c61031fa\nRevises: 1f872d11bbf\nCreate Date: 2016-01-24 17:46:54.879784\n\n\"\"\"\n\n# revision identifiers, used by Alembic.\nrevision = '1f9c61031fa'\ndown_revision = '1f872d11bbf'\n\nfrom alembic import op\nimport sqlalchemy as sa\n\n\ndef upgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('setting', sa.Column('key', sa.String(length=100)))\n op.alter_column('setting', 'name',\n existing_type=sa.VARCHAR(length=100),\n nullable=True) \n op.drop_column('setting', 'id')\n ### end Alembic commands ###\n\n\ndef downgrade():\n ### commands auto generated by Alembic - please adjust! ###\n op.add_column('setting', sa.Column('id', sa.INTEGER(), nullable=False))\n op.alter_column('setting', 'name',\n existing_type=sa.VARCHAR(length=100),\n nullable=False) \n op.drop_column('setting', 'key')\n ### end Alembic commands ###\n","subject":"Fix NOT NULL constraint on Setting name not being removed","message":"Fix NOT NULL constraint on Setting name not being removed\n","lang":"Python","license":"mit","repos":"Encrylize\/flask-blogger,Encrylize\/flask-blogger,Encrylize\/flask-blogger"} {"commit":"26c1daab6095c6110995104b94ad5b6260557c70","old_file":"aiortp\/sdp.py","new_file":"aiortp\/sdp.py","old_contents":"class SDP:\n def __init__(self, local_addr, ptime):\n self.local_addr = local_addr\n self.ptime = ptime\n\n local_addr_desc = f'IN IP4 {self.local_addr[0]}'\n self.payload = '\\r\\n'.join([\n 'v=0',\n f'o=user1 53655765 2353687637 {local_addr_desc}',\n 's=-',\n 't=0 0',\n 'i=aiortp media stream',\n f'm=audio {self.local_addr[1]} RTP\/AVP 0 101 13',\n f'c={local_addr_desc}',\n 'a=rtpmap:0 PCMU\/8000\/1',\n 'a=rtpmap:101 telephone-event\/8000',\n 'a=fmtp:101 0-15',\n f'a=ptime:{self.ptime}',\n 'a=sendrecv',\n '',\n ])\n\n def __str__(self):\n return self.payload\n","new_contents":"class SDP:\n def __init__(self, local_addr, ptime):\n self.local_addr = local_addr\n self.ptime = ptime\n\n local_addr_desc = 'IN IP4 {}'.format(self.local_addr[0])\n self.payload = '\\r\\n'.join([\n 'v=0',\n 'o=user1 53655765 2353687637 {local_addr_desc}',\n 's=-',\n 't=0 0',\n 'i=aiortp media stream',\n 'm=audio {local_port} RTP\/AVP 0 101 13',\n 'c={local_addr_desc}',\n 'a=rtpmap:0 PCMU\/8000\/1',\n 'a=rtpmap:101 telephone-event\/8000',\n 'a=fmtp:101 0-15',\n 'a=ptime:{ptime}',\n 'a=sendrecv',\n '',\n ]).format(local_addr_desc=local_addr_desc,\n local_port=self.local_addr[1],\n ptime=self.ptime)\n\n def __str__(self):\n return self.payload\n","subject":"Remove python 3.6 only format strings","message":"Remove python 3.6 only format strings\n","lang":"Python","license":"apache-2.0","repos":"vodik\/aiortp"} {"commit":"e2ee9045c59e3f03c5342ee41d23e4adece43535","old_file":"weather\/admin.py","new_file":"weather\/admin.py","old_contents":"from django.contrib.admin import ModelAdmin, register\nfrom django.contrib.gis.admin import GeoModelAdmin\nfrom weather.models import WeatherStation, Location\n\n\n@register(Location)\nclass LocationAdmin(GeoModelAdmin):\n pass\n\n\n@register(WeatherStation)\nclass WeatherStationAdmin(ModelAdmin):\n list_display = (\n 'name', 'abbreviation', 'ip_address', 'last_reading',\n 'battery_voltage', 'connect_every', 'active')\n","new_contents":"from django.contrib.admin import ModelAdmin, register\nfrom django.contrib.gis.admin import GeoModelAdmin\nfrom weather.models import WeatherStation, Location\n\n\n@register(Location)\nclass LocationAdmin(GeoModelAdmin):\n openlayers_url = '\/\/static.dpaw.wa.gov.au\/static\/libs\/openlayers\/2.13.1\/OpenLayers.js'\n\n\n@register(WeatherStation)\nclass WeatherStationAdmin(ModelAdmin):\n list_display = (\n 'name', 'abbreviation', 'ip_address', 'last_reading',\n 'battery_voltage', 'connect_every', 'active')\n","subject":"Define URL for OpenLayers.js to DPaW CDN.","message":"Define URL for OpenLayers.js to DPaW CDN.\n","lang":"Python","license":"bsd-3-clause","repos":"parksandwildlife\/resource_tracking,parksandwildlife\/resource_tracking,ropable\/resource_tracking,ropable\/resource_tracking,ropable\/resource_tracking,parksandwildlife\/resource_tracking"} {"commit":"a6bd1cfc5f87d6f9a7ac846665fcab5b02c33c1d","old_file":"tubular\/scripts\/hipchat\/submit_hipchat_msg.py","new_file":"tubular\/scripts\/hipchat\/submit_hipchat_msg.py","old_contents":"import os\nimport sys\nimport requests\nimport click\n\n\nHIPCHAT_API_URL = \"http:\/\/api.hipchat.com\"\nNOTIFICATION_POST = \"\/v2\/room\/{}\/notification\"\nAUTH_HEADER = \"Authorization: Bearer {}\"\n\n\n@click.command()\n@click.option('--auth_token_env_var', '-a',\n help=\"Environment variable containing authentication token to use for HipChat REST API.\",\n )\n@click.option('--channel', '-c',\n default=\"release pipeline\",\n help=\"Channel to which the script should post a message.\",\n )\ndef cli(auth_token_env_var, channel):\n \"\"\"\n Post a message to a HipChat channel.\n \"\"\"\n msg = \"Test message from the demo GoCD release pipeline.\"\n\n headers = {\n \"Authorization\": \"Bearer {}\".format(os.environ[auth_token_env_var])\n }\n msg_payload = {\n \"color\": \"green\",\n \"message\": msg,\n \"notify\": False,\n \"message_format\": \"text\"\n }\n post_url = HIPCHAT_API_URL + NOTIFICATION_POST.format(channel)\n r = requests.post(post_url, headers=headers, json=msg_payload)\n\n # An exit code of 0 means success and non-zero means failure.\n success = r.status_code in (200, 201, 204)\n sys.exit(not success)\n\n\nif __name__ == '__main__':\n cli()\n","new_contents":"import os\nimport sys\nimport requests\nimport click\n\n\nHIPCHAT_API_URL = \"http:\/\/api.hipchat.com\"\nNOTIFICATION_POST = \"\/v2\/room\/{}\/notification\"\nAUTH_HEADER = \"Authorization: Bearer {}\"\n\n\n@click.command()\n@click.option('--auth_token_env_var', '-a',\n help=\"Environment variable containing authentication token to use for HipChat REST API.\",\n )\n@click.option('--channel', '-c',\n default=\"release pipeline\",\n help=\"Channel to which the script should post a message.\",\n )\n@click.option('--message', '-m',\n default=\"Default message.\",\n help=\"Message to send to HipChat channel.\",\n )\ndef cli(auth_token_env_var, channel, message):\n \"\"\"\n Post a message to a HipChat channel.\n \"\"\"\n headers = {\n \"Authorization\": \"Bearer {}\".format(os.environ[auth_token_env_var])\n }\n msg_payload = {\n \"color\": \"green\",\n \"message\": message,\n \"notify\": False,\n \"message_format\": \"text\"\n }\n post_url = HIPCHAT_API_URL + NOTIFICATION_POST.format(channel)\n r = requests.post(post_url, headers=headers, json=msg_payload)\n\n # An exit code of 0 means success and non-zero means failure.\n success = r.status_code in (200, 201, 204)\n sys.exit(not success)\n\n\nif __name__ == '__main__':\n cli()\n","subject":"Add ability to set HipChat message contents.","message":"Add ability to set HipChat message contents.\n","lang":"Python","license":"agpl-3.0","repos":"eltoncarr\/tubular,eltoncarr\/tubular"} {"commit":"e58e33fbce322ba94ce60afc053b3610ad5bf993","old_file":"opps\/__init__.py","new_file":"opps\/__init__.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nimport pkg_resources\n\npkg_resources.declare_namespace(__name__)\n\nVERSION = (0, 1, 4)\n\n__version__ = \".\".join(map(str, VERSION))\n__status__ = \"Development\"\n__description__ = u\"Opps CMS websites magazines and high-traffic\"\n\n__author__ = u\"Thiago Avelino\"\n__credits__ = []\n__email__ = u\"opps-developers@googlegroups.com\"\n__license__ = u\"MIT License\"\n__copyright__ = u\"Copyright 2013, YACOWS\"\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nimport pkg_resources\n\npkg_resources.declare_namespace(__name__)\n\nVERSION = (0, 1, 5)\n\n__version__ = \".\".join(map(str, VERSION))\n__status__ = \"Development\"\n__description__ = u\"Opps CMS websites magazines and high-traffic\"\n\n__author__ = u\"Thiago Avelino\"\n__credits__ = []\n__email__ = u\"opps-developers@googlegroups.com\"\n__license__ = u\"MIT License\"\n__copyright__ = u\"Copyright 2013, YACOWS\"\n","subject":"Upgrade version 0.1.4 to 0.1.5","message":"Upgrade version 0.1.4 to 0.1.5\n","lang":"Python","license":"mit","repos":"jeanmask\/opps,YACOWS\/opps,opps\/opps,williamroot\/opps,williamroot\/opps,YACOWS\/opps,opps\/opps,williamroot\/opps,williamroot\/opps,jeanmask\/opps,YACOWS\/opps,jeanmask\/opps,YACOWS\/opps,jeanmask\/opps,opps\/opps,opps\/opps"} {"commit":"283f4d0dc1896b35e1c6be3458a99c87b9296659","old_file":"amaascore\/asset_managers\/enums.py","new_file":"amaascore\/asset_managers\/enums.py","old_contents":"from __future__ import absolute_import, division, print_function, unicode_literals\n\n\nASSET_MANAGER_TYPES = {'Accredited Investor', 'Bank', 'Broker', 'Corporate Treasury', 'Family Office',\n 'Fund Administrator', 'Fund Manager', 'Hedge Fund', 'Private Equity', 'Retail',\n 'Venture Capital'}\nRELATIONSHIP_TYPES = {'Administrator', 'External', 'Front Office', 'Employee'}\nACCOUNT_TYPES = {'Test', 'Basic', 'Professional'}","new_contents":"from __future__ import absolute_import, division, print_function, unicode_literals\n\n\nASSET_MANAGER_TYPES = {'Accredited Investor', 'Bank', 'Broker', 'Corporate Treasury', 'Family Office',\n 'Fund Administrator', 'Fund Manager', 'Hedge Fund', 'Individual', 'Private Equity',\n 'Venture Capital'}\nRELATIONSHIP_TYPES = {'Administrator', 'External', 'Front Office', 'Employee'}\nACCOUNT_TYPES = {'Test', 'Basic', 'Professional'}","subject":"Migrate “Retail” to “Individual” for clarity. AMAAS-764.","message":"Migrate “Retail” to “Individual” for clarity. AMAAS-764.\n","lang":"Python","license":"apache-2.0","repos":"paul-rs\/amaas-core-sdk-python,amaas-fintech\/amaas-core-sdk-python,nedlowe\/amaas-core-sdk-python,nedlowe\/amaas-core-sdk-python,amaas-fintech\/amaas-core-sdk-python,paul-rs\/amaas-core-sdk-python"} {"commit":"b17104be53389604b4b7f5f109895bdaa6389e43","old_file":"hic\/flow.py","new_file":"hic\/flow.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom __future__ import division\n\nimport numpy as np\nimport numexpr as ne\n\n__all__ = 'qn', 'FlowCumulant'\n\n\ndef qn(n, phi):\n return ne.evaluate('sum(exp(1j*n*phi))')\n\n\nclass FlowCumulant(object):\n def __init__(self, multiplicities, qn):\n self.multiplicities = np.asarray(multiplicities)\n self._qn = dict(qn)\n self._corr2 = {}\n self._corr4 = {}\n\n def _calculate_corr2(self, n):\n try:\n qn = self._qn[n] # noqa\n except KeyError:\n raise\n\n M = self.multiplicities # noqa\n self._corr[n][2] = ne.evaluate(\n 'sum(qn*conj(qn) - M) \/ sum(M*(M-1))'\n )\n\n def _calculate_corr4(self, n):\n pass\n\n def _get_corr(self, n, k):\n pass\n\n def correlation(self, n, k):\n pass\n\n def cumulant(self, n, k, error=False, negative_imaginary=False):\n pass\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom __future__ import division\n\nimport numpy as np\nimport numexpr as ne\n\n__all__ = 'qn', 'FlowCumulant'\n\n\n# If a variable is only ever used by numexpr, flake8 will flag it as unused.\n# The comment 'noqa' prevents this warning.\n\n\ndef qn(n, phi):\n return ne.evaluate('sum(exp(1j*n*phi))')\n\n\nclass FlowCumulant(object):\n def __init__(self, multiplicities, qn):\n self.multiplicities = np.asarray(multiplicities)\n self._qn = dict(qn)\n self._corr2 = {}\n self._corr4 = {}\n\n def _calculate_corr2(self, n):\n try:\n qn = self._qn[n] # noqa\n except KeyError:\n raise\n\n M = self.multiplicities # noqa\n self._corr[n][2] = ne.evaluate(\n 'sum(qn*conj(qn) - M) \/ sum(M*(M-1))'\n )\n\n def _calculate_corr4(self, n):\n pass\n\n def _get_corr(self, n, k):\n pass\n\n def correlation(self, n, k):\n pass\n\n def cumulant(self, n, k, error=False, negative_imaginary=False):\n pass\n","subject":"Add note about flake8 ignore flag.","message":"Add note about flake8 ignore flag.\n","lang":"Python","license":"mit","repos":"jbernhard\/hic,Duke-QCD\/hic"} {"commit":"bbd3190b31a3751d9173b81d6f53c937208969a7","old_file":"tests\/main_test.py","new_file":"tests\/main_test.py","old_contents":"#!\/usr\/bin\/env python3\n\nfrom libpals.util import xor_find_singlechar_key, hamming_distance, fixed_xor\n\ndef test_xor_find_singlechar_key():\n input = '1b37373331363f78151b7f2b783431333d78397828372d363c78373e783a393b3736'\n ciphertext = bytes.fromhex(input)\n result = xor_find_singlechar_key(ciphertext)\n assert result['key'] == 88\n assert result['plaintext'] == b\"Cooking MC's like a pound of bacon\"\n\n\ndef test_hamming_distance():\n assert hamming_distance(b\"this is a test\", b\"wokka wokka!!!\") == 37\n\n\ndef test_fixed_xor():\n input = bytes.fromhex(\"1c0111001f010100061a024b53535009181c\")\n key = bytes.fromhex(\"686974207468652062756c6c277320657965\")\n assert fixed_xor(input, key) == b\"the kid don't play\"\n","new_contents":"#!\/usr\/bin\/env python3\n\nfrom libpals.util import (\n xor_find_singlechar_key,\n hamming_distance,\n fixed_xor\n)\n\ndef test_xor_find_singlechar_key():\n input = '1b37373331363f78151b7f2b783431333d78397828372d363c78373e783a393b3736'\n ciphertext = bytes.fromhex(input)\n result = xor_find_singlechar_key(ciphertext)\n assert result['key'] == 88\n assert result['plaintext'] == b\"Cooking MC's like a pound of bacon\"\n\n\ndef test_hamming_distance():\n assert hamming_distance(b\"this is a test\", b\"wokka wokka!!!\") == 37\n\n\ndef test_fixed_xor():\n input = bytes.fromhex(\"1c0111001f010100061a024b53535009181c\")\n key = bytes.fromhex(\"686974207468652062756c6c277320657965\")\n assert fixed_xor(input, key) == b\"the kid don't play\"\n","subject":"Change to multi-line imports in the test suite","message":"Change to multi-line imports in the test suite\n","lang":"Python","license":"bsd-2-clause","repos":"cpach\/cryptopals-python3"} {"commit":"6a54876a30ec92c0243cf758b7ddf35d7ad3b926","old_file":"pgup\/__init__.py","new_file":"pgup\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n__version__ = \"0.2.3\"\nfrom config import Config\nfrom build_init import build_init\nfrom build_diff import build_diff\n","new_contents":"# -*- coding: utf-8 -*-\n__version__ = \"0.2.4\"\nfrom config import Config\nfrom build_init import build_init\nfrom build_diff import build_diff\n","subject":"Add skiping not Table, Procedure files","message":"Add skiping not Table, Procedure files\n","lang":"Python","license":"mit","repos":"stepan-perlov\/pgup"} {"commit":"e44eb0bd99b4dec1b78707c7343fc6d9b647c7bb","old_file":"scripts\/write_antenna_location_file.py","new_file":"scripts\/write_antenna_location_file.py","old_contents":"#! \/usr\/bin\/env python\n# -*- mode: python; coding: utf-8 -*-\n# Copyright 2016 the HERA Collaboration\n# Licensed under the 2-clause BSD license.\n\n\"\"\"\nScript to write out antenna locations for use in cal files.\n\"\"\"\nimport pandas as pd\nfrom hera_mc import mc, geo_handling\nimport datetime\n\nparser = mc.get_mc_argument_parser()\nparser.add_argument('--file', help=\"file name to save antenna locations to\",\n default='hera_ant_locs_' + datetime.date.today().strftime(\"%m_%d_%Y\") + '.csv')\nargs = parser.parse_args()\nfilename = args.file\ndb = mc.connect_to_mc_db(args)\n\nlocations = geo_handling.get_all_locations(args)\ndf = pd.DataFrame(locations)\ndf = df[['station_name', 'station_type', 'longitude', 'latitude', 'elevation',\n 'antenna_number', 'start_date', 'stop_date']]\ndf.to_csv(filename, index=False)\n","new_contents":"#! \/usr\/bin\/env python\n# -*- mode: python; coding: utf-8 -*-\n# Copyright 2016 the HERA Collaboration\n# Licensed under the 2-clause BSD license.\n\n\"\"\"\nScript to write out antenna locations for use in cal files.\n\"\"\"\nimport pandas as pd\nfrom hera_mc import mc, geo_handling\nimport datetime\n\nparser = mc.get_mc_argument_parser()\nparser.add_argument('--file', help=\"file name to save antenna locations to\",\n default='hera_ant_locs_' + datetime.date.today().strftime(\"%m_%d_%Y\") + '.csv')\nargs = parser.parse_args()\nfilename = args.file\ndb = mc.connect_to_mc_db(args)\n\nlocations = geo_handling.get_all_locations(args)\ncofa_loc = geo_handling.cofa()\nlocations.append({'station_name': cofa_loc.station_name,\n 'station_type': cofa_loc.station_type_name,\n 'longitude': cofa_loc.lon,\n 'latitude': cofa_loc.lat,\n 'elevation': cofa_loc.elevation,\n 'antenna_number': None,\n 'start_date': cofa_loc.created_date,\n 'stop_date': None})\ndf = pd.DataFrame(locations)\ndf = df[['station_name', 'station_type', 'longitude', 'latitude', 'elevation',\n 'antenna_number', 'start_date', 'stop_date']]\ndf.to_csv(filename, index=False)\n","subject":"Add cofa information to antenna location files","message":"Add cofa information to antenna location files\n\n","lang":"Python","license":"bsd-2-clause","repos":"HERA-Team\/hera_mc,HERA-Team\/Monitor_and_Control,HERA-Team\/hera_mc"} {"commit":"6bbee1abf5e53c3e3ce9cc84f35820ee9fe11500","old_file":"dsub\/_dsub_version.py","new_file":"dsub\/_dsub_version.py","old_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Single source of truth for dsub's version.\n\nThis must remain small and dependency-free so that any dsub module may\nimport it without creating circular dependencies. Note that this module\nis parsed as a text file by setup.py and changes to the format of this\nfile could break setup.py.\n\nThe version should follow formatting requirements specified in PEP-440.\n - https:\/\/www.python.org\/dev\/peps\/pep-0440\n\nA typical release sequence will be versioned as:\n 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...\n\"\"\"\n\nDSUB_VERSION = '0.3.9.dev0'\n","new_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Single source of truth for dsub's version.\n\nThis must remain small and dependency-free so that any dsub module may\nimport it without creating circular dependencies. Note that this module\nis parsed as a text file by setup.py and changes to the format of this\nfile could break setup.py.\n\nThe version should follow formatting requirements specified in PEP-440.\n - https:\/\/www.python.org\/dev\/peps\/pep-0440\n\nA typical release sequence will be versioned as:\n 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...\n\"\"\"\n\nDSUB_VERSION = '0.3.9'\n","subject":"Update dsub version to 0.3.9","message":"Update dsub version to 0.3.9\n\nPiperOrigin-RevId: 319808345\n","lang":"Python","license":"apache-2.0","repos":"DataBiosphere\/dsub,DataBiosphere\/dsub"} {"commit":"e3cba925ea106baa99951ac7b3ee72599ee7277d","old_file":"demos\/fs-demo\/main.py","new_file":"demos\/fs-demo\/main.py","old_contents":"import random\nimport os\n\nfrom microbit import *\n\nif 'messages.txt' in os.listdir():\n\n with open('messages.txt') as message_file:\n messages = message_file.read().split('\\n')\n\n while True:\n if button_a.was_pressed():\n display.scroll(random.choice(messages))\n","new_contents":"import random\nimport os\nimport speech\n\nfrom microbit import *\n\nif 'messages.txt' in os.listdir():\n\n with open('messages.txt') as message_file:\n messages = message_file.read().split('\\n')\n\n while True:\n if button_a.was_pressed():\n speech.say(random.choice(messages))\n","subject":"Change output in fs-demo to voice.","message":"Change output in fs-demo to voice.\n","lang":"Python","license":"mit","repos":"mathisgerdes\/microbit-macau"} {"commit":"133792d58e30bd0266355e0a87b1fc3f2d040302","old_file":"dimod\/package_info.py","new_file":"dimod\/package_info.py","old_contents":"# Copyright 2018 D-Wave Systems Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# ================================================================================================\n\n__version__ = '0.7.11'\n__author__ = 'D-Wave Systems Inc.'\n__authoremail__ = 'acondello@dwavesys.com'\n__description__ = 'A shared API for binary quadratic model samplers.'\n","new_contents":"# Copyright 2018 D-Wave Systems Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# ================================================================================================\n\n__version__ = '0.8.0'\n__author__ = 'D-Wave Systems Inc.'\n__authoremail__ = 'acondello@dwavesys.com'\n__description__ = 'A shared API for binary quadratic model samplers.'\n","subject":"Update version 0.7.11 -> 0.8.0","message":"Update version 0.7.11 -> 0.8.0","lang":"Python","license":"apache-2.0","repos":"oneklc\/dimod,oneklc\/dimod"} {"commit":"4545d11c2462ccb6d7848d185f5fe358a51af5f6","old_file":"Trimmer.py","new_file":"Trimmer.py","old_contents":"import sublime\nimport sublime_plugin\n\n\nclass TrimmerCommand(sublime_plugin.TextCommand):\n def run(self, edit):\n view = self.view\n\n trailing_white_space = view.find_all(\"[\\t ]+$\")\n trailing_white_space.reverse()\n\n edit = view.begin_edit()\n\n for r in trailing_white_space:\n view.erase(edit, r)\n\n view.end_edit(edit)\n sublime.set_timeout(lambda: self.save(view), 10)\n\n def save(self, view):\n if view.file_name() is None:\n view.run_command('prompt_save_as')\n else:\n view.run_command('save')\n sublime.status_message('Trimmer: Removed trailing whitespace and saved.')\n","new_contents":"import sublime\nimport sublime_plugin\n\n\nclass TrimmerCommand(sublime_plugin.TextCommand):\n def run(self, edit):\n view = self.view\n\n trailing_white_space = view.find_all(\"[\\t ]+$\")\n trailing_white_space.reverse()\n\n for r in trailing_white_space:\n view.erase(edit, r)\n\n sublime.set_timeout(lambda: self.save(view), 10)\n\n def save(self, view):\n if view.file_name() is None:\n view.run_command('prompt_save_as')\n else:\n view.run_command('save')\n sublime.status_message('Trimmer: Removed trailing whitespace and saved.')\n","subject":"Remove calls to begin, end edit object.","message":"Remove calls to begin, end edit object.\n","lang":"Python","license":"mit","repos":"jonlabelle\/Trimmer,jonlabelle\/Trimmer"} {"commit":"ca6891f3b867fd691c0b682566ffec1fd7f0ac2a","old_file":"pryvate\/blueprints\/simple\/simple.py","new_file":"pryvate\/blueprints\/simple\/simple.py","old_contents":"\"\"\"Simple blueprint.\"\"\"\nimport os\n\nfrom flask import Blueprint, current_app, make_response, render_template\n\nblueprint = Blueprint('simple', __name__, url_prefix='\/simple',\n template_folder='templates')\n\n\n@blueprint.route('', methods=['POST'])\ndef search_simple():\n \"\"\"Handling pip search.\"\"\"\n return make_response('Not implemented', 501)\n\n\n@blueprint.route('', methods=['GET'])\ndef get_simple():\n \"\"\"List all packages.\"\"\"\n packages = os.listdir(current_app.config['BASEDIR'])\n return render_template('simple.html', packages=packages)\n\n\n@blueprint.route('\/<package>', methods=['GET'])\n@blueprint.route('\/<package>\/', methods=['GET'])\ndef get_package(package):\n \"\"\"List versions of a package.\"\"\"\n package_path = os.path.join(current_app.config['BASEDIR'],\n package.lower())\n files = os.listdir(package_path)\n\n packages = []\n for filename in files:\n if filename.endswith('md5'):\n with open(os.path.join(package_path, filename), 'r') as md5_digest: \n item = {\n 'name': package,\n 'version': filename.replace('.md5', ''),\n 'digest': md5_digest.read()\n }\n packages.append(item)\n return render_template('simple_package.html', packages=packages,\n letter=package[:1].lower())\n","new_contents":"\"\"\"Simple blueprint.\"\"\"\nimport os\n\nfrom flask import Blueprint, current_app, make_response, render_template\n\nblueprint = Blueprint('simple', __name__, url_prefix='\/simple',\n template_folder='templates')\n\n\n@blueprint.route('', methods=['POST'])\ndef search_simple():\n \"\"\"Handling pip search.\"\"\"\n return make_response('Not implemented', 501)\n\n\n@blueprint.route('', methods=['GET'])\ndef get_simple():\n \"\"\"List all packages.\"\"\"\n packages = os.listdir(current_app.config['BASEDIR'])\n return render_template('simple.html', packages=packages)\n\n\n@blueprint.route('\/<package>', methods=['GET'])\n@blueprint.route('\/<package>\/', methods=['GET'])\ndef get_package(package):\n \"\"\"List versions of a package.\"\"\"\n package_path = os.path.join(current_app.config['BASEDIR'],\n package.lower())\n if os.path.isdir(package_path):\n files = os.listdir(package_path)\n\n packages = []\n for filename in files:\n if filename.endswith('md5'):\n digest_file = os.path.join(package_path, filename)\n with open(digest_file, 'r') as md5_digest:\n item = {\n 'name': package,\n 'version': filename.replace('.md5', ''),\n 'digest': md5_digest.read()\n }\n packages.append(item)\n return render_template('simple_package.html', packages=packages,\n letter=package[:1].lower())\n else:\n return make_response('404', 404)\n","subject":"Return 404 if package was not found instead of raising an exception","message":"Return 404 if package was not found instead of raising an exception\n","lang":"Python","license":"mit","repos":"Dinoshauer\/pryvate,Dinoshauer\/pryvate"} {"commit":"37c08c15ecc31f90429b20ee2d4561c2886036c2","old_file":"pyaavso\/utils.py","new_file":"pyaavso\/utils.py","old_contents":"from __future__ import unicode_literals\n\nimport logging\n\nimport requests\n\nfrom .parsers import WebObsResultsParser\n\n\nlogger = logging.getLogger(__name__)\n\nWEBOBS_RESULTS_URL = 'http:\/\/www.aavso.org\/apps\/webobs\/results\/'\n\n\ndef download_observations(observer_code):\n \"\"\"\n Downloads all variable star observations by a given observer.\n\n Performs a series of HTTP requests to AAVSO's WebObs search and\n downloads the results page by page. Each page is then passed to\n :py:class:`~pyaavso.parsers.webobs.WebObsResultsParser` and parse results\n are added to the final observation list.\n \"\"\"\n page_number = 1\n observations = []\n while True:\n logger.info('Downloading page %d...', page_number)\n response = requests.get(WEBOBS_RESULTS_URL, params={\n 'obscode': observer_code,\n 'num_results': 200,\n 'obs_types': 'all',\n 'page': page_number,\n })\n parser = WebObsResultsParser(response.content)\n observations.extend(parser.get_observations())\n # kinda silly, but there's no need for lxml machinery here\n if '>Next<\/a>' not in response.content:\n break\n page_number += 1\n return observations\n","new_contents":"from __future__ import unicode_literals\n\nimport logging\n\nimport requests\n\nfrom .parsers import WebObsResultsParser\n\n\nlogger = logging.getLogger(__name__)\n\nWEBOBS_RESULTS_URL = 'http:\/\/www.aavso.org\/apps\/webobs\/results\/'\n\n\ndef download_observations(observer_code):\n \"\"\"\n Downloads all variable star observations by a given observer.\n\n Performs a series of HTTP requests to AAVSO's WebObs search and\n downloads the results page by page. Each page is then passed to\n :py:class:`~pyaavso.parsers.webobs.WebObsResultsParser` and parse results\n are added to the final observation list.\n \"\"\"\n page_number = 1\n observations = []\n while True:\n logger.info('Downloading page %d...', page_number)\n response = requests.get(WEBOBS_RESULTS_URL, params={\n 'obscode': observer_code,\n 'num_results': 200,\n 'obs_types': 'all',\n 'page': page_number,\n })\n parser = WebObsResultsParser(response.text)\n observations.extend(parser.get_observations())\n # kinda silly, but there's no need for lxml machinery here\n if '>Next<\/a>' not in response.text:\n break\n page_number += 1\n return observations\n","subject":"Use response.text for automatic decoding.","message":"Use response.text for automatic decoding.\n","lang":"Python","license":"mit","repos":"zsiciarz\/pyaavso"} {"commit":"52e675ec6789d8ecaddae98a6b36bc8b0c3f6e1e","old_file":"socketio\/sdjango.py","new_file":"socketio\/sdjango.py","old_contents":"import logging\n\nfrom socketio import socketio_manage\nfrom django.conf.urls import patterns, url, include\nfrom django.http import HttpResponse\n\n\nSOCKETIO_NS = {}\n\nclass namespace(object):\n def __init__(self, name=''):\n self.name = name\n\n def __call__(self, handler):\n SOCKETIO_NS[self.name] = handler\n\n\ndef socketio(request):\n try:\n socketio_manage(request.environ, SOCKETIO_NS, request)\n except:\n logging.getLogger(\"socketio\").error(\"Exception while handling socketio connection\", exc_info=True)\n return HttpResponse(\"\")\n\n\nurls = patterns(\"\", (r'', socketio))\n","new_contents":"import logging\n\nfrom socketio import socketio_manage\nfrom django.conf.urls import patterns, url, include\nfrom django.http import HttpResponse\nfrom django.views.decorators.csrf import csrf_exempt\n\n\nSOCKETIO_NS = {}\n\nclass namespace(object):\n def __init__(self, name=''):\n self.name = name\n\n def __call__(self, handler):\n SOCKETIO_NS[self.name] = handler\n\n@csrf_exempt\ndef socketio(request):\n try:\n socketio_manage(request.environ, SOCKETIO_NS, request)\n except:\n logging.getLogger(\"socketio\").error(\"Exception while handling socketio connection\", exc_info=True)\n return HttpResponse(\"\")\n\n\nurls = patterns(\"\", (r'', socketio))\n","subject":"Remove django CSRF protection for socket.io view","message":"Remove django CSRF protection for socket.io view\n","lang":"Python","license":"bsd-3-clause","repos":"abourget\/gevent-socketio,yacneyac\/gevent-socketio,kazmiruk\/gevent-socketio,kazmiruk\/gevent-socketio,yacneyac\/gevent-socketio,arnuschky\/gevent-socketio,gutomaia\/gevent-socketio,smurfix\/gevent-socketio,arnuschky\/gevent-socketio,hzruandd\/gevent-socketio,theskumar-archive\/gevent-socketio,abourget\/gevent-socketio,gutomaia\/gevent-socketio,theskumar-archive\/gevent-socketio,smurfix\/gevent-socketio,smurfix\/gevent-socketio,gutomaia\/gevent-socketio,Eugeny\/gevent-socketio,bobvandevijver\/gevent-socketio,Eugeny\/gevent-socketio,bobvandevijver\/gevent-socketio,hzruandd\/gevent-socketio"} {"commit":"1fdceb6b8072317eff901a8f2b74db60269ca373","old_file":"python\/grade-school\/grade_school.py","new_file":"python\/grade-school\/grade_school.py","old_contents":"from collections import defaultdict\n\nclass School:\n def __init__(self):\n self.db = defaultdict(list)\n\n def add_student(self, name, grade):\n self.db[grade].append(name)\n self.db[grade] = sorted(self.db[grade])\n\n def roster(self):\n all_student_names = []\n for grade_number in sorted(self.db.keys()):\n all_student_names.extend(self.db[grade_number])\n return all_student_names\n\n def grade(self, grade_number):\n return self.db[grade_number]\n","new_contents":"from collections import defaultdict\n\n\nclass School:\n def __init__(self):\n self.db = defaultdict(list)\n\n def add_student(self, name, grade):\n self.db[grade].append(name)\n self.db[grade] = sorted(self.db[grade])\n\n def roster(self):\n all_student_names = []\n for grade_number in sorted(self.db.keys()):\n all_student_names.extend(self.db[grade_number])\n return all_student_names\n\n def grade(self, grade_number):\n return self.db[grade_number]\n","subject":"Add newline to fix lint warning","message":"Add newline to fix lint warning\n","lang":"Python","license":"mit","repos":"rootulp\/exercism,rootulp\/exercism,rootulp\/exercism,rootulp\/exercism,rootulp\/exercism,rootulp\/exercism,rootulp\/exercism,rootulp\/exercism"} {"commit":"8ff8b9400adf24e082908befed7788099b01f328","old_file":"bench\/pact-suite\/scripts\/opcount_merge.py","new_file":"bench\/pact-suite\/scripts\/opcount_merge.py","old_contents":"#!\/usr\/bin\/env python2.7\nimport sys\n\nfiles = sys.argv[1:]\n\nkeys = set()\nfileVals = []\nfor file in files:\n vals = {}\n fileVals.append(vals)\n try:\n for line in open(file).readlines():\n k, v = line.split()\n vals[k] = v\n keys.add(k)\n except Exception, e:\n print \"Error in line \\\"%s\\\" of file %s\" % (line, file)\n raise e\n\n\nfor key in sorted(keys):\n sys.stdout.write(key)\n for vals in fileVals:\n sys.stdout.write(\"\\t\")\n sys.stdout.write(str(vals.get(key, 0)))\n sys.stdout.write(\"\\n\")\n","new_contents":"#!\/usr\/bin\/env python2.7\nimport sys\n\nfiles = sys.argv[1:]\n\nkeys = set()\nfileVals = []\nfor file in files:\n vals = {}\n fileVals.append(vals)\n try:\n for line in open(file).readlines():\n toks = line.split()\n if len(toks) != 2:\n print >> sys.stderr, \"Bad line: %s\" % repr(toks)\n else:\n k, v = toks\n vals[k] = v\n keys.add(k)\n except Exception, e:\n print >> sys.stderr, \"Error in line \\\"%s\\\" of file %s\" % (line, file)\n raise e\n\n\nfor key in sorted(keys):\n sys.stdout.write(key)\n for vals in fileVals:\n sys.stdout.write(\"\\t\")\n sys.stdout.write(str(vals.get(key, 0)))\n sys.stdout.write(\"\\n\")\n","subject":"Add initial data for operations counts","message":"Add initial data for operations counts\n\ngit-svn-id: 0c5512015aa96f7d3f5c3ad598bd98edc52008b1@12204 dc4e9af1-7f46-4ead-bba6-71afc04862de\n","lang":"Python","license":"apache-2.0","repos":"basheersubei\/swift-t,blue42u\/swift-t,swift-lang\/swift-t,JohnPJenkins\/swift-t,swift-lang\/swift-t,blue42u\/swift-t,swift-lang\/swift-t,swift-lang\/swift-t,JohnPJenkins\/swift-t,basheersubei\/swift-t,JohnPJenkins\/swift-t,JohnPJenkins\/swift-t,JohnPJenkins\/swift-t,basheersubei\/swift-t,basheersubei\/swift-t,basheersubei\/swift-t,JohnPJenkins\/swift-t,JohnPJenkins\/swift-t,swift-lang\/swift-t,basheersubei\/swift-t,blue42u\/swift-t,blue42u\/swift-t,swift-lang\/swift-t,blue42u\/swift-t,swift-lang\/swift-t,blue42u\/swift-t,blue42u\/swift-t,basheersubei\/swift-t"} {"commit":"b6099b64efb6e74e754d0911bd64512b6b6b631b","old_file":"quokka\/modules\/accounts\/tests\/test_model.py","new_file":"quokka\/modules\/accounts\/tests\/test_model.py","old_contents":"# coding: utf-8\nfrom flask.ext.testing import TestCase\n\nfrom quokka import create_app\nfrom flask.ext.security.utils import encrypt_password\nfrom ..models import User\n\n\nclass TestAuthModels(TestCase):\n def setUp(self):\n self.db = self.app.extensions.get('mongoengine')\n self.user_dict = {\n 'name': u'Guybrush Treepwood',\n 'email': u'guybrush@monkeyisland.com',\n 'password': encrypt_password(u'lechucksucks'),\n }\n self.user = User.objects.create(**self.user_dict)\n\n def tearDown(self):\n User.objects.all().delete()\n\n def create_app(self):\n return create_app(config='quokka.test_settings',\n DEBUG=False,\n test=True)\n\n def test_user_fields(self):\n self.assertIsInstance(self.user, User)\n self.assertEqual(self.user.username, u'guybrush_monkeyisland_com')\n self.assertEqual(self.user.name, u'Guybrush Treepwood')\n self.assertEqual(self.user.email, u'guybrush@monkeyisland.com')\n self.assertEqual(self.user.password, self.user_dict['password'])\n self.assertEqual(self.user.display_name, self.user.name)\n","new_contents":"# coding: utf-8\nfrom flask.ext.testing import TestCase\n\nfrom quokka import create_app\nfrom flask.ext.security.utils import encrypt_password\nfrom ..models import User\n\n\nclass TestAuthModels(TestCase):\n def setUp(self):\n self.user_dict = {\n 'name': u'Guybrush Treepwood',\n 'email': u'guybrush@monkeyisland.com',\n 'password': encrypt_password(u'lechucksucks'),\n }\n self.user = User.objects.create(**self.user_dict)\n\n def tearDown(self):\n User.objects.all().delete()\n\n def create_app(self):\n return create_app(config='quokka.test_settings',\n DEBUG=False,\n test=True)\n\n def test_user_fields(self):\n self.assertIsInstance(self.user, User)\n self.assertEqual(self.user.username, u'guybrush_monkeyisland_com')\n self.assertEqual(self.user.name, u'Guybrush Treepwood')\n self.assertEqual(self.user.email, u'guybrush@monkeyisland.com')\n self.assertEqual(self.user.password, self.user_dict['password'])\n self.assertEqual(self.user.display_name, self.user.name)\n","subject":"Remove unused config on accounts tests","message":"Remove unused config on accounts tests\n","lang":"Python","license":"mit","repos":"maurobaraldi\/quokka,ChengChiongWah\/quokka,wushuyi\/quokka,cbeloni\/quokka,alexandre\/quokka,wushuyi\/quokka,cbeloni\/quokka,wushuyi\/quokka,fdumpling\/quokka,fdumpling\/quokka,lnick\/quokka,maurobaraldi\/quokka,fdumpling\/quokka,ChengChiongWah\/quokka,romulocollopy\/quokka,CoolCloud\/quokka,romulocollopy\/quokka,CoolCloud\/quokka,Ckai1991\/quokka,cbeloni\/quokka,lnick\/quokka,felipevolpone\/quokka,Ckai1991\/quokka,romulocollopy\/quokka,felipevolpone\/quokka,maurobaraldi\/quokka,cbeloni\/quokka,lnick\/quokka,romulocollopy\/quokka,Ckai1991\/quokka,ChengChiongWah\/quokka,fdumpling\/quokka,wushuyi\/quokka,felipevolpone\/quokka,maurobaraldi\/quokka,lnick\/quokka,CoolCloud\/quokka,felipevolpone\/quokka,ChengChiongWah\/quokka,Ckai1991\/quokka,CoolCloud\/quokka,alexandre\/quokka"} {"commit":"24e780dd0f30e4bf9696a6fd185d20fb297f0bd0","old_file":"rsk_mind\/transformer\/transformer.py","new_file":"rsk_mind\/transformer\/transformer.py","old_contents":"class Transformer(object):\n class Feats():\n exclude = None\n\n def __init__(self):\n for field in self.get_feats():\n getattr(self.Feats, field).bind(field, self)\n\n def get_feats(self):\n return [x for x in dir(self.Feats) if not (x.startswith('__') or x in ['exclude'])]\n\n def get_transformer_func(self, feat_name):\n return getattr(self.Feats, feat_name).transform\n","new_contents":"class Transformer(object):\n \"\"\"\n Base class for all transformer\n \"\"\"\n\n class Feats:\n \"\"\"\n Define feats on dataset\n \"\"\"\n exclude = None\n\n def __init__(self):\n for field in self.get_feats():\n getattr(self.Feats, field).bind(field, self)\n\n def get_feats(self):\n \"\"\"\n\n :return: a list of feats\n \"\"\"\n return [x for x in dir(self.Feats) if not (x.startswith('__') or x in ['exclude'])]\n\n def get_transformer_func(self, feat_name):\n \"\"\"\n\n :param feat_name: name of feat\n :return: a transformer function on feat\n \"\"\"\n return getattr(self.Feats, feat_name).transform\n\n def get_excluded_feats(self):\n \"\"\"\n\n :return: a list with excluded feats\n \"\"\"\n return self.Feats.exclude\n","subject":"Add documentation and some methods","message":"Add documentation and some methods\n","lang":"Python","license":"mit","repos":"rsk-mind\/rsk-mind-framework"} {"commit":"644c69a25d81ae4473d19bfe1faa6b7fe10e8afd","old_file":"dsub\/_dsub_version.py","new_file":"dsub\/_dsub_version.py","old_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Single source of truth for dsub's version.\n\nThis must remain small and dependency-free so that any dsub module may\nimport it without creating circular dependencies. Note that this module\nis parsed as a text file by setup.py and changes to the format of this\nfile could break setup.py.\n\nThe version should follow formatting requirements specified in PEP-440.\n - https:\/\/www.python.org\/dev\/peps\/pep-0440\n\nA typical release sequence will be versioned as:\n 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...\n\"\"\"\n\nDSUB_VERSION = '0.3.8'\n","new_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Single source of truth for dsub's version.\n\nThis must remain small and dependency-free so that any dsub module may\nimport it without creating circular dependencies. Note that this module\nis parsed as a text file by setup.py and changes to the format of this\nfile could break setup.py.\n\nThe version should follow formatting requirements specified in PEP-440.\n - https:\/\/www.python.org\/dev\/peps\/pep-0440\n\nA typical release sequence will be versioned as:\n 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...\n\"\"\"\n\nDSUB_VERSION = '0.3.9.dev0'\n","subject":"Update dsub version to 0.3.9.dev0","message":"Update dsub version to 0.3.9.dev0\n\nPiperOrigin-RevId: 313432033\n","lang":"Python","license":"apache-2.0","repos":"DataBiosphere\/dsub,DataBiosphere\/dsub"} {"commit":"a6ae4171de33dd77e9109523380c1330d4037f9f","old_file":"gengine\/app\/tests\/runner.py","new_file":"gengine\/app\/tests\/runner.py","old_contents":"from gengine.app.tests import db as db\nfrom gengine.metadata import init_declarative_base, init_session\nimport unittest\nimport os\nimport pkgutil\nimport testing.redis\nimport logging\n\nlog = logging.getLogger(__name__)\n\ninit_session()\ninit_declarative_base()\n\n__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]\n\ndef create_test_suite():\n suite = unittest.TestSuite()\n for imp, modname, _ in pkgutil.walk_packages(__path__):\n mod = imp.find_module(modname).load_module(modname)\n for test in unittest.defaultTestLoader.loadTestsFromModule(mod):\n suite.addTests(test)\n return suite\n\nif __name__==\"__main__\":\n exit = 1\n try:\n redis = testing.redis.RedisServer()\n\n from gengine.base.cache import setup_redis_cache\n dsn = redis.dsn()\n setup_redis_cache(dsn[\"host\"], dsn[\"port\"], dsn[\"db\"])\n\n from gengine.app.cache import init_caches\n init_caches()\n\n db.setupDB()\n testSuite = create_test_suite()\n text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)\n if text_runner.wasSuccessful():\n exit = 0\n finally:\n try:\n db.unsetupDB()\n except:\n log.exception()\n try:\n redis.stop()\n except:\n log.exception()\n sys.exit(exit)\n","new_contents":"from gengine.app.tests import db as db\nfrom gengine.metadata import init_declarative_base, init_session\nimport unittest\nimport os\nimport pkgutil\nimport testing.redis\nimport logging\nimport sys\n\nlog = logging.getLogger(__name__)\n\ninit_session()\ninit_declarative_base()\n\n__path__ = [x[0] for x in os.walk(os.path.dirname(__file__))]\n\ndef create_test_suite():\n suite = unittest.TestSuite()\n for imp, modname, _ in pkgutil.walk_packages(__path__):\n mod = imp.find_module(modname).load_module(modname)\n for test in unittest.defaultTestLoader.loadTestsFromModule(mod):\n suite.addTests(test)\n return suite\n\nif __name__==\"__main__\":\n exit = 1\n try:\n redis = testing.redis.RedisServer()\n\n from gengine.base.cache import setup_redis_cache\n dsn = redis.dsn()\n setup_redis_cache(dsn[\"host\"], dsn[\"port\"], dsn[\"db\"])\n\n from gengine.app.cache import init_caches\n init_caches()\n\n db.setupDB()\n testSuite = create_test_suite()\n text_runner = unittest.TextTestRunner(failfast=True).run(testSuite)\n if text_runner.wasSuccessful():\n exit = 0\n finally:\n try:\n db.unsetupDB()\n except:\n log.exception()\n try:\n redis.stop()\n except:\n log.exception()\n sys.exit(exit)\n","subject":"Add missing import for sys","message":"Add missing import for sys","lang":"Python","license":"mit","repos":"ActiDoo\/gamification-engine,ActiDoo\/gamification-engine,ActiDoo\/gamification-engine,ActiDoo\/gamification-engine"} {"commit":"2ebb667b38b3d74003948347f411f177ca584834","old_file":"boardinghouse\/contrib\/template\/models.py","new_file":"boardinghouse\/contrib\/template\/models.py","old_contents":"from django.db import models\nfrom django.utils import six\n\nfrom boardinghouse.base import SharedSchemaMixin\nfrom boardinghouse.schema import activate_schema, deactivate_schema\n\n\n@six.python_2_unicode_compatible\nclass SchemaTemplate(SharedSchemaMixin, models.Model):\n \"\"\"\n A ``boardinghouse.contrib.template.models.SchemaTemplate`` can be used\n for creating a new schema complete with some initial data.\n \"\"\"\n template_schema_id = models.AutoField(primary_key=True)\n name = models.CharField(max_length=128, unique=True)\n is_active = models.BooleanField(default=True)\n description = models.TextField(null=True, blank=True)\n\n class Meta:\n default_permissions = ('add', 'change', 'delete', 'view', 'activate', 'clone')\n verbose_name_plural = u'template schemata'\n\n def __str__(self):\n return self.name\n\n @property\n def schema(self):\n return '__template_{}'.format(self.pk)\n\n def activate(self):\n activate_schema(self.schema)\n\n def deactivate(self):\n deactivate_schema()\n","new_contents":"from django.db import models\nfrom django.utils import six\nfrom django.utils.functional import lazy\n\nfrom boardinghouse.base import SharedSchemaMixin\nfrom boardinghouse.schema import activate_schema, deactivate_schema, get_schema_model\n\n\ndef verbose_name_plural():\n return u'template {}'.format(get_schema_model()._meta.verbose_name_plural)\n\n\ndef verbose_name():\n return u'template {}'.format(get_schema_model()._meta.verbose_name)\n\n\n@six.python_2_unicode_compatible\nclass SchemaTemplate(SharedSchemaMixin, models.Model):\n \"\"\"\n A ``boardinghouse.contrib.template.models.SchemaTemplate`` can be used\n for creating a new schema complete with some initial data.\n \"\"\"\n template_schema_id = models.AutoField(primary_key=True)\n name = models.CharField(max_length=128, unique=True)\n is_active = models.BooleanField(default=True)\n description = models.TextField(null=True, blank=True)\n\n class Meta:\n default_permissions = ('add', 'change', 'delete', 'view', 'activate', 'clone')\n verbose_name = lazy(verbose_name, six.text_type)()\n verbose_name_plural = lazy(verbose_name_plural, six.text_type)()\n\n def __str__(self):\n return self.name\n\n @property\n def schema(self):\n return '__template_{}'.format(self.pk)\n\n def activate(self):\n activate_schema(self.schema)\n\n def deactivate(self):\n deactivate_schema()\n","subject":"Use 'template ...' for the SchemaTemplate verbose_name*","message":"Use 'template ...' for the SchemaTemplate verbose_name*\n","lang":"Python","license":"bsd-3-clause","repos":"schinckel\/django-boardinghouse,schinckel\/django-boardinghouse,schinckel\/django-boardinghouse"} {"commit":"0ea4abe8b2e44bdd02308ad590ffb1e846201300","old_file":"terms\/sitemaps.py","new_file":"terms\/sitemaps.py","old_contents":"from django.contrib.sitemaps import Sitemap\nfrom .models import Term\n\n\nclass TermsSitemap(Sitemap):\n changefreq = 'yearly'\n priority = 0.1\n\n def items(self):\n return Term.objects.all()\n","new_contents":"from django.contrib.sitemaps import Sitemap\nfrom django.db.models import Q\nfrom .models import Term\n\n\nclass TermsSitemap(Sitemap):\n changefreq = 'yearly'\n priority = 0.1\n\n def items(self):\n return Term.objects.filter(Q(url__startswith='\/') | Q(url=''))\n","subject":"Exclude external urls from the sitemap.","message":"Exclude external urls from the sitemap.\n","lang":"Python","license":"bsd-3-clause","repos":"philippeowagner\/django-terms,BertrandBordage\/django-terms,philippeowagner\/django-terms,BertrandBordage\/django-terms"} {"commit":"ddf2075228a8c250cf75ec85914801262cb73177","old_file":"zerver\/migrations\/0032_verify_all_medium_avatar_images.py","new_file":"zerver\/migrations\/0032_verify_all_medium_avatar_images.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations\nfrom django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor\nfrom django.db.migrations.state import StateApps\n\nfrom zerver.lib.upload import upload_backend\n\n\ndef verify_medium_avatar_image(apps, schema_editor):\n # type: (StateApps, DatabaseSchemaEditor) -> None\n user_profile_model = apps.get_model('zerver', 'UserProfile')\n for user_profile in user_profile_model.objects.filter(avatar_source=u\"U\"):\n upload_backend.ensure_medium_avatar_image(user_profile)\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('zerver', '0031_remove_system_avatar_source'),\n ]\n\n operations = [\n migrations.RunPython(verify_medium_avatar_image)\n ]\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom django.conf import settings\nfrom django.db import migrations\nfrom django.db.backends.postgresql_psycopg2.schema import DatabaseSchemaEditor\nfrom django.db.migrations.state import StateApps\n\nfrom mock import patch\nfrom zerver.lib.utils import make_safe_digest\nfrom zerver.lib.upload import upload_backend\nfrom zerver.models import UserProfile\nfrom typing import Text\nimport hashlib\n\n# We hackishly patch this function in order to revert it to the state\n# it had when this migration was first written. This is a balance\n# between copying in a historical version of hundreds of lines of code\n# from zerver.lib.upload (which would pretty annoying, but would be a\n# pain) and just using the current version, which doesn't work\n# since we rearranged the avatars in Zulip 1.6.\ndef patched_user_avatar_path(user_profile):\n # type: (UserProfile) -> Text\n email = user_profile.email\n user_key = email.lower() + settings.AVATAR_SALT\n return make_safe_digest(user_key, hashlib.sha1)\n\n@patch('zerver.lib.upload.user_avatar_path', patched_user_avatar_path)\ndef verify_medium_avatar_image(apps, schema_editor):\n # type: (StateApps, DatabaseSchemaEditor) -> None\n user_profile_model = apps.get_model('zerver', 'UserProfile')\n for user_profile in user_profile_model.objects.filter(avatar_source=u\"U\"):\n upload_backend.ensure_medium_avatar_image(user_profile)\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('zerver', '0031_remove_system_avatar_source'),\n ]\n\n operations = [\n migrations.RunPython(verify_medium_avatar_image)\n ]\n","subject":"Make migration 0032 use an old version of user_avatar_path.","message":"Make migration 0032 use an old version of user_avatar_path.\n\nThis fixes upgrading from very old Zulip servers (e.g. 1.4.3) all the\nway to current.\n\nFixes: #6516.\n","lang":"Python","license":"apache-2.0","repos":"hackerkid\/zulip,kou\/zulip,amanharitsh123\/zulip,brockwhittaker\/zulip,showell\/zulip,hackerkid\/zulip,rishig\/zulip,verma-varsha\/zulip,synicalsyntax\/zulip,zulip\/zulip,amanharitsh123\/zulip,showell\/zulip,punchagan\/zulip,amanharitsh123\/zulip,punchagan\/zulip,timabbott\/zulip,rht\/zulip,tommyip\/zulip,eeshangarg\/zulip,Galexrt\/zulip,eeshangarg\/zulip,rishig\/zulip,rishig\/zulip,Galexrt\/zulip,dhcrzf\/zulip,rht\/zulip,shubhamdhama\/zulip,rht\/zulip,tommyip\/zulip,mahim97\/zulip,kou\/zulip,kou\/zulip,timabbott\/zulip,zulip\/zulip,brainwane\/zulip,verma-varsha\/zulip,verma-varsha\/zulip,punchagan\/zulip,brainwane\/zulip,zulip\/zulip,eeshangarg\/zulip,timabbott\/zulip,synicalsyntax\/zulip,brockwhittaker\/zulip,brockwhittaker\/zulip,brainwane\/zulip,showell\/zulip,brainwane\/zulip,brainwane\/zulip,rishig\/zulip,tommyip\/zulip,Galexrt\/zulip,tommyip\/zulip,synicalsyntax\/zulip,zulip\/zulip,jackrzhang\/zulip,eeshangarg\/zulip,andersk\/zulip,kou\/zulip,punchagan\/zulip,rht\/zulip,kou\/zulip,andersk\/zulip,kou\/zulip,timabbott\/zulip,jackrzhang\/zulip,synicalsyntax\/zulip,eeshangarg\/zulip,brockwhittaker\/zulip,tommyip\/zulip,brainwane\/zulip,dhcrzf\/zulip,dhcrzf\/zulip,rishig\/zulip,showell\/zulip,showell\/zulip,andersk\/zulip,hackerkid\/zulip,brockwhittaker\/zulip,andersk\/zulip,showell\/zulip,timabbott\/zulip,dhcrzf\/zulip,mahim97\/zulip,jackrzhang\/zulip,andersk\/zulip,hackerkid\/zulip,jackrzhang\/zulip,jackrzhang\/zulip,amanharitsh123\/zulip,shubhamdhama\/zulip,amanharitsh123\/zulip,mahim97\/zulip,kou\/zulip,tommyip\/zulip,dhcrzf\/zulip,shubhamdhama\/zulip,rht\/zulip,rht\/zulip,eeshangarg\/zulip,shubhamdhama\/zulip,jackrzhang\/zulip,shubhamdhama\/zulip,zulip\/zulip,punchagan\/zulip,rishig\/zulip,shubhamdhama\/zulip,hackerkid\/zulip,shubhamdhama\/zulip,dhcrzf\/zulip,timabbott\/zulip,rishig\/zulip,punchagan\/zulip,zulip\/zulip,verma-varsha\/zulip,showell\/zulip,Galexrt\/zulip,mahim97\/zulip,Galexrt\/zulip,verma-varsha\/zulip,punchagan\/zulip,hackerkid\/zulip,jackrzhang\/zulip,Galexrt\/zulip,brainwane\/zulip,hackerkid\/zulip,synicalsyntax\/zulip,eeshangarg\/zulip,zulip\/zulip,dhcrzf\/zulip,andersk\/zulip,brockwhittaker\/zulip,mahim97\/zulip,amanharitsh123\/zulip,timabbott\/zulip,synicalsyntax\/zulip,rht\/zulip,tommyip\/zulip,synicalsyntax\/zulip,verma-varsha\/zulip,mahim97\/zulip,Galexrt\/zulip,andersk\/zulip"} {"commit":"674dfb000cca79998674cd0b490ae6f3f992b313","old_file":"blazarclient\/tests\/__init__.py","new_file":"blazarclient\/tests\/__init__.py","old_contents":"\n# Copyright (c) 2014 Mirantis.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\nfrom oslotest import base\nfrom oslotest import mockpatch\n\n\nclass TestCase(base.BaseTestCase):\n \"\"\"Test case base class for all unit tests.\"\"\"\n\n def patch(self, obj, attr):\n \"\"\"Returns a Mocked object on the patched attribute.\"\"\"\n mockfixture = self.useFixture(mockpatch.PatchObject(obj, attr))\n return mockfixture.mock\n","new_contents":"\n# Copyright (c) 2014 Mirantis.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or\n# implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport fixtures\nfrom oslotest import base\n\n\nclass TestCase(base.BaseTestCase):\n \"\"\"Test case base class for all unit tests.\"\"\"\n\n def patch(self, obj, attr):\n \"\"\"Returns a Mocked object on the patched attribute.\"\"\"\n mockfixture = self.useFixture(fixtures.MockPatchObject(obj, attr))\n return mockfixture.mock\n","subject":"Use fixtures instead of deprecated mockpatch module","message":"Use fixtures instead of deprecated mockpatch module\n\nThe mockpatch module of oslotest is deprecated since version 1.13\nand may be removed in version 2.0. Use fixtures.Mock* classes instead.\n\nChange-Id: I0ea834d41664efe84aa28ef2362467e2ad8b1928\n","lang":"Python","license":"apache-2.0","repos":"openstack\/python-blazarclient,ChameleonCloud\/python-blazarclient,stackforge\/python-blazarclient"} {"commit":"191d73fb6d30b691da8d9c55bfd36f055aea19d5","old_file":"backend\/pokehelper.py","new_file":"backend\/pokehelper.py","old_contents":"import json\nimport os\n\nclass Pokehelper(object):\n\n def __init__(self):\n\n basepath = os.path.dirname(__file__)\n filepath = os.path.abspath(os.path.join(basepath, 'data\/pokemon.json' ))\n with open(filepath) as pokejson:\n self.pokelist = json.load(pokejson)\n\n###\n### LIST STARTS AT 0, EVERY PokeNumber needs a -1\n###\n\n def get_pokefamily(self, pokemon_number):\n return self.pokelist[pokemon_number-1]['family']\n\n def get_evolution_name(self, pokemon_number):\n # NOT a safe method to use, just for testing purposes\n return self.pokelist[pokemon_number]['name']\n\n def get_pokename(self, poke_number):\n return self.pokelist[poke_number-1]['name']\n\n def get_base_attack(self, poke_number):\n return self.pokelist[poke_number-1]['stats']['attack']\n\n def get_base_defense(self, poke_number):\n return self.pokelist[poke_number-1]['stats']['defense']\n","new_contents":"import json\nimport os\n\nemptymon = {'moves1': [], 'family': 1, 'name': 'not-in-database', 'moves2': [], 'type2': 'nil', 'id': -1, 'candy': -1, 'type1': 'nil', 'stats': {'stamina': -1, 'attack': -1, 'defense': -1}}\n\n\nclass Pokehelper(object):\n\n def __init__(self):\n\n basepath = os.path.dirname(__file__)\n filepath = os.path.abspath(os.path.join(basepath, 'data\/pokemon.json'))\n with open(filepath) as pokejson:\n self.pokelist = json.load(pokejson)\n\n ###\n # LIST STARTS AT 0, EVERY PokeNumber needs a -1\n ###\n\n def get_pokefamily(self, poke_number):\n if (poke_number > 151):\n return emptymon['family']\n return self.pokelist[poke_number-1]['family']\n\n def get_evolution_name(self, poke_number):\n # NOT a safe method to use, just for testing purposes\n if (poke_number > 151):\n return emptymon['name']\n return self.pokelist[poke_number]['name']\n\n def get_pokename(self, poke_number):\n if (poke_number > 151):\n return emptymon['name']\n return self.pokelist[poke_number-1]['name']\n\n def get_base_attack(self, poke_number):\n if (poke_number > 151):\n return emptymon['stats']['attack']\n return self.pokelist[poke_number-1]['stats']['attack']\n\n def get_base_defense(self, poke_number):\n if (poke_number > 151):\n return emptymon['stats']['defense']\n return self.pokelist[poke_number-1]['stats']['defense']\n","subject":"Add fallback if pokemon_id > 151","message":"Add fallback if pokemon_id > 151\n","lang":"Python","license":"mit","repos":"Phaetec\/pogo-cruncher,Phaetec\/pogo-cruncher,Phaetec\/pogo-cruncher"} {"commit":"4b3e2289dbf20c0e2a7e0f83c7bd5963f2aa311f","old_file":"longshot.py","new_file":"longshot.py","old_contents":"\n\nHOME_URL = 'https:\/\/github.com\/ftobia\/longshot\/blob\/master\/longshot.py'\n\n\ndef upgrade():\n backup_self()\n download_and_overwrite()\n restart()\n\n\ndef backup_self():\n import shutil\n new_name = __file__ + '.bak'\n shutil.copy(__file__, new_name)\n\n\ndef download_and_overwrite():\n import urllib2\n response = urllib2.urlopen(HOME_URL)\n with open(__file__, 'w') as f:\n f.write(response.read())\n\n\ndef restart():\n import os\n os.execlp('python', __name__)\n\n\nif __name__ == '__main__':\n backup_self()\n download_and_overwrite()\n","new_contents":"\n\nHOME_URL = 'https:\/\/raw.githubusercontent.com\/ftobia\/longshot\/develop\/longshot.py'\n\n\ndef upgrade():\n backup_self()\n download_and_overwrite()\n restart()\n\n\ndef backup_self():\n import shutil\n new_name = __file__ + '.bak'\n shutil.copy(__file__, new_name)\n\n\ndef download_and_overwrite():\n import urllib2\n response = urllib2.urlopen(HOME_URL)\n with open(__file__, 'w') as f:\n f.write(response.read())\n\n\ndef restart():\n import os\n os.execlp('python', __name__)\n\n\nif __name__ == '__main__':\n backup_self()\n download_and_overwrite()\n","subject":"Use the right download URL.","message":"Use the right download URL.\n","lang":"Python","license":"bsd-3-clause","repos":"ftobia\/longshot"} {"commit":"7574528d839dc627ea53032b547e0e1c23a51f6b","old_file":"rdioexport\/_client\/__init__.py","new_file":"rdioexport\/_client\/__init__.py","old_contents":"import json\nfrom ._base import get_base_rdio_client\n\n\nclass _RdioExportClient(object):\n def __init__(self, base_client):\n self.base_client = base_client\n\n def get_current_user_key(self):\n return self.base_client.call('currentUser')['key']\n\n def get_collection_by_album(self, batch_size=100):\n current_user_key = self.get_current_user_key()\n\n start = 0\n result = []\n\n while True:\n batch = self.base_client.call(\n 'getAlbumsInCollection',\n user=current_user_key,\n sort='dateAdded',\n start=start,\n count=batch_size,\n extras=json.dumps([\n {'field': '*', 'exclude': True},\n {'field': 'key'},\n {'field': 'artist'},\n {'field': 'trackKeys'},\n ]),\n )\n\n for album in batch:\n yield album\n\n if (len(batch) < batch_size):\n break\n else:\n start += batch_size\n\n def get_album_data(self, album_key):\n return self.base_client.call(\n 'get',\n keys=album_key,\n extras=json.dumps([\n {'field': '*'},\n {\n 'field': 'track',\n 'extras': [\n {'field': '*'},\n ],\n },\n ]),\n )\n\n\ndef get_rdio_client():\n base_client = get_base_rdio_client()\n return _RdioExportClient(base_client)\n","new_contents":"import json\nfrom ._base import get_base_rdio_client\n\n\nclass _RdioExportClient(object):\n def __init__(self, base_client):\n self.base_client = base_client\n\n def get_current_user_key(self):\n return self.base_client.call('currentUser')['key']\n\n def get_collection_by_album(self, batch_size=100):\n current_user_key = self.get_current_user_key()\n\n start = 0\n result = []\n\n while True:\n batch = self.base_client.call(\n 'getAlbumsInCollection',\n user=current_user_key,\n sort='dateAdded',\n start=start,\n count=batch_size,\n extras=json.dumps([\n {'field': '*', 'exclude': True},\n {'field': 'key'},\n {'field': 'trackKeys'},\n ]),\n )\n\n for album in batch:\n yield album\n\n if (len(batch) < batch_size):\n break\n else:\n start += batch_size\n\n def get_album_data(self, album_key):\n return self.base_client.call(\n 'get',\n keys=album_key,\n extras=json.dumps([\n {'field': '*'},\n {\n 'field': 'track',\n 'extras': [\n {'field': '*'},\n ],\n },\n ]),\n )\n\n\ndef get_rdio_client():\n base_client = get_base_rdio_client()\n return _RdioExportClient(base_client)\n","subject":"Remove unused field from request.","message":"Remove unused field from request.\n","lang":"Python","license":"isc","repos":"alexhanson\/rdio-export"} {"commit":"e5d78dcfca7afffda7126e4e71944f40cdd9c272","old_file":"tests\/__init__.py","new_file":"tests\/__init__.py","old_contents":"#\n# For the license of this file, please consult the LICENSE file in the\n# root directory of this distribution.\n#\n\n# All tests in the test suite.\n__all__ = ( \"bitfield_tests\", \"zscii_tests\" )\n","new_contents":"#\n# For the license of this file, please consult the LICENSE file in the\n# root directory of this distribution.\n#\n\n# All tests in the test suite.\n__all__ = ( \"bitfield_tests\", \"zscii_tests\", \"lexer_tests\", \"glk_tests\" )\n","subject":"Make run_tests run all tests if no arguments are provided.","message":"Make run_tests run all tests if no arguments are provided.\n","lang":"Python","license":"bsd-3-clause","repos":"sussman\/zvm,sussman\/zvm"} {"commit":"65cd819b73c4a28b67a30b46b264b330d9967582","old_file":"flicks\/users\/forms.py","new_file":"flicks\/users\/forms.py","old_contents":"from django import forms\n\nfrom tower import ugettext_lazy as _lazy\n\nfrom flicks.base.util import country_choices\nfrom flicks.users.models import UserProfile\n\n\nclass UserProfileForm(forms.ModelForm):\n # L10n: Used in a choice field where users can choose between receiving\n # L10n: HTML-based or Text-only newsletter emails.\n NEWSLETTER_FORMATS = (('html', 'HTML'), ('text', _lazy('Text')))\n\n privacy_policy_agree = forms.BooleanField(required=True)\n mailing_list_signup = forms.BooleanField(required=False)\n mailing_list_format = forms.ChoiceField(required=False,\n choices=NEWSLETTER_FORMATS,\n initial='html')\n\n class Meta:\n model = UserProfile\n fields = ('full_name', 'nickname', 'country', 'address1', 'address2',\n 'city', 'mailing_country', 'state', 'postal_code')\n widgets = {\n 'full_name': forms.TextInput(attrs={'required': 'required'}),\n 'privacy_policy_agree': forms.CheckboxInput(\n attrs={'required': 'required'}),\n }\n\n def __init__(self, *args, **kwargs):\n super(UserProfileForm, self).__init__(*args, **kwargs)\n\n # Localize countries list\n self.fields['country'].choices = country_choices(allow_empty=False)\n self.fields['mailing_country'].choices = country_choices()\n","new_contents":"from django import forms\n\nfrom tower import ugettext_lazy as _lazy\n\nfrom flicks.base.util import country_choices\nfrom flicks.users.models import UserProfile\n\n\nclass UserProfileForm(forms.ModelForm):\n # L10n: Used in a choice field where users can choose between receiving\n # L10n: HTML-based or Text-only newsletter emails.\n NEWSLETTER_FORMATS = (('html', 'HTML'), ('text', _lazy('Text')))\n\n privacy_policy_agree = forms.BooleanField(\n required=True,\n widget=forms.CheckboxInput(attrs={'required': 'required'}))\n\n mailing_list_signup = forms.BooleanField(required=False)\n mailing_list_format = forms.ChoiceField(required=False,\n choices=NEWSLETTER_FORMATS,\n initial='html')\n\n class Meta:\n model = UserProfile\n fields = ('full_name', 'nickname', 'country', 'address1', 'address2',\n 'city', 'mailing_country', 'state', 'postal_code')\n widgets = {\n 'full_name': forms.TextInput(attrs={'required': 'required'}),\n }\n\n def __init__(self, *args, **kwargs):\n super(UserProfileForm, self).__init__(*args, **kwargs)\n\n # Localize countries list\n self.fields['country'].choices = country_choices(allow_empty=False)\n self.fields['mailing_country'].choices = country_choices()\n","subject":"Make privacy checkbox on user form required via required attribute.","message":"Make privacy checkbox on user form required via required attribute.","lang":"Python","license":"bsd-3-clause","repos":"mozilla\/firefox-flicks,mozilla\/firefox-flicks,mozilla\/firefox-flicks,mozilla\/firefox-flicks"} {"commit":"1e6ccfe615ee5d3e873e341a3d38553c3b07a3a0","old_file":"athumb\/validators.py","new_file":"athumb\/validators.py","old_contents":"from django.conf import settings\nfrom django.core.validators import ValidationError\n\n# A list of allowable thumbnail file extensions.\nALLOWABLE_THUMBNAIL_EXTENSIONS = getattr(\n settings, 'ALLOWABLE_THUMBNAIL_EXTENSIONS', ['png', 'jpg', 'jpeg', 'gif'])\n\nclass ImageUploadExtensionValidator(object):\n \"\"\"\n Perform some basic image uploading extension validation.\n \"\"\"\n compare = lambda self, a, b: a is not b\n clean = lambda self, x: x\n\n def __call__(self, value):\n filename = value.name\n filename_split = filename.split('.')\n extension = filename_split[-1]\n \n # Decided to require file extensions.\n if len(filename_split) < 2:\n raise ValidationError(\n \"Your file lacks an extension such as .jpg or .png. \"\n \"Please re-name it on your computer and re-upload it.\",\n code='no_extension'\n )\n\n # Restrict allowable extensions.\n if extension.lower() not in ALLOWABLE_THUMBNAIL_EXTENSIONS:\n # Format for your viewing pleasure.\n allowable_str = ' '.join(ALLOWABLE_THUMBNAIL_EXTENSIONS)\n raise ValidationError(\n \"Your file is not one of the allowable types: %s\" % allowable_str,\n code='extension_not_allowed'\n )","new_contents":"from django.conf import settings\nfrom django.core.validators import ValidationError\n\n# A list of allowable thumbnail file extensions.\nALLOWABLE_THUMBNAIL_EXTENSIONS = getattr(\n settings, 'ALLOWABLE_THUMBNAIL_EXTENSIONS', ['png', 'jpg', 'jpeg', 'gif'])\n\nclass ImageUploadExtensionValidator(object):\n \"\"\"\n Perform some basic image uploading extension validation.\n \"\"\"\n compare = lambda self, a, b: a is not b\n clean = lambda self, x: x\n\n\n def deconstruct(self):\n path = \"athumb.validators.ImageUploadExtensionValidator\"\n args = []\n kwargs = {}\n return (path, args, kwargs)\n\n\n def __call__(self, value):\n filename = value.name\n filename_split = filename.split('.')\n extension = filename_split[-1]\n \n # Decided to require file extensions.\n if len(filename_split) < 2:\n raise ValidationError(\n \"Your file lacks an extension such as .jpg or .png. \"\n \"Please re-name it on your computer and re-upload it.\",\n code='no_extension'\n )\n\n # Restrict allowable extensions.\n if extension.lower() not in ALLOWABLE_THUMBNAIL_EXTENSIONS:\n # Format for your viewing pleasure.\n allowable_str = ' '.join(ALLOWABLE_THUMBNAIL_EXTENSIONS)\n raise ValidationError(\n \"Your file is not one of the allowable types: %s\" % allowable_str,\n code='extension_not_allowed'\n )","subject":"Make ImageUploadExtensionValidator work with django 1.7 migrations","message":"Make ImageUploadExtensionValidator work with django 1.7 migrations\n","lang":"Python","license":"bsd-3-clause","repos":"ligonier\/django-athumb"} {"commit":"a5f3aa3d1098645f1b24539714bf22244141811c","old_file":"src\/pyprobe\/sensors\/system\/sensor_uptime.py","new_file":"src\/pyprobe\/sensors\/system\/sensor_uptime.py","old_contents":"# coding=utf-8\nfrom datetime import datetime\n\nimport psutil\n\nfrom pyprobe.sensors import *\n\n\n__author__ = 'Dirk Dittert'\n\n\nclass UptimeSensor(BaseSensor):\n KIND = u'uptime'\n\n def define(self, configuration):\n result = SensorDescription(u\"Laufzeit\", self.KIND)\n result.description = u\"Monitort die Laufzeit eines Geräts.\"\n return result\n\n def execute(self, sensorid, host, parameters, configuration):\n uptime = datetime.now() - datetime.fromtimestamp(psutil.get_boot_time())\n result = SensorResult(sensorid)\n channel = SensorChannel(u\"System-Laufzeit\", ModeType.FLOAT, ValueType.TIME_SECONDS, uptime.total_seconds())\n result.channel.append(channel)\n return result","new_contents":"# coding=utf-8\nfrom datetime import datetime\n\nimport psutil\n\nfrom pyprobe.sensors import *\n\n\n__author__ = 'Dirk Dittert'\n\n\nclass UptimeSensor(BaseSensor):\n KIND = u'uptime'\n\n def define(self, configuration):\n result = SensorDescription(u\"Laufzeit\", self.KIND)\n result.description = u\"Monitort die Laufzeit eines Geräts.\"\n return result\n\n def execute(self, sensorid, host, parameters, configuration):\n uptime = datetime.now() - datetime.fromtimestamp(psutil.boot_time())\n result = SensorResult(sensorid)\n channel = SensorChannel(u\"System-Laufzeit\", ModeType.FLOAT, ValueType.TIME_SECONDS, uptime.total_seconds())\n result.channel.append(channel)\n return result","subject":"Use proper method to determine system uptime.","message":"Use proper method to determine system uptime.\n\nFixes https:\/\/github.com\/dittert\/pyprobe\/issues\/9\n","lang":"Python","license":"apache-2.0","repos":"dittert\/pyprobe,dittert\/pyprobe"} {"commit":"bb575cfdf4a6781c878a12f80987fb3e62fe56d4","old_file":"chandl\/model\/posts.py","new_file":"chandl\/model\/posts.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\n\nclass Posts(list):\n \"\"\"\n Represents a list of posts in a thread.\n \"\"\"\n\n def __init__(self, *args):\n \"\"\"\n Initialise a new posts list.\n\n :param args: The list of posts.\n \"\"\"\n\n super(Posts, self).__init__(*args)\n\n def filter(self, predicate):\n \"\"\"\n Take a subset of this list of posts.\n\n :param predicate: The predicate to use to choose which posts make the\n cut.\n :return: The filtered posts.\n \"\"\"\n\n return Posts(filter(predicate, self))\n\n def map(self, transformation):\n \"\"\"\n Applies a transformation function to each post, returning a list of this\n function's returned values.\n\n :param transformation: The transformation function.\n :return: The transformed list of posts.\n \"\"\"\n\n return map(transformation, self)\n\n def foreach(self, function):\n \"\"\"\n Call a function for each post.\n\n :param function: A function taking a post argument. Return values are\n ignored.\n \"\"\"\n\n for post in self:\n function(post)\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\n\nclass Posts(list):\n \"\"\"\n Represents a list of posts in a thread.\n \"\"\"\n\n def __init__(self, *args):\n \"\"\"\n Initialise a new posts list.\n\n :param args: The list of posts.\n \"\"\"\n\n super(Posts, self).__init__(*args)\n\n def filter(self, predicate):\n \"\"\"\n Take a subset of this list of posts.\n\n :param predicate: The predicate to use to choose which posts make the\n cut.\n :return: The filtered posts.\n \"\"\"\n\n return Posts([post for post in self if predicate(post)])\n\n def map(self, transform):\n \"\"\"\n Applies a transformation function to each post, returning a list of this\n function's returned values.\n\n :param transform: The transformation function.\n :return: The transformed list of posts.\n \"\"\"\n\n return [transform(post) for post in self]\n\n def foreach(self, function):\n \"\"\"\n Call a function for each post.\n\n :param function: A function taking a post argument. Return values are\n ignored.\n \"\"\"\n\n for post in self:\n function(post)\n","subject":"Make post filtering and mapping more pythonic","message":"Make post filtering and mapping more pythonic\n","lang":"Python","license":"mit","repos":"gebn\/chandl,gebn\/chandl"} {"commit":"1fa7fed7d63fc7819ae5378f9a9addf7439e9b92","old_file":"messages.py","new_file":"messages.py","old_contents":"class Dispatched(object):\n\tdef __init__(self, w=None, cb=None):\n\t\tself.w = w\n\t\tself.cb = cb\n\t\tself.retval = None\n\t\tself.output = None\n\nclass DispatchInquiry(object):\n\tdef __init__(self, msg=None, payload=None, resp=None):\n\t\tself.msg = msg\n\t\tself.resp = resp\n\t\tself.payload = payload\n","new_contents":"class Dispatched(object):\n\tdef __init__(self, w=None, _id=None):\n\t\tself.w = w\n\t\tself.id = _id if _id != None else id(self)\n\nclass DispatchedState(object):\n\tdef __init__(self, retval=None, output=None, exc=None, _id=None):\n\t\tself.retval = retval\n\t\tself.output = output\n\t\tself.exc = exc\n\t\tself.id = _id if _id != None else id(self)\n\nclass DispatchInquiry(object):\n\tdef __init__(self, msg=None, payload=None, tgt=None, resp=None, _id=None):\n\t\tself.id = _id if _id != None else id(self)\n\t\tself.msg = msg\n\t\tself.tgt = tgt\n\t\tself.resp = resp\n\t\tself.payload = payload\n","subject":"Add DispatchedState, add target, add id","message":"Add DispatchedState, add target, add id\n","lang":"Python","license":"mit","repos":"joushou\/dispatch,joushou\/dispatch"} {"commit":"074dcf9c822827c6609dc11c0047aa79dd8c1ad8","old_file":"tests\/test_cli.py","new_file":"tests\/test_cli.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n\"\"\"Tests for `pyutrack` package.\"\"\"\n\n\nimport unittest\nfrom click.testing import CliRunner\n\nfrom pyutrack import cli\n\n\nclass TestYoutrack_cli(unittest.TestCase):\n\n def test_improt(self):\n import pyutrack\n\n def test_command_line_interface(self):\n runner = CliRunner()\n result = runner.invoke(cli.cli)\n assert result.exit_code == 0\n assert 'YouTrack' in result.output\n help_result = runner.invoke(cli.cli, ['--help'])\n assert help_result.exit_code == 0\n assert 'Show this message and exit.' in help_result.output\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\"\"\"Tests for `pyutrack` package.\"\"\"\n\nimport unittest\nfrom click.testing import CliRunner\n\nfrom pyutrack import cli\nfrom tests import PyutrackTest\n\n\nclass TestYoutrack_cli(PyutrackTest):\n def test_improt(self):\n import pyutrack\n\n def test_command_line_interface(self):\n runner = CliRunner()\n result = runner.invoke(cli.cli)\n assert result.exit_code == 0\n assert 'YouTrack' in result.output\n help_result = runner.invoke(cli.cli, ['--help'])\n assert help_result.exit_code == 0\n assert 'Show this message and exit.' in help_result.output\n","subject":"Set cli tests to base test class","message":"Set cli tests to base test class\n","lang":"Python","license":"mit","repos":"alisaifee\/pyutrack,alisaifee\/pyutrack"} {"commit":"5bd56894f3f4bbb3fadc3430238c394660b249c4","old_file":"core\/bootstrap.py","new_file":"core\/bootstrap.py","old_contents":"from __future__ import print_function\nfrom tinydb import TinyDB\nfrom core.models import Movie\n\nTABLE_POPULAR = \"popular\"\nTABLE_NAME_TO_ID = \"name_to_id_mapping\"\nTABLE_MOVIES = \"movies\"\n\nclass Application(object):\n def __init__(self, settings):\n database = TinyDB(settings[\"DATABASE\"])\n self.Movie = Movie(database, TABLE_MOVIES)\n self.settings = settings\n\n def setting(self, key):\n return self.settings[key]\n\n def debug(self, message):\n if self.settings.get(\"DEBUG\", False):\n print(message)\n\n def output(self, message):\n print(message)\n\n def debug_or_dot(self, message):\n if self.settings.get(\"DEBUG\", False):\n print(message)\n else:\n print(\".\", end=\"\")\n","new_contents":"from __future__ import print_function\nfrom tinydb import TinyDB\nfrom core.models import Movie\n\nTABLE_POPULAR = \"popular\"\nTABLE_NAME_TO_ID = \"name_to_id_mapping\"\nTABLE_MOVIES = \"movies\"\n\nclass Application(object):\n def __init__(self, settings):\n database = TinyDB(settings[\"DATABASE\"], indent=4)\n self.Movie = Movie(database, TABLE_MOVIES)\n self.settings = settings\n\n def setting(self, key):\n return self.settings[key]\n\n def debug(self, message):\n if self.settings.get(\"DEBUG\", False):\n print(message)\n\n def output(self, message):\n print(message)\n\n def debug_or_dot(self, message):\n if self.settings.get(\"DEBUG\", False):\n print(message)\n else:\n print(\".\", end=\"\")\n","subject":"Make the database JSON easier to read.","message":"Make the database JSON easier to read.\n","lang":"Python","license":"mit","repos":"EmilStenstrom\/nephele"} {"commit":"fc318739066a9310967f5ca9c261b670143f0fab","old_file":"codejail\/safe_exec.py","new_file":"codejail\/safe_exec.py","old_contents":"\"\"\"Safe execution of untrusted Python code.\"\"\"\n\nimport json\n\nfrom .lazymod import LazyModule\n\ndef straw(v):\n return json.loads(json.dumps(jsonable_dict(v)))\n\ndef jsonable_dict(d):\n jd = {}\n for k,v in d.iteritems():\n try:\n json.dumps(v)\n except TypeError:\n continue\n else:\n jd[k] = v\n return jd\n\ndef safe_exec(code, globals_dict, locals_dict=None, future_division=False, assumed_imports=None):\n if future_division:\n code = \"from __future__ import division\\n\" + code\n\n g_dict = straw(globals_dict)\n\n if locals_dict is None:\n l_dict = g_dict\n else:\n l_dict = straw(locals_dict)\n\n for modname in assumed_imports or ():\n if isinstance(modname, tuple):\n name, modname = modname\n else:\n name = modname\n g_dict[name] = LazyModule(modname)\n\n exec code in g_dict, l_dict\n\n globals_dict.update(straw(g_dict))\n if locals_dict is not None:\n locals_dict.update(straw(l_dict))\n","new_contents":"\"\"\"Safe execution of untrusted Python code.\"\"\"\n\nimport json\n\nfrom .lazymod import LazyModule\n\ndef straw(v):\n return json.loads(json.dumps(jsonable_dict(v)))\n\ndef jsonable_dict(d):\n jd = {}\n for k,v in d.iteritems():\n try:\n json.dumps(v)\n except TypeError:\n continue\n else:\n jd[k] = v\n return jd\n\ndef safe_exec(code, globals_dict, locals_dict=None, future_division=False, assumed_imports=None):\n \"\"\"Execute code safely.\n\n Returns None. The code can modify globals in `global_dict`.\n\n \"\"\"\n if future_division:\n code = \"from __future__ import division\\n\" + code\n\n g_dict = straw(globals_dict)\n\n if locals_dict is None:\n l_dict = g_dict\n else:\n l_dict = straw(locals_dict)\n\n for modname in assumed_imports or ():\n if isinstance(modname, tuple):\n name, modname = modname\n else:\n name = modname\n g_dict[name] = LazyModule(modname)\n\n exec code in g_dict, l_dict\n\n globals_dict.update(straw(g_dict))\n if locals_dict is not None:\n locals_dict.update(straw(l_dict))\n","subject":"Work in progress to sandbox the uses of eval in LMS.","message":"Work in progress to sandbox the uses of eval in LMS.\n","lang":"Python","license":"apache-2.0","repos":"edx\/codejail,StepicOrg\/codejail"} {"commit":"c9c0aace029dd07a96ceed4f14303d5f0eadee13","old_file":"blackjax\/__init__.py","new_file":"blackjax\/__init__.py","old_contents":"from .mcmc import hmc, nuts, rmh\nfrom .mcmc_adaptation import window_adaptation\n\n__version__ = \"0.3.0\"\n\n__all__ = [\n \"hmc\",\n \"nuts\",\n \"rmh\",\n \"window_adaptation\",\n \"adaptive_tempered_smc\",\n \"tempered_smc\",\n \"inference\",\n \"adaptation\",\n \"diagnostics\",\n]\n","new_contents":"from .diagnostics import effective_sample_size as ess\nfrom .diagnostics import potential_scale_reduction as rhat\nfrom .mcmc import hmc, nuts, rmh\nfrom .mcmc_adaptation import window_adaptation\n\n__version__ = \"0.3.0\"\n\n__all__ = [\n \"hmc\",\n \"nuts\",\n \"rmh\",\n \"window_adaptation\",\n \"adaptive_tempered_smc\",\n \"tempered_smc\",\n \"ess\",\n \"rhat\",\n]\n","subject":"Add diagnostics to blackjax namespace","message":"Add diagnostics to blackjax namespace\n","lang":"Python","license":"apache-2.0","repos":"blackjax-devs\/blackjax"} {"commit":"d3992b1677a5186b8b4072c9fdf50e4cb44dc5ef","old_file":"base_accounts\/models.py","new_file":"base_accounts\/models.py","old_contents":"from django.db import models\nfrom django.contrib.auth.models import AbstractUser\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.template.defaultfilters import slugify\n\n\nclass BaseUser(AbstractUser):\n slug = models.SlugField(_('slug'), max_length=255)\n name = models.CharField(_('name'), max_length=255, blank=True)\n first_login = models.BooleanField(_('first login'), default=True)\n image = models.ImageField(_('image'), blank=True, null=True, upload_to=\"images\/avatars\/%Y\/%m\/%d\", max_length=255)\n\n class Meta:\n abstract = True\n\n def save(self, *args, **kwargs):\n if not self.id:\n self.slug = slugify(self.username)\n if not self.name.strip():\n self.name = \"%s %s\" % (self.first_name, self.last_name)\n super(BaseUser, self).save(*args, **kwargs)\n\n def get_display_name(self):\n return self.name or self.username\n","new_contents":"from django.db import models\nfrom django.contrib.auth.models import AbstractUser\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.template.defaultfilters import slugify\n\n\nclass BaseUser(AbstractUser):\n slug = models.SlugField(_('slug'), max_length=255)\n name = models.CharField(_('name'), max_length=255, blank=True)\n first_login = models.BooleanField(_('first login'), default=True)\n image = models.ImageField(_('image'), blank=True, null=True, upload_to=\"images\/avatars\/%Y\/%m\/%d\", max_length=255)\n\n class Meta:\n abstract = True\n\n def save(self, *args, **kwargs):\n\n # Create slug from username. Altough field is not unique at database\n # level, it will be as long as username stays unique as well.\n if not self.id:\n self.slug = slugify(self.username)\n\n # Assign username as name if empty\n if not self.name.strip():\n if not self.first_name:\n self.first_name = self.username\n name = \"%s %s\" % (self.first_name, self.last_name)\n self.name = name.strip()\n\n super(BaseUser, self).save(*args, **kwargs)\n\n def get_display_name(self):\n return self.name or self.username\n","subject":"Fix name field for empty values","message":"Fix name field for empty values\n","lang":"Python","license":"bsd-3-clause","repos":"Nomadblue\/django-nomad-base-accounts,Nomadblue\/django-nomad-base-accounts"} {"commit":"e1bfa7170d4cf6a78cd0f2ca9c3d5302e04323f5","old_file":"utensils\/forms.py","new_file":"utensils\/forms.py","old_contents":"# encoding: utf-8\nfrom django import forms\n\n\nclass SearchForm(forms.Form):\n search = forms.CharField(\n label='', required=False,\n widget=forms.widgets.TextInput())\n","new_contents":"# encoding: utf-8\nfrom django import forms\nfrom django.utils.functional import curry\n\n\nclass SearchForm(forms.Form):\n search = forms.CharField(\n label='', required=False,\n widget=forms.widgets.TextInput())\n\n\nclass UniqueModelFieldsMixin(object):\n \"\"\"\n Mixin that enforces unique fields on ModelForm forms.\n\n Must be left of ModelForm when defining the form class (see\n https:\/\/code.djangoproject.com\/ticket\/13075).\n\n unique_fields = ['name', 'username']\n unique_fields = ['name', {'field': 'username', case_insensitive=True}]\n \"\"\"\n unique_fields = []\n\n def __init__(self, *args, **kwargs):\n super(UniqueModelFieldsMixin, self).__init__(*args, **kwargs)\n\n def _make_validator(field, case_insensitive):\n model = self.Meta.model\n value = self.cleaned_data.get(field)\n\n case = 'i' if case_insensitive else ''\n qs = model.objects.filter(\n **{field + '__{}exact'.format(case): value})\n if self.instance.pk:\n qs = qs.exclude(pk=self.instance.pk)\n\n if qs.exists():\n raise forms.ValidationError(\n \"That {} is not available.\".format(field))\n return value\n\n for field in self.unique_fields:\n if isinstance(field, dict):\n case_insensitive = field.get('case_insensitive', False)\n field_name = field['field']\n else:\n field_name = field\n case_insensitive = False\n func_name = \"clean_{}\".format(field_name)\n setattr(self, func_name,\n curry(_make_validator, field_name, case_insensitive))\n","subject":"Add unique model fields form mixin.","message":"Add unique model fields form mixin.\n","lang":"Python","license":"mit","repos":"code-kitchen\/django-utensils,code-kitchen\/django-utensils,code-kitchen\/django-utensils"} {"commit":"3d84e8e871b1049102815136ef23e3e630461918","old_file":"connman_dispatcher\/utils.py","new_file":"connman_dispatcher\/utils.py","old_contents":"import os\nimport subprocess\nimport logbook\nlogger = logbook.Logger('connman-dispatcher')\n\n\ndef execute_scripts_in_dirs(paths, state):\n for path in sorted(paths):\n if os.path.exists(path) and os.path.isdir(path):\n execute_scripts_in_dir(path, state)\n\ndef execute_scripts_in_dir(path, state):\n for script in sorted(os.listdir(path)):\n full_scirpt_path = os.path.join(path, script)\n if os.path.exists(full_scirpt_path):\n logger.info('executing: %s %s' % (full_scirpt_path, state))\n subprocess.Popen([full_scirpt_path, state])\n\n","new_contents":"import os\nimport subprocess\nimport logbook\nlogger = logbook.Logger('connman-dispatcher')\n\n\ndef is_executable(path):\n return all([os.path.isfile(path), os.access(path, os.X_OK)])\n\ndef execute_scripts_in_dirs(paths, state):\n for path in sorted(paths):\n if os.path.exists(path) and os.path.isdir(path):\n execute_scripts_in_dir(path, state)\n\ndef execute_scripts_in_dir(path, state):\n for script in sorted(os.listdir(path)):\n full_scirpt_path = os.path.join(path, script)\n if os.path.exists(full_scirpt_path):\n if is_executable(full_scirpt_path):\n logger.info('executing: %s %s' % (full_scirpt_path, state))\n subprocess.Popen([full_scirpt_path, state])\n else:\n logger.error('%s is not executable file' % full_scirpt_path)\n\n","subject":"Check if file is executable, before executing it","message":"Check if file is executable, before executing it\n","lang":"Python","license":"isc","repos":"a-sk\/connman-dispatcher"} {"commit":"7a901a8edd850dc5e2e75c89362444768722592c","old_file":"svs_interface.py","new_file":"svs_interface.py","old_contents":"#!\/usr\/bin\/env python\n\nimport subprocess\nfrom Tkinter import *\nfrom tkFileDialog import *\nimport os\n\nclass GpgApp(object):\n def __init__(self, master):\n frame = Frame(master)\n frame.pack()\n self.text = Text()\n self.text.pack()\n menu = Menu(master)\n root.config(menu=menu)\n\n filemenu = Menu(menu, tearoff=0)\n menu.add_cascade(label=\"File\", menu=filemenu)\n filemenu.add_command(label=\"Open\", command=self.filename_open)\n filemenu.add_separator()\n filemenu.add_command(label=\"Exit\", command=self.do_exit)\n def filename_open(self):\n fin = askopenfilenames()\n if fin:\n self.text.insert(END,fin)\n return fin\n def do_exit(self):\n root.destroy()\n\n\nroot = Tk()\nroot.title(\"a simple GnuPG interface\")\napp = GpgApp(root)\nroot.mainloop()\n\n\n","new_contents":"#!\/usr\/bin\/env python\n\nimport subprocess\nfrom Tkinter import *\nfrom tkFileDialog import *\nimport os\n\nGPG = 'gpg2'\nSERVER_KEY = '' # replace with gpg key ID of server key\n\nclass GpgApp(object):\n def __init__(self, master):\n frame = Frame(master)\n frame.pack()\n self.text = Text()\n self.text.pack()\n menu = Menu(master)\n root.config(menu=menu)\n\n filemenu = Menu(menu, tearoff=0)\n menu.add_cascade(label=\"File\", menu=filemenu)\n filemenu.add_command(label=\"Open\", command=self.filename_open)\n filemenu.add_separator()\n filemenu.add_command(label=\"Exit\", command=self.do_exit)\n def filename_open(self):\n fin = askopenfilenames()\n if fin:\n self.text.insert(END,fin)\n return fin\n def encrypt_file(self, input_file, output_file, recipient):\n args = [GPG, '--output', output_file, '--recipient', recipient, '-sea', input_file]\n subprocess.call(args)\n def do_exit(self):\n root.destroy()\n\n\nroot = Tk()\nroot.title(\"a simple GnuPG interface\")\napp = GpgApp(root)\nroot.mainloop()\n\n\n","subject":"Add method to encrypt files","message":"Add method to encrypt files\n","lang":"Python","license":"agpl-3.0","repos":"jeann2013\/securedrop,pwplus\/securedrop,chadmiller\/securedrop,jrosco\/securedrop,GabeIsman\/securedrop,GabeIsman\/securedrop,jaseg\/securedrop,chadmiller\/securedrop,jrosco\/securedrop,micahflee\/securedrop,jaseg\/securedrop,conorsch\/securedrop,kelcecil\/securedrop,chadmiller\/securedrop,pwplus\/securedrop,jaseg\/securedrop,ehartsuyker\/securedrop,ehartsuyker\/securedrop,micahflee\/securedrop,garrettr\/securedrop,harlo\/securedrop,pwplus\/securedrop,micahflee\/securedrop,harlo\/securedrop,jaseg\/securedrop,jrosco\/securedrop,ageis\/securedrop,jeann2013\/securedrop,jaseg\/securedrop,jeann2013\/securedrop,GabeIsman\/securedrop,heartsucker\/securedrop,harlo\/securedrop,kelcecil\/securedrop,pwplus\/securedrop,harlo\/securedrop,jrosco\/securedrop,kelcecil\/securedrop,ageis\/securedrop,ehartsuyker\/securedrop,ageis\/securedrop,conorsch\/securedrop,jrosco\/securedrop,micahflee\/securedrop,jeann2013\/securedrop,jeann2013\/securedrop,chadmiller\/securedrop,heartsucker\/securedrop,garrettr\/securedrop,kelcecil\/securedrop,kelcecil\/securedrop,harlo\/securedrop,garrettr\/securedrop,ehartsuyker\/securedrop,heartsucker\/securedrop,GabeIsman\/securedrop,heartsucker\/securedrop,GabeIsman\/securedrop,ehartsuyker\/securedrop,garrettr\/securedrop,pwplus\/securedrop,conorsch\/securedrop,chadmiller\/securedrop,heartsucker\/securedrop,GabeIsman\/securedrop,ehartsuyker\/securedrop,jrosco\/securedrop,conorsch\/securedrop,chadmiller\/securedrop,pwplus\/securedrop,kelcecil\/securedrop,conorsch\/securedrop,harlo\/securedrop,ageis\/securedrop,jeann2013\/securedrop,jaseg\/securedrop"} {"commit":"f3cbe52e0d65e8d6647815b25c79a836db93fb41","old_file":"gitcd\/Cli\/Command.py","new_file":"gitcd\/Cli\/Command.py","old_contents":"import subprocess\nimport string\n\nclass Command(object):\n\n def execute(self, command: str):\n cliArgs = self.parseCliArgs(command)\n\n process = subprocess.Popen(cliArgs, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n output, err = process.communicate()\n if process.returncode != 0:\n return False\n\n return output.decode(\"utf-8\").strip()\n\n def parseCliArgs(self, command: str):\n rawArgs = command.split(\" \")\n\n parsedArgs = []\n tmpString = False\n isSingle = False\n isDouble = False\n for arg in rawArgs:\n # handle strings in single quotes\n if arg.startswith(\"'\") and isSingle == False and isDouble == False:\n isSingle = True\n tmpString = arg\n\n elif arg.endswith(\"'\") and isSingle == True:\n arg = \"%s %s\" % (tmpString, arg)\n parsedArgs.append(arg)\n isSingle = False\n tmpString = False\n\n # handle strings in double quotes\n elif arg.startswith('\"') and isDouble == False and isSingle == False:\n isDouble = True\n tmpString = arg\n\n elif arg.endswith('\"') and isDouble == True:\n arg = \"%s %s\" % (tmpString, arg)\n parsedArgs.append(arg)\n isDouble = False\n tmpString = False\n\n # extend current string\n elif tmpString != False:\n tmpString = \"%s %s\" % (tmpString, arg)\n\n else:\n parsedArgs.append(arg)\n\n return parsedArgs","new_contents":"import subprocess\nimport string\n\nfrom pprint import pprint\n\nclass Command(object):\n\n def execute(self, command: str):\n cliArgs = self.parseCliArgs(command)\n\n pprint(cliArgs)\n\n process = subprocess.Popen(cliArgs, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE)\n output, err = process.communicate()\n if process.returncode != 0:\n return False\n\n return output.decode(\"utf-8\").strip()\n\n def parseCliArgs(self, command: str):\n rawArgs = command.split(\" \")\n\n parsedArgs = []\n tmpString = False\n isSingle = False\n isDouble = False\n for arg in rawArgs:\n # handle strings in single quotes\n if arg.startswith(\"'\") and isSingle == False and isDouble == False:\n isSingle = True\n tmpString = arg\n\n elif arg.endswith(\"'\") and isSingle == True:\n arg = \"%s %s\" % (tmpString, arg)\n parsedArgs.append(arg)\n isSingle = False\n tmpString = False\n\n # handle strings in double quotes\n elif arg.startswith('\"') and isDouble == False and isSingle == False:\n isDouble = True\n tmpString = arg\n\n elif arg.endswith('\"') and isDouble == True:\n arg = \"%s %s\" % (tmpString, arg)\n parsedArgs.append(arg)\n isDouble = False\n tmpString = False\n\n # extend current string\n elif tmpString != False:\n tmpString = \"%s %s\" % (tmpString, arg)\n\n else:\n parsedArgs.append(arg)\n\n return parsedArgs","subject":"Add some debug for debian box","message":"Add some debug for debian box\n","lang":"Python","license":"apache-2.0","repos":"claudio-walser\/gitcd,claudio-walser\/gitcd"} {"commit":"c72b712cf84e63dd2d72fdc6d64c50a65b8a88a0","old_file":"courant\/core\/search\/urls.py","new_file":"courant\/core\/search\/urls.py","old_contents":"from django.conf.urls.defaults import *\r\nfrom courant.core.search.views import *\r\n\r\nfrom haystack.forms import ModelSearchForm\r\nfrom haystack.query import SearchQuerySet\r\nfrom haystack.views import SearchView\r\n\r\nurlpatterns = patterns('',\r\n url(r'', CourantSearchView(template='search\/results_page.html',\r\n form_class=ModelSearchForm,\r\n searchqueryset=SearchQuerySet().all()), name=\"search\"),\r\n)\r\n","new_contents":"from django.conf.urls.defaults import *\r\nfrom courant.core.search.views import *\r\n\r\nfrom haystack.forms import ModelSearchForm\r\nfrom haystack.query import SearchQuerySet\r\nfrom haystack.views import SearchView\r\n\r\nurlpatterns = patterns('',\r\n url(r'', SearchView(template='search\/results_page.html',\r\n load_all=True,\r\n form_class=ModelSearchForm,\r\n searchqueryset=SearchQuerySet().all()), name=\"search\"),\r\n)\r\n","subject":"Remove all Haystack customization of search view pending further investigations.","message":"Remove all Haystack customization of search view pending further investigations.\n","lang":"Python","license":"bsd-3-clause","repos":"maxcutler\/Courant-News,maxcutler\/Courant-News"} {"commit":"1504710d748a86bbd4eed717b4bcc2f5d15ec1b7","old_file":"SatNOGS\/base\/management\/commands\/initialize.py","new_file":"SatNOGS\/base\/management\/commands\/initialize.py","old_contents":"from orbit import satellite\n\nfrom django.core.management.base import BaseCommand\n\nfrom base.tests import ObservationFactory, StationFactory\nfrom base.models import Satellite\n\n\nclass Command(BaseCommand):\n help = 'Create initial fixtures'\n\n def handle(self, *args, **options):\n ObservationFactory.create_batch(200)\n StationFactory.create_batch(200)\n\n satellites = Satellite.objects.all()\n\n for obj in satellites:\n try:\n sat = satellite(obj.norad_cat_id)\n except:\n self.stdout.write(('Satellite {} with Identifier {} does '\n 'not exist [deleted]').format(obj.name, obj.norad_cat_id))\n obj.delete()\n continue\n\n obj.name = sat.name()\n tle = sat.tle()\n obj.tle0 = tle[0]\n obj.tle1 = tle[1]\n obj.tle2 = tle[2]\n obj.save()\n self.stdout.write(('Satellite {} with Identifier {} '\n 'found [updated]').format(obj.norad_cat_id, obj.name))\n","new_contents":"from orbit import satellite\n\nfrom django.core.management.base import BaseCommand\n\nfrom base.tests import ObservationFactory, StationFactory\nfrom base.models import Satellite\n\n\nclass Command(BaseCommand):\n help = 'Create initial fixtures'\n\n def handle(self, *args, **options):\n ObservationFactory.create_batch(20)\n StationFactory.create_batch(20)\n\n satellites = Satellite.objects.all()\n\n for obj in satellites:\n try:\n sat = satellite(obj.norad_cat_id)\n except:\n self.stdout.write(('Satellite {} with Identifier {} does '\n 'not exist [deleted]').format(obj.name, obj.norad_cat_id))\n obj.delete()\n continue\n\n obj.name = sat.name()\n tle = sat.tle()\n obj.tle0 = tle[0]\n obj.tle1 = tle[1]\n obj.tle2 = tle[2]\n obj.save()\n self.stdout.write(('Satellite {} with Identifier {} '\n 'found [updated]').format(obj.norad_cat_id, obj.name))\n","subject":"Use more sane numbers for initial data","message":"Use more sane numbers for initial data\n","lang":"Python","license":"agpl-3.0","repos":"cshields\/satnogs-network,cshields\/satnogs-network,cshields\/satnogs-network,cshields\/satnogs-network"} {"commit":"d413345197abe9092979e324498c766f7410d34b","old_file":"bazaar\/goods\/utils.py","new_file":"bazaar\/goods\/utils.py","old_contents":"from __future__ import unicode_literals\n\nfrom .models import Product, PriceList\n\n\ndef create_product_for_good(good, price, quantity=1):\n \"\"\"\n Creates a product for the specified `good` with `quantity`. `price` is set to the default price list.\n Returns the new product instance\n \"\"\"\n\n product = Product.objects.create(name=good.name, description=good.description)\n product.save()\n\n # Add good to product elements list\n product.elements.create(good=good, quantity=quantity)\n\n # Set product's base price on default price list\n default_price_list = PriceList.objects.get_default()\n product.prices.create(product=product, price_list=default_price_list, price=price)\n\n return product\n","new_contents":"from __future__ import unicode_literals\n\nfrom .models import Product, PriceList\n\n\ndef create_product_for_good(good, price, quantity=1, name=None):\n \"\"\"\n Creates a product for the specified `good` with `quantity`. `price` is set to the default price list.\n Returns the new product instance\n \"\"\"\n product_name = name or good.name\n\n product = Product.objects.create(name=product_name, description=good.description)\n product.save()\n\n # Add good to product elements list\n product.elements.create(good=good, quantity=quantity)\n\n # Set product's base price on default price list\n default_price_list = PriceList.objects.get_default()\n product.prices.create(product=product, price_list=default_price_list, price=price)\n\n return product\n","subject":"Add name parameter to create_product_for_good which defaults to good.name","message":"Add name parameter to create_product_for_good which defaults to good.name\n","lang":"Python","license":"bsd-2-clause","repos":"evonove\/django-bazaar,evonove\/django-bazaar,meghabhoj\/NEWBAZAAR,evonove\/django-bazaar,meghabhoj\/NEWBAZAAR,meghabhoj\/NEWBAZAAR"} {"commit":"f21da23d45c328acffaba69a6f2fbf2056ca326b","old_file":"datapipe\/denoising\/__init__.py","new_file":"datapipe\/denoising\/__init__.py","old_contents":"#!\/usr\/bin\/env python3\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2016 Jérémie DECOCK (http:\/\/www.jdhp.org)\n\n# This script is provided under the terms and conditions of the MIT license:\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and\/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\n__all__ = ['abstract_cleaning_algorithm',\n 'fft',\n 'null',\n 'null_ref',\n 'tailcut',\n 'tailcut_jd',\n 'wavelets_mrfilter',\n 'wavelets_mrtransform']\n\n","new_contents":"#!\/usr\/bin\/env python3\n# -*- coding: utf-8 -*-\n\n# Copyright (c) 2016 Jérémie DECOCK (http:\/\/www.jdhp.org)\n\n# This script is provided under the terms and conditions of the MIT license:\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and\/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\n\n__all__ = ['abstract_cleaning_algorithm',\n 'fft',\n 'null',\n 'null_ref',\n 'tailcut',\n 'tailcut_jd',\n 'wavelets_mrfilter',\n 'wavelets_mrtransform',\n 'inverse_transform_sampling']\n\n","subject":"Add a module to the __all__ list.","message":"Add a module to the __all__ list.\n","lang":"Python","license":"mit","repos":"jdhp-sap\/sap-cta-data-pipeline,jdhp-sap\/sap-cta-data-pipeline,jdhp-sap\/data-pipeline-standalone-scripts,jdhp-sap\/data-pipeline-standalone-scripts"} {"commit":"c6a56604562460b2a8a72b25b23ce2ff4958d184","old_file":"test\/test_api.py","new_file":"test\/test_api.py","old_contents":"import unittest\nimport timevis\n\n\nclass TestAPIs(unittest.TestCase):\n def setUp(self):\n self.app = timevis.app.test_client()\n\n def test_api(self):\n resp = self.app.get('\/api\/v2\/experiment')\n self.assertIsNotNone(resp.data)\n\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"import unittest\nimport timevis\nimport os.path\nimport json\n\n\n# The folder holding the test data\ntest_path = os.path.dirname(__file__)\n\n\nclass TestExperiment(unittest.TestCase):\n def setUp(self):\n self.app = timevis.app.test_client()\n self.url = '\/api\/v2\/experiment'\n # TODO create test db\n timevis.models.init_db()\n\n def test_post(self):\n with open(test_path + '\/post_exp.json') as file:\n obj = json.load(file)\n resp = self.app.post(self.url, data=json.dumps(obj),\n content_type='application\/json')\n self.assertIsNotNone(resp.data)\n\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Add test case: Post method for experiment","message":"Add test case: Post method for experiment\n","lang":"Python","license":"mit","repos":"gaoce\/TimeVis,gaoce\/TimeVis,gaoce\/TimeVis"} {"commit":"c9b38972486b588790371ab41c961e68609e0b4b","old_file":"fabfile.py","new_file":"fabfile.py","old_contents":"from fabric.api import sudo, cd, env, run, local\n\nenv.hosts = ['ibadaw@sableamd2.cs.mcgill.ca']\n\nDEPLOY_DIR = '\/var\/www\/mcbench\/mcbench'\n\n\ndef deploy():\n with cd(DEPLOY_DIR):\n run('git pull origin master')\n restart()\n\n\ndef restart():\n sudo('service httpd restart')\n\n\ndef test():\n local('nosetests')\n\n\ndef coverage():\n nose_flags = [\n '--with-coverage',\n '--cover-html',\n '--cover-package=app,manage,mcbench'\n ]\n local('nosetests ' + ' '.join(nose_flags))\n","new_contents":"from fabric.api import sudo, cd, env, run, local\n\nenv.hosts = ['ibadaw@sableamd2.cs.mcgill.ca']\n\nDEPLOY_DIR = '\/var\/www\/mcbench\/mcbench'\n\n\ndef deploy():\n with cd(DEPLOY_DIR):\n run('git pull origin master')\n restart()\n\n\ndef restart():\n sudo('service httpd restart')\n\n\ndef test():\n local('nosetests')\n\n\ndef coverage():\n nose_flags = [\n '--with-coverage',\n '--cover-html',\n '--cover-package=app,manage,mcbench'\n ]\n local('nosetests ' + ' '.join(nose_flags))\n\n\ndef up():\n local('python manage.py runserver')\n","subject":"Add command to bring up dev server.","message":"Add command to bring up dev server.\n","lang":"Python","license":"mit","repos":"isbadawi\/mcbench,isbadawi\/mcbench"} {"commit":"0257d01e53a314b176f3a3b97259b46a271a08be","old_file":"tests\/test_tx.py","new_file":"tests\/test_tx.py","old_contents":"from __future__ import absolute_import, division, print_function\n\nimport pytest\n\npytest.importorskip(\"twisted\")\n\nfrom twisted.internet.defer import Deferred, succeed, fail\n\nfrom prometheus_async import tx\n\n\nclass TestTime(object):\n @pytest.inlineCallbacks\n def test_decorator(self, fo, patch_timer):\n \"\"\"\n time works with functions returning Deferreds.\n \"\"\"\n @tx.time(fo)\n def func():\n return succeed(42)\n\n rv = func()\n\n # Twisted runs fires callbacks immediately.\n assert [1] == fo._observed\n assert 42 == (yield rv)\n assert [1] == fo._observed\n\n @pytest.inlineCallbacks\n def test_decorator_exc(self, fo, patch_timer):\n \"\"\"\n Does not swallow exceptions.\n \"\"\"\n v = ValueError(\"foo\")\n\n @tx.time(fo)\n def func():\n return fail(v)\n\n with pytest.raises(ValueError) as e:\n yield func()\n\n assert v is e.value\n\n @pytest.inlineCallbacks\n def test_deferred(self, fo, patch_timer):\n \"\"\"\n time works with Deferreds.\n \"\"\"\n d = tx.time(fo, Deferred())\n\n assert [] == fo._observed\n\n d.callback(42)\n\n assert 42 == (yield d)\n assert [1] == fo._observed\n","new_contents":"from __future__ import absolute_import, division, print_function\n\nimport pytest\n\npytest.importorskip(\"twisted\")\n\nfrom twisted.internet.defer import Deferred, succeed, fail\n\nfrom prometheus_async import tx\n\n\nclass TestTime(object):\n @pytest.inlineCallbacks\n def test_decorator_sync(self, fo, patch_timer):\n \"\"\"\n time works with sync results functions.\n \"\"\"\n @tx.time(fo)\n def func():\n return 42\n\n assert 42 == (yield func())\n assert [1] == fo._observed\n\n @pytest.inlineCallbacks\n def test_decorator(self, fo, patch_timer):\n \"\"\"\n time works with functions returning Deferreds.\n \"\"\"\n @tx.time(fo)\n def func():\n return succeed(42)\n\n rv = func()\n\n # Twisted runs fires callbacks immediately.\n assert [1] == fo._observed\n assert 42 == (yield rv)\n assert [1] == fo._observed\n\n @pytest.inlineCallbacks\n def test_decorator_exc(self, fo, patch_timer):\n \"\"\"\n Does not swallow exceptions.\n \"\"\"\n v = ValueError(\"foo\")\n\n @tx.time(fo)\n def func():\n return fail(v)\n\n with pytest.raises(ValueError) as e:\n yield func()\n\n assert v is e.value\n\n @pytest.inlineCallbacks\n def test_deferred(self, fo, patch_timer):\n \"\"\"\n time works with Deferreds.\n \"\"\"\n d = tx.time(fo, Deferred())\n\n assert [] == fo._observed\n\n d.callback(42)\n\n assert 42 == (yield d)\n assert [1] == fo._observed\n","subject":"Test sync return for Twisted too","message":"Test sync return for Twisted too\n","lang":"Python","license":"apache-2.0","repos":"hynek\/prometheus_async"} {"commit":"7ef23761c64c1e1b1ac47c72a78d5109c36761d0","old_file":"tests\/testing.py","new_file":"tests\/testing.py","old_contents":"import os\nimport os.path\nimport subprocess\n\n\nclass HelloWorld(object):\n BUILD = r\"\"\"#!\/bin\/sh\nset -e\ncd $1\n\ncat > hello << EOF\n#!\/bin\/sh\necho Hello world!\nEOF\n\nchmod +x hello\n \"\"\"\n\n EXPECTED_OUTPUT = \"Hello world!\\n\"\n\ndef write_package_source(package_dir, scripts):\n whack_dir = os.path.join(package_dir, \"whack\")\n os.makedirs(whack_dir)\n for name, contents in scripts.iteritems():\n _write_script(os.path.join(whack_dir, name), contents)\n\ndef _write_script(path, contents):\n _write_file(path, contents)\n _make_executable(path)\n\ndef _make_executable(path):\n subprocess.check_call([\"chmod\", \"u+x\", path])\n\ndef _write_file(path, contents):\n open(path, \"w\").write(contents)\n\n","new_contents":"import os\nimport os.path\nimport subprocess\n\nfrom whack.files import write_file\n\n\nclass HelloWorld(object):\n BUILD = r\"\"\"#!\/bin\/sh\nset -e\ncd $1\n\ncat > hello << EOF\n#!\/bin\/sh\necho Hello world!\nEOF\n\nchmod +x hello\n \"\"\"\n\n EXPECTED_OUTPUT = \"Hello world!\\n\"\n\ndef write_package_source(package_dir, scripts):\n whack_dir = os.path.join(package_dir, \"whack\")\n os.makedirs(whack_dir)\n for name, contents in scripts.iteritems():\n _write_script(os.path.join(whack_dir, name), contents)\n\ndef _write_script(path, contents):\n write_file(path, contents)\n _make_executable(path)\n\ndef _make_executable(path):\n subprocess.check_call([\"chmod\", \"u+x\", path])\n","subject":"Remove duplicate definition of write_file","message":"Remove duplicate definition of write_file\n","lang":"Python","license":"bsd-2-clause","repos":"mwilliamson\/whack"} {"commit":"d8444cec60f38baa75b89892dda6163bf63917af","old_file":"todo\/__init__.py","new_file":"todo\/__init__.py","old_contents":"\"\"\"django todo\"\"\"\n__version__ = '1.5.dev'\n\n__author__ = 'Scot Hacker'\n__email__ = 'shacker@birdhouse.org'\n\n__url__ = 'https:\/\/github.com\/shacker\/django-todo'\n__license__ = 'BSD License'\n","new_contents":"\"\"\"django todo\"\"\"\n__version__ = '1.5'\n\n__author__ = 'Scot Hacker'\n__email__ = 'shacker@birdhouse.org'\n\n__url__ = 'https:\/\/github.com\/shacker\/django-todo'\n__license__ = 'BSD License'\n","subject":"Bump version number for release 1.5","message":"Bump version number for release 1.5\n","lang":"Python","license":"bsd-3-clause","repos":"jwiltshire\/django-todo,shacker\/django-todo,jwiltshire\/django-todo,shacker\/django-todo,jwiltshire\/django-todo,shacker\/django-todo"} {"commit":"f5728e24ba6dec2d2d7c2eff7888137e91469094","old_file":"overlay\/Data.py","new_file":"overlay\/Data.py","old_contents":"import time\n\n\nclass Data:\n def __init__(self, secs_since_epoch, depth_chart, temperature_chart, frame_path):\n # general settings\n self.width = 1296\n self.height = 972\n self.padding = 5\n self.frame_path = frame_path\n\n # date\/time settings\n self.time = time.localtime(secs_since_epoch)\n self.frame_date = time.strftime(\"%B %d, %Y\", self.time)\n self.frame_time = time.strftime(\"%I:%M:%S %p\", self.time)\n self.font_size = 22\n self.text_color = \"rgb(255,255,255)\"\n \n # charts\n self.depth_chart = depth_chart.to_svg()\n self.temperature_chart = temperature_chart.to_svg()\n\n @property\n def datetime_x(self):\n return self.width - self.padding\n\n @property\n def depth_background_y(self):\n return self.height - 3 * self.padding - self.depth_graph_height\n\n @property\n def depth_background_width(self):\n return self.depth_graph_width + 2 * self.padding\n\n @property\n def depth_background_height(self):\n return self.depth_graph_height + 2 * self.padding\n\n @property\n def depth_text_x(self):\n return self.depth_background_width * 0.5\n\n @property\n def depth_text_y(self):\n return self.depth_background_height - self.padding\n","new_contents":"import time\n\n\nclass Data:\n def __init__(self, secs_since_epoch, depth_chart, temperature_chart, frame_path):\n # general settings\n self.width = 1296\n self.height = 972\n self.padding = 5\n self.frame_path = frame_path\n\n # date\/time settings\n local_time = time.localtime(secs_since_epoch)\n self.frame_date = time.strftime(\"%B %d, %Y\", local_time)\n self.frame_time = time.strftime(\"%I:%M:%S %p\", local_time)\n self.font_size = 22\n self.text_color = \"rgb(255,255,255)\"\n self.datetime_x = self.width - self.padding\n \n # charts\n self.depth_chart = depth_chart.to_svg()\n self.temperature_chart = temperature_chart.to_svg()\n","subject":"Remove unneeded properties from main data object","message":"Remove unneeded properties from main data object\n","lang":"Python","license":"mit","repos":"thelonious\/g2x,gizmo-cda\/g2x,gizmo-cda\/g2x,gizmo-cda\/g2x,gizmo-cda\/g2x,thelonious\/g2x"} {"commit":"29f3bb4fc549f78771294f90f5168b20f9ea7b5e","old_file":"sdi\/corestick.py","new_file":"sdi\/corestick.py","old_contents":"def read(filename):\n \"\"\"\n Reads in a corestick file and returns a dictionary keyed by core_id.\n Layer interface depths are positive and are relative to the lake bottom.\n depths are returned in meters. Northing and Easting are typically in the\n coordinate system used in the rest of the lake survey. We ignore the display\n related color and width fields in the file.\n \"\"\"\n\n cores = {}\n with open(filename) as f:\n units = f.readline().strip('\\r\\n').lower()\n\n if units not in ['feet', 'meters', 'meter']:\n raise NotImplementedError('Only units of FEET and METERS\/METER are supported ')\n\n conv_factor = 1.0\n\n if units == 'feet':\n conv_factor = 0.3048\n\n f.readline()\n for line in f.readlines():\n fields = line.split()\n core_id = fields[2]\n data = {}\n data['easting'] = float(fields[0])\n data['northing'] = float(fields[1])\n data['layer_interface_depths'] = [\n float(fields[i]) * conv_factor\n for i in range(5, len(fields), 4)\n ]\n cores[core_id] = data\n\n return cores\n","new_contents":"def read(filename):\n \"\"\"\n Reads in a corestick file and returns a dictionary keyed by core_id.\n Layer interface depths are positive and are relative to the lake bottom.\n depths are returned in meters. Northing and Easting are typically in the\n coordinate system used in the rest of the lake survey. We ignore the\n width fields in the file.\n \"\"\"\n\n cores = {}\n with open(filename) as f:\n units = f.readline().strip('\\r\\n').lower()\n\n if units not in ['feet', 'meters', 'meter']:\n raise NotImplementedError('Only units of FEET and METERS\/METER are supported ')\n\n conv_factor = 1.0\n\n if units == 'feet':\n conv_factor = 0.3048\n\n f.readline()\n for line in f.readlines():\n fields = line.split()\n core_id = fields[2]\n data = {}\n data['easting'] = float(fields[0])\n data['northing'] = float(fields[1])\n data['layer_interface_depths'] = [\n float(fields[i]) * conv_factor\n for i in range(5, len(fields), 4)\n ]\n data['layer_colors'] = [i for i in range(6, len(fields), 4)]\n cores[core_id] = data\n\n return cores\n","subject":"Modify to read layer colors.","message":"Modify to read layer colors.\n","lang":"Python","license":"bsd-3-clause","repos":"twdb\/sdi"} {"commit":"51aaf9c19f92db4b3ad5d7aa646bf6ef8d5e62a6","old_file":"runtests.py","new_file":"runtests.py","old_contents":"#!\/usr\/bin\/env python\nimport sys\nfrom optparse import OptionParser\n\nfrom django.conf import settings\n\nif not settings.configured:\n settings.configure(\n DATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': 'django_celery_rpc',\n 'USER': '',\n 'PASSWORD': '',\n 'HOST': '',\n 'PORT': '',\n }\n },\n\n INSTALLED_APPS=[\n 'django.contrib.contenttypes',\n 'celery_rpc',\n 'celery_rpc.tests',\n ],\n ROOT_URLCONF='',\n DEBUG=False,\n )\n\n\nfrom django_nose import NoseTestSuiteRunner\n\n\ndef runtests(*test_args, **kwargs):\n if 'south' in settings.INSTALLED_APPS:\n from south.management.commands import patch_for_test_db_setup\n patch_for_test_db_setup()\n\n if not test_args:\n test_args = ['celery_rpc']\n\n test_runner = NoseTestSuiteRunner(**kwargs)\n\n failures = test_runner.run_tests(test_args)\n sys.exit(failures)\n\nif __name__ == '__main__':\n parser = OptionParser()\n parser.add_option('--verbosity', dest='verbosity', action='store',\n default=1, type=int)\n parser.add_options(NoseTestSuiteRunner.options)\n (options, args) = parser.parse_args()\n\n runtests(*args, **options.__dict__)\n","new_contents":"#!\/usr\/bin\/env python\nimport sys\nfrom optparse import OptionParser\n\nfrom django.conf import settings\n\nif not settings.configured:\n settings.configure(\n DATABASES = {\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': ':memory:',\n 'USER': '',\n 'PASSWORD': '',\n 'HOST': '',\n 'PORT': '',\n }\n },\n\n INSTALLED_APPS=[\n 'django.contrib.contenttypes',\n 'celery_rpc',\n 'celery_rpc.tests',\n ],\n ROOT_URLCONF='',\n DEBUG=True,\n\n CELERY_RPC_CONFIG = {\n 'CELERY_ALWAYS_EAGER': True\n },\n )\n\n\nfrom django_nose import NoseTestSuiteRunner\n\n\ndef runtests(*test_args, **kwargs):\n if 'south' in settings.INSTALLED_APPS:\n from south.management.commands import patch_for_test_db_setup\n patch_for_test_db_setup()\n\n if not test_args:\n test_args = ['celery_rpc']\n\n test_runner = NoseTestSuiteRunner(**kwargs)\n\n failures = test_runner.run_tests(test_args)\n sys.exit(failures)\n\nif __name__ == '__main__':\n parser = OptionParser()\n parser.add_option('--verbosity', dest='verbosity', action='store',\n default=1, type=int)\n parser.add_options(NoseTestSuiteRunner.options)\n (options, args) = parser.parse_args()\n\n runtests(*args, **options.__dict__)\n","subject":"Fix run tests with Celery","message":"Fix run tests with Celery\n","lang":"Python","license":"unlicense","repos":"bourivouh\/django-celery-rpc,ttyS15\/django-celery-rpc,tumb1er\/django-celery-rpc"} {"commit":"6290d72458a470947c5da651968cf20f8714c646","old_file":"fancypages\/contrib\/oscar_fancypages\/views.py","new_file":"fancypages\/contrib\/oscar_fancypages\/views.py","old_contents":"from oscar.core.loading import load_class\n\nfrom . import mixins\n\n\nProductCategoryView = load_class('catalogue.views', 'ProductCategoryView')\n\n\nclass FancyPageDetailView(mixins.OscarFancyPageMixin, ProductCategoryView):\n pass\n","new_contents":"from oscar.core.loading import get_class\n\nfrom . import mixins\n\n\nProductCategoryView = get_class('catalogue.views', 'ProductCategoryView')\n\n\nclass FancyPageDetailView(mixins.OscarFancyPageMixin, ProductCategoryView):\n pass\n","subject":"Fix class loader called in Oscar contrib package","message":"Fix class loader called in Oscar contrib package\n","lang":"Python","license":"bsd-3-clause","repos":"socradev\/django-fancypages,socradev\/django-fancypages,socradev\/django-fancypages,tangentlabs\/django-fancypages,tangentlabs\/django-fancypages,tangentlabs\/django-fancypages"} {"commit":"9f5ed14f24aecdd46699e84e13e9fa1f90cbf793","old_file":"script\/lib\/config.py","new_file":"script\/lib\/config.py","old_contents":"#!\/usr\/bin\/env python\n\nimport platform\nimport sys\n\nBASE_URL = 'http:\/\/gh-contractor-zcbenz.s3.amazonaws.com\/libchromiumcontent'\nLIBCHROMIUMCONTENT_COMMIT = '17a0e24666d0198810752284690bc2d0d87094d7'\n\nARCH = {\n 'cygwin': '32bit',\n 'darwin': '64bit',\n 'linux2': platform.architecture()[0],\n 'win32': '32bit',\n}[sys.platform]\nDIST_ARCH = {\n '32bit': 'ia32',\n '64bit': 'x64',\n}[ARCH]\n\nTARGET_PLATFORM = {\n 'cygwin': 'win32',\n 'darwin': 'darwin',\n 'linux2': 'linux',\n 'win32': 'win32',\n}[sys.platform]\n\nverbose_mode = False\n\ndef enable_verbose_mode():\n print 'Running in verbose mode'\n global verbose_mode\n verbose_mode = True\n\ndef is_verbose_mode():\n return verbose_mode\n","new_contents":"#!\/usr\/bin\/env python\n\nimport platform\nimport sys\n\nBASE_URL = 'http:\/\/gh-contractor-zcbenz.s3.amazonaws.com\/libchromiumcontent'\nLIBCHROMIUMCONTENT_COMMIT = '6300862b4b16bd171f00ae566b697098c29743f7'\n\nARCH = {\n 'cygwin': '32bit',\n 'darwin': '64bit',\n 'linux2': platform.architecture()[0],\n 'win32': '32bit',\n}[sys.platform]\nDIST_ARCH = {\n '32bit': 'ia32',\n '64bit': 'x64',\n}[ARCH]\n\nTARGET_PLATFORM = {\n 'cygwin': 'win32',\n 'darwin': 'darwin',\n 'linux2': 'linux',\n 'win32': 'win32',\n}[sys.platform]\n\nverbose_mode = False\n\ndef enable_verbose_mode():\n print 'Running in verbose mode'\n global verbose_mode\n verbose_mode = True\n\ndef is_verbose_mode():\n return verbose_mode\n","subject":"Upgrade libchromiumcontent to fix linking error","message":"mac: Upgrade libchromiumcontent to fix linking error\n","lang":"Python","license":"mit","repos":"brave\/muon,chriskdon\/electron,Gerhut\/electron,twolfson\/electron,icattlecoder\/electron,the-ress\/electron,bruce\/electron,soulteary\/electron,nekuz0r\/electron,tylergibson\/electron,Zagorakiss\/electron,tomashanacek\/electron,chrisswk\/electron,destan\/electron,roadev\/electron,simongregory\/electron,gbn972\/electron,jiaz\/electron,leolujuyi\/electron,seanchas116\/electron,baiwyc119\/electron,GoooIce\/electron,arturts\/electron,bruce\/electron,rhencke\/electron,nicholasess\/electron,tincan24\/electron,John-Lin\/electron,kcrt\/electron,seanchas116\/electron,mirrh\/electron,destan\/electron,fffej\/electron,bwiggs\/electron,chriskdon\/electron,jaanus\/electron,voidbridge\/electron,webmechanicx\/electron,joaomoreno\/atom-shell,Andrey-Pavlov\/electron,wolfflow\/electron,mirrh\/electron,mattdesl\/electron,pandoraui\/electron,Neron-X5\/electron,bbondy\/electron,sircharleswatson\/electron,tinydew4\/electron,jacksondc\/electron,IonicaBizauKitchen\/electron,GoooIce\/electron,seanchas116\/electron,jaanus\/electron,webmechanicx\/electron,JesselJohn\/electron,simonfork\/electron,wolfflow\/electron,nekuz0r\/electron,fffej\/electron,dahal\/electron,simongregory\/electron,fritx\/electron,bitemyapp\/electron,rreimann\/electron,MaxGraey\/electron,kazupon\/electron,sshiting\/electron,shiftkey\/electron,coderhaoxin\/electron,leftstick\/electron,lzpfmh\/electron,stevekinney\/electron,jacksondc\/electron,dongjoon-hyun\/electron,felixrieseberg\/electron,preco21\/electron,nicholasess\/electron,kikong\/electron,synaptek\/electron,jsutcodes\/electron,arusakov\/electron,trigrass2\/electron,kokdemo\/electron,setzer777\/electron,christian-bromann\/electron,voidbridge\/electron,gstack\/infinium-shell,timruffles\/electron,faizalpribadi\/electron,jaanus\/electron,aichingm\/electron,brave\/electron,shockone\/electron,egoist\/electron,michaelchiche\/electron,hokein\/atom-shell,minggo\/electron,noikiy\/electron,bwiggs\/electron,MaxWhere\/electron,LadyNaggaga\/electron,aliib\/electron,synaptek\/electron,joneit\/electron,Jonekee\/electron,rajatsingla28\/electron,iftekeriba\/electron,jsutcodes\/electron,dahal\/electron,stevemao\/electron,benweissmann\/electron,kokdemo\/electron,stevekinney\/electron,abhishekgahlot\/electron,John-Lin\/electron,Faiz7412\/electron,fomojola\/electron,hokein\/atom-shell,astoilkov\/electron,mrwizard82d1\/electron,kostia\/electron,matiasinsaurralde\/electron,maxogden\/atom-shell,Jacobichou\/electron,neutrous\/electron,d-salas\/electron,joaomoreno\/atom-shell,astoilkov\/electron,Ivshti\/electron,xiruibing\/electron,mattotodd\/electron,stevemao\/electron,tomashanacek\/electron,nicobot\/electron,anko\/electron,bwiggs\/electron,setzer777\/electron,RobertJGabriel\/electron,smczk\/electron,MaxWhere\/electron,digideskio\/electron,jlhbaseball15\/electron,jaanus\/electron,biblerule\/UMCTelnetHub,beni55\/electron,gerhardberger\/electron,meowlab\/electron,zhakui\/electron,jcblw\/electron,deed02392\/electron,renaesop\/electron,wolfflow\/electron,leethomas\/electron,adamjgray\/electron,Floato\/electron,leftstick\/electron,eric-seekas\/electron,rajatsingla28\/electron,BionicClick\/electron,leethomas\/electron,kokdemo\/electron,gstack\/infinium-shell,MaxWhere\/electron,coderhaoxin\/electron,saronwei\/electron,robinvandernoord\/electron,xfstudio\/electron,darwin\/electron,aichingm\/electron,vipulroxx\/electron,jonatasfreitasv\/electron,d-salas\/electron,LadyNaggaga\/electron,chriskdon\/electron,pirafrank\/electron,Gerhut\/electron,fritx\/electron,mubassirhayat\/electron,edulan\/electron,egoist\/electron,minggo\/electron,brave\/electron,minggo\/electron,jsutcodes\/electron,yalexx\/electron,kazupon\/electron,jsutcodes\/electron,bright-sparks\/electron,mrwizard82d1\/electron,greyhwndz\/electron,IonicaBizauKitchen\/electron,MaxWhere\/electron,gstack\/infinium-shell,vipulroxx\/electron,tincan24\/electron,leftstick\/electron,howmuchcomputer\/electron,BionicClick\/electron,faizalpribadi\/electron,voidbridge\/electron,adcentury\/electron,edulan\/electron,renaesop\/electron,synaptek\/electron,noikiy\/electron,stevekinney\/electron,jannishuebl\/electron,minggo\/electron,eriser\/electron,oiledCode\/electron,leolujuyi\/electron,nicobot\/electron,eriser\/electron,egoist\/electron,Jonekee\/electron,farmisen\/electron,yalexx\/electron,sshiting\/electron,biblerule\/UMCTelnetHub,aecca\/electron,shiftkey\/electron,leftstick\/electron,mhkeller\/electron,egoist\/electron,RIAEvangelist\/electron,mjaniszew\/electron,aecca\/electron,d-salas\/electron,brave\/electron,vaginessa\/electron,icattlecoder\/electron,bbondy\/electron,destan\/electron,xfstudio\/electron,aliib\/electron,bbondy\/electron,joaomoreno\/atom-shell,gamedevsam\/electron,jjz\/electron,gabrielPeart\/electron,posix4e\/electron,wan-qy\/electron,lrlna\/electron,robinvandernoord\/electron,kostia\/electron,DivyaKMenon\/electron,robinvandernoord\/electron,bpasero\/electron,zhakui\/electron,kazupon\/electron,simongregory\/electron,astoilkov\/electron,tinydew4\/electron,SufianHassan\/electron,michaelchiche\/electron,gbn972\/electron,gerhardberger\/electron,arturts\/electron,digideskio\/electron,cqqccqc\/electron,posix4e\/electron,MaxGraey\/electron,Neron-X5\/electron,joneit\/electron,dkfiresky\/electron,Ivshti\/electron,thingsinjars\/electron,yalexx\/electron,Evercoder\/electron,tonyganch\/electron,wan-qy\/electron,pandoraui\/electron,thomsonreuters\/electron,rhencke\/electron,cos2004\/electron,xfstudio\/electron,fireball-x\/atom-shell,jlord\/electron,kostia\/electron,fireball-x\/atom-shell,shiftkey\/electron,wan-qy\/electron,arusakov\/electron,medixdev\/electron,kostia\/electron,deepak1556\/atom-shell,neutrous\/electron,bobwol\/electron,gerhardberger\/electron,lrlna\/electron,jjz\/electron,synaptek\/electron,systembugtj\/electron,the-ress\/electron,mirrh\/electron,sky7sea\/electron,bpasero\/electron,mhkeller\/electron,faizalpribadi\/electron,wolfflow\/electron,takashi\/electron,natgolov\/electron,kenmozi\/electron,jacksondc\/electron,digideskio\/electron,brenca\/electron,vHanda\/electron,Jonekee\/electron,jlord\/electron,Andrey-Pavlov\/electron,dongjoon-hyun\/electron,beni55\/electron,Rokt33r\/electron,aaron-goshine\/electron,icattlecoder\/electron,felixrieseberg\/electron,soulteary\/electron,simonfork\/electron,stevemao\/electron,coderhaoxin\/electron,adcentury\/electron,rsvip\/electron,mjaniszew\/electron,oiledCode\/electron,gamedevsam\/electron,d-salas\/electron,timruffles\/electron,yan-foto\/electron,natgolov\/electron,egoist\/electron,farmisen\/electron,lzpfmh\/electron,bitemyapp\/electron,twolfson\/electron,posix4e\/electron,davazp\/electron,mattotodd\/electron,biblerule\/UMCTelnetHub,matiasinsaurralde\/electron,jannishuebl\/electron,chriskdon\/electron,anko\/electron,SufianHassan\/electron,davazp\/electron,joneit\/electron,SufianHassan\/electron,natgolov\/electron,ianscrivener\/electron,webmechanicx\/electron,nekuz0r\/electron,christian-bromann\/electron,miniak\/electron,yalexx\/electron,ianscrivener\/electron,fireball-x\/atom-shell,synaptek\/electron,neutrous\/electron,lrlna\/electron,etiktin\/electron,mubassirhayat\/electron,sky7sea\/electron,jtburke\/electron,d-salas\/electron,mhkeller\/electron,kokdemo\/electron,electron\/electron,subblue\/electron,destan\/electron,leethomas\/electron,eric-seekas\/electron,deed02392\/electron,yan-foto\/electron,astoilkov\/electron,Zagorakiss\/electron,zhakui\/electron,smczk\/electron,beni55\/electron,meowlab\/electron,bbondy\/electron,GoooIce\/electron,hokein\/atom-shell,bobwol\/electron,arusakov\/electron,aaron-goshine\/electron,mattotodd\/electron,shennushi\/electron,bright-sparks\/electron,MaxGraey\/electron,shaundunne\/electron,ervinb\/electron,gstack\/infinium-shell,Jacobichou\/electron,aliib\/electron,bitemyapp\/electron,carsonmcdonald\/electron,sky7sea\/electron,JesselJohn\/electron,sircharleswatson\/electron,jonatasfreitasv\/electron,fffej\/electron,jonatasfreitasv\/electron,mirrh\/electron,MaxGraey\/electron,ankitaggarwal011\/electron,RobertJGabriel\/electron,robinvandernoord\/electron,Floato\/electron,xfstudio\/electron,rreimann\/electron,thomsonreuters\/electron,eriser\/electron,coderhaoxin\/electron,deed02392\/electron,xiruibing\/electron,aaron-goshine\/electron,lzpfmh\/electron,edulan\/electron,eriser\/electron,edulan\/electron,mjaniszew\/electron,baiwyc119\/electron,rreimann\/electron,simongregory\/electron,fritx\/electron,jlhbaseball15\/electron,cqqccqc\/electron,fffej\/electron,zhakui\/electron,gamedevsam\/electron,evgenyzinoviev\/electron,Zagorakiss\/electron,deepak1556\/atom-shell,aichingm\/electron,gabrielPeart\/electron,aaron-goshine\/electron,natgolov\/electron,gamedevsam\/electron,brave\/electron,arturts\/electron,stevekinney\/electron,aichingm\/electron,sshiting\/electron,zhakui\/electron,pombredanne\/electron,Rokt33r\/electron,thompsonemerson\/electron,jlhbaseball15\/electron,leethomas\/electron,maxogden\/atom-shell,ankitaggarwal011\/electron,neutrous\/electron,nicholasess\/electron,thompsonemerson\/electron,darwin\/electron,subblue\/electron,jiaz\/electron,trankmichael\/electron,joaomoreno\/atom-shell,icattlecoder\/electron,ianscrivener\/electron,bitemyapp\/electron,adamjgray\/electron,ankitaggarwal011\/electron,etiktin\/electron,brenca\/electron,jiaz\/electron,jlord\/electron,seanchas116\/electron,electron\/electron,ankitaggarwal011\/electron,jjz\/electron,subblue\/electron,jonatasfreitasv\/electron,greyhwndz\/electron,anko\/electron,takashi\/electron,voidbridge\/electron,fffej\/electron,bruce\/electron,arusakov\/electron,GoooIce\/electron,adcentury\/electron,egoist\/electron,micalan\/electron,JussMee15\/electron,roadev\/electron,aichingm\/electron,michaelchiche\/electron,bitemyapp\/electron,darwin\/electron,xfstudio\/electron,SufianHassan\/electron,Neron-X5\/electron,Evercoder\/electron,bruce\/electron,trankmichael\/electron,evgenyzinoviev\/electron,chrisswk\/electron,lzpfmh\/electron,rreimann\/electron,farmisen\/electron,thomsonreuters\/electron,jcblw\/electron,Floato\/electron,destan\/electron,setzer777\/electron,lrlna\/electron,pandoraui\/electron,tylergibson\/electron,GoooIce\/electron,etiktin\/electron,Jonekee\/electron,eric-seekas\/electron,BionicClick\/electron,jhen0409\/electron,renaesop\/electron,stevemao\/electron,miniak\/electron,Gerhut\/electron,webmechanicx\/electron,leftstick\/electron,rajatsingla28\/electron,pandoraui\/electron,tylergibson\/electron,Andrey-Pavlov\/electron,nagyistoce\/electron-atom-shell,vaginessa\/electron,Jacobichou\/electron,Ivshti\/electron,brave\/muon,shockone\/electron,oiledCode\/electron,LadyNaggaga\/electron,edulan\/electron,miniak\/electron,simonfork\/electron,leolujuyi\/electron,ervinb\/electron,posix4e\/electron,deepak1556\/atom-shell,beni55\/electron,seanchas116\/electron,setzer777\/electron,baiwyc119\/electron,kikong\/electron,DivyaKMenon\/electron,jonatasfreitasv\/electron,JussMee15\/electron,roadev\/electron,leolujuyi\/electron,destan\/electron,bobwol\/electron,saronwei\/electron,brenca\/electron,rsvip\/electron,deed02392\/electron,kokdemo\/electron,mhkeller\/electron,Jacobichou\/electron,maxogden\/atom-shell,kcrt\/electron,preco21\/electron,kenmozi\/electron,fomojola\/electron,jlhbaseball15\/electron,gabriel\/electron,IonicaBizauKitchen\/electron,RobertJGabriel\/electron,dkfiresky\/electron,lzpfmh\/electron,IonicaBizauKitchen\/electron,yalexx\/electron,michaelchiche\/electron,Andrey-Pavlov\/electron,trigrass2\/electron,tomashanacek\/electron,pombredanne\/electron,nekuz0r\/electron,DivyaKMenon\/electron,fomojola\/electron,simonfork\/electron,xfstudio\/electron,tinydew4\/electron,tinydew4\/electron,baiwyc119\/electron,trigrass2\/electron,wan-qy\/electron,ervinb\/electron,SufianHassan\/electron,stevekinney\/electron,rsvip\/electron,lzpfmh\/electron,rhencke\/electron,kenmozi\/electron,xiruibing\/electron,micalan\/electron,mrwizard82d1\/electron,RobertJGabriel\/electron,leolujuyi\/electron,RIAEvangelist\/electron,shaundunne\/electron,fabien-d\/electron,voidbridge\/electron,bwiggs\/electron,John-Lin\/electron,jtburke\/electron,simonfork\/electron,cos2004\/electron,evgenyzinoviev\/electron,faizalpribadi\/electron,gbn972\/electron,mrwizard82d1\/electron,electron\/electron,shiftkey\/electron,carsonmcdonald\/electron,kikong\/electron,brenca\/electron,kcrt\/electron,mhkeller\/electron,gamedevsam\/electron,thompsonemerson\/electron,kazupon\/electron,xiruibing\/electron,jcblw\/electron,jlord\/electron,joneit\/electron,beni55\/electron,tincan24\/electron,brave\/electron,matiasinsaurralde\/electron,gbn972\/electron,tonyganch\/electron,saronwei\/electron,gabrielPeart\/electron,etiktin\/electron,shockone\/electron,rajatsingla28\/electron,icattlecoder\/electron,shiftkey\/electron,the-ress\/electron,sircharleswatson\/electron,shaundunne\/electron,tonyganch\/electron,nagyistoce\/electron-atom-shell,arusakov\/electron,chrisswk\/electron,webmechanicx\/electron,kazupon\/electron,ervinb\/electron,meowlab\/electron,gerhardberger\/electron,shennushi\/electron,mjaniszew\/electron,the-ress\/electron,christian-bromann\/electron,Rokt33r\/electron,RobertJGabriel\/electron,matiasinsaurralde\/electron,pirafrank\/electron,bruce\/electron,RIAEvangelist\/electron,adamjgray\/electron,miniak\/electron,cos2004\/electron,jiaz\/electron,JesselJohn\/electron,brave\/electron,maxogden\/atom-shell,Faiz7412\/electron,iftekeriba\/electron,dahal\/electron,stevemao\/electron,adamjgray\/electron,evgenyzinoviev\/electron,leethomas\/electron,nicobot\/electron,zhakui\/electron,kcrt\/electron,yan-foto\/electron,BionicClick\/electron,shennushi\/electron,RIAEvangelist\/electron,abhishekgahlot\/electron,tonyganch\/electron,vHanda\/electron,nicholasess\/electron,BionicClick\/electron,eriser\/electron,maxogden\/atom-shell,wan-qy\/electron,carsonmcdonald\/electron,jhen0409\/electron,rsvip\/electron,aichingm\/electron,jannishuebl\/electron,Jacobichou\/electron,jtburke\/electron,voidbridge\/electron,kenmozi\/electron,vaginessa\/electron,oiledCode\/electron,shockone\/electron,LadyNaggaga\/electron,Zagorakiss\/electron,thomsonreuters\/electron,nicobot\/electron,cos2004\/electron,MaxWhere\/electron,noikiy\/electron,christian-bromann\/electron,mattotodd\/electron,anko\/electron,Gerhut\/electron,felixrieseberg\/electron,Floato\/electron,saronwei\/electron,thingsinjars\/electron,trankmichael\/electron,oiledCode\/electron,timruffles\/electron,mjaniszew\/electron,rreimann\/electron,gstack\/infinium-shell,neutrous\/electron,Floato\/electron,benweissmann\/electron,takashi\/electron,micalan\/electron,brave\/muon,hokein\/atom-shell,vaginessa\/electron,pirafrank\/electron,vHanda\/electron,sircharleswatson\/electron,jjz\/electron,greyhwndz\/electron,tincan24\/electron,yalexx\/electron,roadev\/electron,gabriel\/electron,Faiz7412\/electron,takashi\/electron,kcrt\/electron,jsutcodes\/electron,chrisswk\/electron,greyhwndz\/electron,tylergibson\/electron,roadev\/electron,chriskdon\/electron,pirafrank\/electron,jhen0409\/electron,jtburke\/electron,mirrh\/electron,Neron-X5\/electron,faizalpribadi\/electron,bobwol\/electron,shockone\/electron,shaundunne\/electron,Rokt33r\/electron,shiftkey\/electron,ervinb\/electron,sky7sea\/electron,setzer777\/electron,John-Lin\/electron,gabriel\/electron,evgenyzinoviev\/electron,icattlecoder\/electron,bwiggs\/electron,vaginessa\/electron,cos2004\/electron,medixdev\/electron,twolfson\/electron,hokein\/atom-shell,deed02392\/electron,posix4e\/electron,aliib\/electron,mrwizard82d1\/electron,gerhardberger\/electron,Evercoder\/electron,dongjoon-hyun\/electron,Jonekee\/electron,pandoraui\/electron,timruffles\/electron,twolfson\/electron,Ivshti\/electron,benweissmann\/electron,thingsinjars\/electron,bbondy\/electron,joaomoreno\/atom-shell,cqqccqc\/electron,bitemyapp\/electron,abhishekgahlot\/electron,BionicClick\/electron,bruce\/electron,jhen0409\/electron,iftekeriba\/electron,mubassirhayat\/electron,miniak\/electron,smczk\/electron,gbn972\/electron,vipulroxx\/electron,davazp\/electron,digideskio\/electron,dkfiresky\/electron,nagyistoce\/electron-atom-shell,simongregory\/electron,d-salas\/electron,nicobot\/electron,jhen0409\/electron,Evercoder\/electron,vHanda\/electron,eric-seekas\/electron,jsutcodes\/electron,sky7sea\/electron,thingsinjars\/electron,howmuchcomputer\/electron,iftekeriba\/electron,bright-sparks\/electron,Andrey-Pavlov\/electron,posix4e\/electron,electron\/electron,bpasero\/electron,twolfson\/electron,fireball-x\/atom-shell,kokdemo\/electron,jiaz\/electron,greyhwndz\/electron,meowlab\/electron,John-Lin\/electron,chriskdon\/electron,farmisen\/electron,micalan\/electron,nekuz0r\/electron,fabien-d\/electron,nicholasess\/electron,rsvip\/electron,LadyNaggaga\/electron,fabien-d\/electron,aecca\/electron,cqqccqc\/electron,tinydew4\/electron,natgolov\/electron,rajatsingla28\/electron,fomojola\/electron,brave\/muon,bpasero\/electron,kcrt\/electron,smczk\/electron,bobwol\/electron,eric-seekas\/electron,gabrielPeart\/electron,sshiting\/electron,Neron-X5\/electron,twolfson\/electron,renaesop\/electron,ianscrivener\/electron,adcentury\/electron,kenmozi\/electron,synaptek\/electron,gamedevsam\/electron,rhencke\/electron,Evercoder\/electron,jaanus\/electron,aecca\/electron,howmuchcomputer\/electron,davazp\/electron,iftekeriba\/electron,electron\/electron,aliib\/electron,joneit\/electron,trigrass2\/electron,dkfiresky\/electron,takashi\/electron,biblerule\/UMCTelnetHub,medixdev\/electron,sky7sea\/electron,fffej\/electron,bobwol\/electron,jlord\/electron,benweissmann\/electron,fabien-d\/electron,thompsonemerson\/electron,gabriel\/electron,simonfork\/electron,xiruibing\/electron,jcblw\/electron,mattotodd\/electron,John-Lin\/electron,fabien-d\/electron,electron\/electron,ianscrivener\/electron,digideskio\/electron,pombredanne\/electron,stevemao\/electron,davazp\/electron,jhen0409\/electron,edulan\/electron,joaomoreno\/atom-shell,mrwizard82d1\/electron,carsonmcdonald\/electron,Rokt33r\/electron,bpasero\/electron,aaron-goshine\/electron,electron\/electron,Jonekee\/electron,benweissmann\/electron,gabriel\/electron,mhkeller\/electron,bright-sparks\/electron,JesselJohn\/electron,rreimann\/electron,kikong\/electron,astoilkov\/electron,christian-bromann\/electron,soulteary\/electron,coderhaoxin\/electron,Ivshti\/electron,mattotodd\/electron,shennushi\/electron,pombredanne\/electron,eric-seekas\/electron,DivyaKMenon\/electron,ervinb\/electron,nicholasess\/electron,jjz\/electron,JussMee15\/electron,jiaz\/electron,mubassirhayat\/electron,thomsonreuters\/electron,anko\/electron,GoooIce\/electron,davazp\/electron,arturts\/electron,leolujuyi\/electron,etiktin\/electron,DivyaKMenon\/electron,stevekinney\/electron,wan-qy\/electron,jacksondc\/electron,mattdesl\/electron,Zagorakiss\/electron,robinvandernoord\/electron,rajatsingla28\/electron,brave\/muon,tincan24\/electron,brave\/muon,gabrielPeart\/electron,jonatasfreitasv\/electron,yan-foto\/electron,thingsinjars\/electron,mjaniszew\/electron,setzer777\/electron,simongregory\/electron,kenmozi\/electron,biblerule\/UMCTelnetHub,pirafrank\/electron,yan-foto\/electron,felixrieseberg\/electron,smczk\/electron,medixdev\/electron,bright-sparks\/electron,kostia\/electron,leftstick\/electron,Neron-X5\/electron,carsonmcdonald\/electron,JussMee15\/electron,ianscrivener\/electron,oiledCode\/electron,jcblw\/electron,etiktin\/electron,fireball-x\/atom-shell,meowlab\/electron,Floato\/electron,RobertJGabriel\/electron,gbn972\/electron,nicobot\/electron,noikiy\/electron,jlhbaseball15\/electron,miniak\/electron,pirafrank\/electron,tonyganch\/electron,adamjgray\/electron,preco21\/electron,tomashanacek\/electron,Rokt33r\/electron,bwiggs\/electron,gabrielPeart\/electron,dongjoon-hyun\/electron,iftekeriba\/electron,xiruibing\/electron,jtburke\/electron,deepak1556\/atom-shell,mattdesl\/electron,evgenyzinoviev\/electron,jannishuebl\/electron,medixdev\/electron,saronwei\/electron,Andrey-Pavlov\/electron,seanchas116\/electron,joneit\/electron,vipulroxx\/electron,Evercoder\/electron,tonyganch\/electron,shennushi\/electron,Zagorakiss\/electron,yan-foto\/electron,astoilkov\/electron,MaxGraey\/electron,systembugtj\/electron,sshiting\/electron,thompsonemerson\/electron,thompsonemerson\/electron,howmuchcomputer\/electron,dongjoon-hyun\/electron,aliib\/electron,michaelchiche\/electron,Gerhut\/electron,carsonmcdonald\/electron,ankitaggarwal011\/electron,the-ress\/electron,tinydew4\/electron,bpasero\/electron,vipulroxx\/electron,shaundunne\/electron,deed02392\/electron,greyhwndz\/electron,jjz\/electron,jacksondc\/electron,DivyaKMenon\/electron,Jacobichou\/electron,subblue\/electron,webmechanicx\/electron,brenca\/electron,timruffles\/electron,micalan\/electron,robinvandernoord\/electron,adcentury\/electron,rhencke\/electron,baiwyc119\/electron,Faiz7412\/electron,jlhbaseball15\/electron,matiasinsaurralde\/electron,faizalpribadi\/electron,systembugtj\/electron,nagyistoce\/electron-atom-shell,vaginessa\/electron,coderhaoxin\/electron,jannishuebl\/electron,soulteary\/electron,lrlna\/electron,roadev\/electron,baiwyc119\/electron,RIAEvangelist\/electron,fritx\/electron,preco21\/electron,micalan\/electron,beni55\/electron,ankitaggarwal011\/electron,dahal\/electron,mattdesl\/electron,SufianHassan\/electron,trigrass2\/electron,shennushi\/electron,howmuchcomputer\/electron,the-ress\/electron,biblerule\/UMCTelnetHub,kikong\/electron,preco21\/electron,howmuchcomputer\/electron,soulteary\/electron,JussMee15\/electron,neutrous\/electron,farmisen\/electron,adamjgray\/electron,saronwei\/electron,smczk\/electron,dongjoon-hyun\/electron,lrlna\/electron,IonicaBizauKitchen\/electron,nagyistoce\/electron-atom-shell,jcblw\/electron,kazupon\/electron,sshiting\/electron,brenca\/electron,Faiz7412\/electron,medixdev\/electron,thomsonreuters\/electron,fomojola\/electron,MaxWhere\/electron,systembugtj\/electron,chrisswk\/electron,aecca\/electron,dkfiresky\/electron,sircharleswatson\/electron,digideskio\/electron,pandoraui\/electron,mattdesl\/electron,shaundunne\/electron,fritx\/electron,tomashanacek\/electron,bright-sparks\/electron,cos2004\/electron,thingsinjars\/electron,tylergibson\/electron,Gerhut\/electron,minggo\/electron,abhishekgahlot\/electron,dahal\/electron,abhishekgahlot\/electron,jacksondc\/electron,arusakov\/electron,RIAEvangelist\/electron,wolfflow\/electron,vHanda\/electron,anko\/electron,matiasinsaurralde\/electron,natgolov\/electron,cqqccqc\/electron,IonicaBizauKitchen\/electron,gerhardberger\/electron,subblue\/electron,felixrieseberg\/electron,bpasero\/electron,bbondy\/electron,jaanus\/electron,trankmichael\/electron,christian-bromann\/electron,dkfiresky\/electron,soulteary\/electron,pombredanne\/electron,aaron-goshine\/electron,mirrh\/electron,eriser\/electron,subblue\/electron,vipulroxx\/electron,michaelchiche\/electron,tincan24\/electron,deepak1556\/atom-shell,noikiy\/electron,gerhardberger\/electron,noikiy\/electron,preco21\/electron,gabriel\/electron,fritx\/electron,vHanda\/electron,cqqccqc\/electron,the-ress\/electron,pombredanne\/electron,JesselJohn\/electron,trankmichael\/electron,rhencke\/electron,arturts\/electron,trigrass2\/electron,takashi\/electron,JesselJohn\/electron,trankmichael\/electron,JussMee15\/electron,dahal\/electron,renaesop\/electron,nekuz0r\/electron,mubassirhayat\/electron,jannishuebl\/electron,abhishekgahlot\/electron,darwin\/electron,felixrieseberg\/electron,wolfflow\/electron,darwin\/electron,jtburke\/electron,farmisen\/electron,systembugtj\/electron,LadyNaggaga\/electron,fomojola\/electron,shockone\/electron,meowlab\/electron,renaesop\/electron,tylergibson\/electron,leethomas\/electron,aecca\/electron,mattdesl\/electron,sircharleswatson\/electron,tomashanacek\/electron,benweissmann\/electron,arturts\/electron,minggo\/electron,kostia\/electron,adcentury\/electron,systembugtj\/electron"} {"commit":"4146be648f04ed409eb82e43528bc700751ef03c","old_file":"src\/qtlayoutbuilder\/builderror_test.py","new_file":"src\/qtlayoutbuilder\/builderror_test.py","old_contents":"from unittest import TestCase\n\n\nclass TestBuildError(TestCase):\n def test_push_message(self):\n self.fail()\n\n def test_format_as_single_string(self):\n self.faildoo()\n","new_contents":"from unittest import TestCase\n\nfrom builderror import BuildError\n\nclass TestBuildError(TestCase):\n\n def test_that_multiple_pushed_messages_are_formatted_properly_when_asked_for(self):\n err = BuildError()\n err.push_message('message about error details')\n err.push_message('message about error context')\n formatted_message = err.format_as_single_string()\n self.assertEquals(formatted_message, 'message about error context\\nmessage about error details')\n\n","subject":"Put in first proper tiny package class and unit test.","message":"Put in first proper tiny package class and unit test.\n","lang":"Python","license":"mit","repos":"peterhoward42\/qt-layout-gen"} {"commit":"1713cf8553d7f21d1192ed58138ecf7875c4b181","old_file":"icebergsdk\/front_modules.py","new_file":"icebergsdk\/front_modules.py","old_contents":"# -*- coding: utf-8 -*-\nimport logging\n\nfrom icebergsdk.mixins.request_mixin import IcebergRequestBase\n\nlogger = logging.getLogger('icebergsdk.frontmodules')\n\nclass FrontModules(IcebergRequestBase):\n cache_key = \"icebergsdk:frontmodule:data\"\n cache_expire = 60*60 # one hour\n\n def __init__(self, *args, **kwargs):\n super(FrontModules, self).__init__(*args, **kwargs)\n self.cache = kwargs.get('cache', None)\n self.lang = kwargs.get('lang', \"en\")\n\n\n def get_module_data(self, module_name):\n return self.modules_data['modules'][module_name]\n\n ####\n # Loader\n ####\n @property\n def modules_data(self):\n \"\"\"\n Helper to fetch Iceberg client side javascript templates\n \"\"\"\n if hasattr(self, \"_modules_data\"):\n return getattr(self, \"_modules_data\")\n\n if self.cache:\n data = self.cache.get(\"%s:%s\" % (self.cache_key, self.lang), False)\n if data:\n setattr(self, '_modules_data', data)\n return data\n\n data = self.request(self.conf.ICEBERG_MODULES_URL) # Do to, add lang\n setattr(self, '_modules_data', data)\n if self.cache:\n self.cache.set(\"%s:%s\" % (self.cache_key, self.lang), data, self.cache_expire)\n\n return data\n\n\n","new_contents":"# -*- coding: utf-8 -*-\nimport logging\n\nfrom icebergsdk.mixins.request_mixin import IcebergRequestBase\n\nlogger = logging.getLogger('icebergsdk.frontmodules')\n\nclass FrontModules(IcebergRequestBase):\n cache_key = \"icebergsdk:frontmodule:data\"\n cache_expire = 60*60 # one hour\n\n def __init__(self, *args, **kwargs):\n super(FrontModules, self).__init__(*args, **kwargs)\n self.cache = kwargs.get('cache', None)\n self.lang = kwargs.get('lang', \"en\")\n self.debug = kwargs.get('debug', False)\n\n\n def get_module_data(self, module_name):\n return self.modules_data['modules'][module_name]\n\n ####\n # Loader\n ####\n @property\n def modules_data(self):\n \"\"\"\n Helper to fetch Iceberg client side javascript templates\n \"\"\"\n if hasattr(self, \"_modules_data\"):\n return getattr(self, \"_modules_data\")\n\n if self.cache:\n data = self.cache.get(\"%s:%s\" % (self.cache_key, self.lang), False)\n if data:\n setattr(self, '_modules_data', data)\n return data\n\n data = self.request(self.conf.ICEBERG_MODULES_URL, args = {\n \"lang\": self.lang,\n \"enviro\": self.conf.ICEBERG_ENV,\n \"debug\": self.debug\n }) # Do to, add lang\n setattr(self, '_modules_data', data)\n if self.cache:\n self.cache.set(\"%s:%s\" % (self.cache_key, self.lang), data, self.cache_expire)\n\n return data\n\n\n","subject":"Add lang, enviro in request","message":"Add lang, enviro in request\n","lang":"Python","license":"mit","repos":"izberg-marketplace\/izberg-api-python,Iceberg-Marketplace\/Iceberg-API-PYTHON"} {"commit":"adf747998641b1aeb75feada25470aa2a072bd37","old_file":"examples\/test-mh\/policies\/participant_3.py","new_file":"examples\/test-mh\/policies\/participant_3.py","old_contents":"{\n \"inbound\": [\n {\n \"cookie\": 1,\n \"match\": \n {\n \"tcp_dst\": 4321\n },\n \"action\": \n {\n \"fwd\": 0\n }\n },\n {\n \"cookie\": 2,\n \"match\": \n {\n \"tcp_dst\": 4322\n },\n \"action\": \n {\n \"fwd\": 1\n }\n },\n {\n \"cookie\": 3,\n \"match\": \n {\n \"tcp_dst\": 4323\n },\n \"action\": \n {\n \"drop\": 0\n }\n }\n ]\n}","new_contents":"{\n \"inbound\": [\n {\n \"cookie\": 1,\n \"match\": \n {\n \"tcp_dst\": 4321\n },\n \"action\": \n {\n \"fwd\": 0\n }\n },\n {\n \"cookie\": 2,\n \"match\": \n {\n \"tcp_dst\": 4322\n },\n \"action\": \n {\n \"fwd\": 1\n }\n },\n {\n \"cookie\": 3,\n \"match\": \n {\n \"eth_src\": '08:00:27:89:3b:9f'\n },\n \"action\": \n {\n \"drop\": 0\n }\n }\n ]\n}","subject":"Add inbound drop policy for participant 3 based on eth_src of participant 1","message":"Add inbound drop policy for participant 3 based on eth_src of participant 1\n","lang":"Python","license":"apache-2.0","repos":"h2020-endeavour\/endeavour,h2020-endeavour\/endeavour"} {"commit":"626345a10ee3d0aa90b7791e88c9b09544bdfa88","old_file":"daiquiri\/files\/views.py","new_file":"daiquiri\/files\/views.py","old_contents":"import logging\n\nfrom django.contrib.auth.views import redirect_to_login\nfrom django.core.exceptions import PermissionDenied\nfrom django.http import Http404\nfrom django.views.generic import View\n\nfrom .utils import file_exists, get_directory, render_with_layout, send_file\n\nlogger = logging.getLogger(__name__)\n\n\nclass FileView(View):\n\n def get(self, request, file_path):\n # append 'index.html' when the file_path is a directory\n if not file_path or file_path.endswith('\/'):\n file_path += 'index.html'\n\n if not file_exists(file_path):\n logger.debug('%s not found', file_path)\n raise Http404\n\n directory = get_directory(request.user, file_path)\n if directory is None:\n logger.debug('%s if forbidden', file_path)\n if request.user.is_authenticated:\n raise PermissionDenied\n else:\n return redirect_to_login(request.path_info)\n\n if directory.layout:\n return render_with_layout(request, file_path)\n else:\n return send_file(request, file_path)\n","new_contents":"import logging\n\nfrom django.contrib.auth.views import redirect_to_login\nfrom django.core.exceptions import PermissionDenied\nfrom django.http import Http404\nfrom django.views.generic import View\n\nfrom .utils import file_exists, get_directory, render_with_layout, send_file\n\nlogger = logging.getLogger(__name__)\n\n\nclass FileView(View):\n\n def get(self, request, file_path):\n # append 'index.html' when the file_path is a directory\n if not file_path or file_path.endswith('\/'):\n file_path += 'index.html'\n\n if not file_exists(file_path):\n logger.debug('%s not found', file_path)\n raise Http404\n\n directory = get_directory(request.user, file_path)\n if directory is None:\n logger.debug('%s if forbidden', file_path)\n if request.user.is_authenticated:\n raise PermissionDenied\n else:\n return redirect_to_login(request.path_info)\n\n if file_path.endswith('.html') and directory.layout:\n return render_with_layout(request, file_path)\n else:\n return send_file(request, file_path)\n","subject":"Remove layout for non-html files","message":"Remove layout for non-html files\n","lang":"Python","license":"apache-2.0","repos":"aipescience\/django-daiquiri,aipescience\/django-daiquiri,aipescience\/django-daiquiri"} {"commit":"b71f076bbb745764d6ed5724d494e878a8fbd785","old_file":"syncplay\/__init__.py","new_file":"syncplay\/__init__.py","old_contents":"version = '1.6.5'\nrevision = ' development'\nmilestone = 'Yoitsu'\nrelease_number = '85'\nprojectURL = 'https:\/\/syncplay.pl\/'\n","new_contents":"version = '1.6.5'\nrevision = ' release'\nmilestone = 'Yoitsu'\nrelease_number = '86'\nprojectURL = 'https:\/\/syncplay.pl\/'\n","subject":"Mark build 86 and v1.6.5 release","message":"Mark build 86 and v1.6.5 release","lang":"Python","license":"apache-2.0","repos":"Syncplay\/syncplay,Syncplay\/syncplay,alby128\/syncplay,alby128\/syncplay"} {"commit":"b5206de1d6a99cd4a22f1ef5681e7964a7289b2e","old_file":"debug_toolbar_multilang\/pseudo\/expander_pseudo_language.py","new_file":"debug_toolbar_multilang\/pseudo\/expander_pseudo_language.py","old_contents":"from django.utils import six\nfrom debug_toolbar_multilang.pseudo import STR_FORMAT_PATTERN, \\\n STR_FORMAT_NAMED_PATTERN\nfrom debug_toolbar_multilang.pseudo.pseudo_language import PseudoLanguage\n\n\nclass ExpanderPseudoLanguage(PseudoLanguage):\n \"\"\"\n Pseudo Language for expanding the strings. This is useful\n for verifying that the message still fits on the screen.\n Remember that some words are much more longer in other\n languages than in English. For instance, German words\n that 30% more space in average.\n \"\"\"\n\n def make_pseudo(self, message):\n # message without %s or {} in it.\n # {test} or %(test)s is allowed, though.\n safeMessage = list(message)\n\n # find every matching string\n for match in reversed(list(STR_FORMAT_PATTERN.finditer(message))):\n # Check if string uses the \"named format\".\n # If not, the string will be replaced and saved\n # into safeMessage\n if not STR_FORMAT_NAMED_PATTERN.match(match.group()):\n start, end = match.span()\n safeMessage[start:end] = \"???\"\n\n # create complete message by using the original, appending\n # a space and finally converting the safeMessage to a string\n # again.\n return six.u(\"%s %s\" % (message, \"\".join(safeMessage)))\n\n def language(self):\n return \"pse-expander\"\n\n @property\n def name(self):\n return \"Pseudo-Expander Language\"\n","new_contents":"from django.utils import six\nfrom debug_toolbar_multilang.pseudo import STR_FORMAT_PATTERN, \\\n STR_FORMAT_NAMED_PATTERN\nfrom debug_toolbar_multilang.pseudo.pseudo_language import PseudoLanguage\n\n\nclass ExpanderPseudoLanguage(PseudoLanguage):\n \"\"\"\n Pseudo Language for expanding the strings. This is useful\n for verifying that the message still fits on the screen.\n Remember that some words are much more longer in other\n languages than in English. For instance, German words\n that 30% more space in average.\n \"\"\"\n\n def make_pseudo(self, message):\n # message without %s or {} in it.\n # {test} or %(test)s is allowed, though.\n safeMessage = list(message)\n\n # find every matching string\n for match in reversed(list(STR_FORMAT_PATTERN.finditer(message))):\n # Check if string uses the \"named format\".\n # If not, the string will be replaced and saved\n # into safeMessage\n if not STR_FORMAT_NAMED_PATTERN.match(match.group()):\n start, end = match.span()\n safeMessage[start:end] = \"???\"\n\n # create complete message by using the original, appending\n # a space and finally converting the safeMessage to a string\n # again.\n return \"%s %s\" % (message, \"\".join(safeMessage))\n\n def language(self):\n return \"pse-expander\"\n\n @property\n def name(self):\n return \"Pseudo-Expander Language\"\n","subject":"Fix python 2 unicode issue.","message":"Fix python 2 unicode issue.\n","lang":"Python","license":"mit","repos":"Matt3o12\/django-debug-toolbar-multilang,Matt3o12\/django-debug-toolbar-multilang"} {"commit":"c27010a3d5265d9eb783f627adca7cb0c25dcb9a","old_file":"ctypeslib\/test\/stdio.py","new_file":"ctypeslib\/test\/stdio.py","old_contents":"import os\nfrom ctypeslib.dynamic_module import include\nfrom ctypes import *\nimport logging\nlogging.basicConfig(level=logging.INFO)\n\nif os.name == \"nt\":\n _libc = CDLL(\"msvcrt\")\nelse:\n _libc = CDLL(None)\n\ninclude(\"\"\"\\\n#include <stdio.h>\n\n#ifdef _MSC_VER\n# include <fcntl.h>\n#else\n# include <sys\/fcntl.h>\n#endif\n\"\"\",\n persist=False)\n","new_contents":"import os\nfrom ctypeslib.dynamic_module import include\nfrom ctypes import *\n\nif os.name == \"nt\":\n _libc = CDLL(\"msvcrt\")\nelse:\n _libc = CDLL(None)\n\ninclude(\"\"\"\\\n#include <stdio.h>\n\n#ifdef _MSC_VER\n# include <fcntl.h>\n#else\n# include <sys\/fcntl.h>\n#endif\n\"\"\",\n persist=False)\n","subject":"Remove the logging setup call.","message":"Remove the logging setup call.","lang":"Python","license":"mit","repos":"sugarmanz\/ctypeslib"} {"commit":"5d0541f5b5b8cc18b2e3f86b237c01ed915d5c0a","old_file":"dhcp2nest\/util.py","new_file":"dhcp2nest\/util.py","old_contents":"\"\"\"\nUtility functions for dhcp2nest\n\"\"\"\nfrom queue import Queue\nfrom subprocess import Popen, PIPE\nfrom threading import Thread\n\n\ndef follow_file(fn, max_lines=100):\n \"\"\"\n Return a Queue that is fed lines (up to max_lines) from the given file (fn)\n continuously\n\n The implementation given here was inspired by\n http:\/\/stackoverflow.com\/questions\/12523044\/how-can-i-tail-a-log-file-in-python\n \"\"\"\n fq = Queue(maxsize=max_lines)\n\n # Declare the helper routine\n def _follow_file_thread(fn, fq):\n # Use system tail with name-based following and retry\n p = Popen([\"tail\", \"-F\", fn], stdout=PIPE)\n\n # Loop forever on pulling data from tail\n line = True\n while line:\n line = p.stdout.readline()\n fq.put(line)\n\n # Spawn a thread to read data from tail\n Thread(target=_follow_file_thread, args=(fn, fq)).start()\n\n # Return the queue\n return fq\n","new_contents":"\"\"\"\nUtility functions for dhcp2nest\n\"\"\"\nfrom queue import Queue\nfrom subprocess import Popen, PIPE\nfrom threading import Thread\n\n\ndef follow_file(fn, max_lines=100):\n \"\"\"\n Return a Queue that is fed lines (up to max_lines) from the given file (fn)\n continuously\n\n The implementation given here was inspired by\n http:\/\/stackoverflow.com\/questions\/12523044\/how-can-i-tail-a-log-file-in-python\n \"\"\"\n fq = Queue(maxsize=max_lines)\n\n # Declare the helper routine\n def _follow_file_thread(fn, fq):\n # Use system tail with name-based following and retry\n p = Popen([\"tail\", \"-F\", fn], stdout=PIPE)\n\n # Loop forever on pulling data from tail\n line = True\n while line:\n line = p.stdout.readline().decode('utf-8')\n fq.put(line)\n\n # Spawn a thread to read data from tail\n Thread(target=_follow_file_thread, args=(fn, fq)).start()\n\n # Return the queue\n return fq\n","subject":"Make sure that follow-file decodes utf-8 from its input","message":"Make sure that follow-file decodes utf-8 from its input\n\nSigned-off-by: Jason Bernardino Alonso <f71c42a1353bbcdbe07e24c2a1c893f8ea1d05ee@hackorp.com>\n","lang":"Python","license":"mit","repos":"jbalonso\/dhcp2nest"} {"commit":"cb75a7ad69b273a57d2b10378712388f179abca3","old_file":"pande_gas\/features\/tests\/test_fingerprints.py","new_file":"pande_gas\/features\/tests\/test_fingerprints.py","old_contents":"\"\"\"\nTest topological fingerprints.\n\"\"\"\nimport unittest\n\nfrom rdkit import Chem\n\nfrom pande_gas.features import fingerprints as fp\n\n\nclass TestCircularFingerprint(unittest.TestCase):\n \"\"\"\n Tests for CircularFingerprint.\n \"\"\"\n def setUp(self):\n \"\"\"\n Set up tests.\n \"\"\"\n smiles = 'CC(=O)OC1=CC=CC=C1C(=O)O'\n self.mol = Chem.MolFromSmiles(smiles)\n self.engine = fp.CircularFingerprint()\n\n def test_circular_fingerprints(self):\n \"\"\"\n Test CircularFingerprint.\n \"\"\"\n rval = self.engine([self.mol])\n assert rval.shape == (1, self.engine.size)\n\n def test_sparse_circular_fingerprints(self):\n \"\"\"\n Test CircularFingerprint with sparse encoding.\n \"\"\"\n self.engine = fp.CircularFingerprint(sparse=True)\n rval = self.engine([self.mol])\n assert rval.shape == (1,)\n assert isinstance(rval[0], dict)\n assert len(rval[0])\n","new_contents":"\"\"\"\nTest topological fingerprints.\n\"\"\"\nimport unittest\n\nfrom rdkit import Chem\n\nfrom pande_gas.features import fingerprints as fp\n\n\nclass TestCircularFingerprint(unittest.TestCase):\n \"\"\"\n Tests for CircularFingerprint.\n \"\"\"\n def setUp(self):\n \"\"\"\n Set up tests.\n \"\"\"\n smiles = 'CC(=O)OC1=CC=CC=C1C(=O)O'\n self.mol = Chem.MolFromSmiles(smiles)\n self.engine = fp.CircularFingerprint()\n\n def test_circular_fingerprints(self):\n \"\"\"\n Test CircularFingerprint.\n \"\"\"\n rval = self.engine([self.mol])\n assert rval.shape == (1, self.engine.size)\n\n def test_sparse_circular_fingerprints(self):\n \"\"\"\n Test CircularFingerprint with sparse encoding.\n \"\"\"\n self.engine = fp.CircularFingerprint(sparse=True)\n rval = self.engine([self.mol])\n assert rval.shape == (1,)\n assert isinstance(rval[0], dict)\n assert len(rval[0])\n\n def test_sparse_circular_fingerprints_with_smiles(self):\n \"\"\"\n Test CircularFingerprint with sparse encoding and SMILES for each\n fragment.\n \"\"\"\n self.engine = fp.CircularFingerprint(sparse=True, smiles=True)\n rval = self.engine([self.mol])\n assert rval.shape == (1,)\n assert isinstance(rval[0], dict)\n assert len(rval[0])\n\n # check for separate count and SMILES entries for each fragment\n for fragment_id, value in rval[0].items():\n assert 'count' in value\n assert 'smiles' in value\n","subject":"Add test for fragment SMILES","message":"Add test for fragment SMILES\n","lang":"Python","license":"bsd-3-clause","repos":"rbharath\/pande-gas,rbharath\/pande-gas"} {"commit":"891e9e05f8c3fde75bb217d8d8132cdf6003e827","old_file":"locust\/shape.py","new_file":"locust\/shape.py","old_contents":"from __future__ import annotations\nimport time\nfrom typing import Optional, Tuple, List, Type\n\nfrom . import User\nfrom .runners import Runner\n\n\nclass LoadTestShape:\n \"\"\"\n A simple load test shape class used to control the shape of load generated\n during a load test.\n \"\"\"\n\n runner: Optional[Runner] = None\n \"\"\"Reference to the :class:`Runner <locust.runners.Runner>` instance\"\"\"\n\n def __init__(self):\n self.start_time = time.perf_counter()\n\n def reset_time(self):\n \"\"\"\n Resets start time back to 0\n \"\"\"\n self.start_time = time.perf_counter()\n\n def get_run_time(self):\n \"\"\"\n Calculates run time in seconds of the load test\n \"\"\"\n return time.perf_counter() - self.start_time\n\n def get_current_user_count(self):\n \"\"\"\n Returns current actual number of users from the runner\n \"\"\"\n return self.runner.user_count\n\n def tick(self) -> Tuple[int, float] | Tuple[int, float, Optional[List[Type[User]]]] | None:\n \"\"\"\n Returns a tuple with 2 elements to control the running load test:\n\n user_count -- Total user count\n spawn_rate -- Number of users to start\/stop per second when changing number of users\n user_classes -- None or a List of userclasses to be spawned in it tick\n\n If `None` is returned then the running load test will be stopped.\n\n \"\"\"\n\n return None\n","new_contents":"from __future__ import annotations\nimport time\nfrom typing import Optional, Tuple, List, Type\nfrom abc import ABC, abstractmethod\n\nfrom . import User\nfrom .runners import Runner\n\n\nclass LoadTestShape(ABC):\n \"\"\"\n Base class for custom load shapes.\n \"\"\"\n\n runner: Optional[Runner] = None\n \"\"\"Reference to the :class:`Runner <locust.runners.Runner>` instance\"\"\"\n\n def __init__(self):\n self.start_time = time.perf_counter()\n\n def reset_time(self):\n \"\"\"\n Resets start time back to 0\n \"\"\"\n self.start_time = time.perf_counter()\n\n def get_run_time(self):\n \"\"\"\n Calculates run time in seconds of the load test\n \"\"\"\n return time.perf_counter() - self.start_time\n\n def get_current_user_count(self):\n \"\"\"\n Returns current actual number of users from the runner\n \"\"\"\n return self.runner.user_count\n\n @abstractmethod\n def tick(self) -> Tuple[int, float] | Tuple[int, float, Optional[List[Type[User]]]] | None:\n \"\"\"\n Returns a tuple with 2 elements to control the running load test:\n\n user_count -- Total user count\n spawn_rate -- Number of users to start\/stop per second when changing number of users\n user_classes -- None or a List of userclasses to be spawned in it tick\n\n If `None` is returned then the running load test will be stopped.\n\n \"\"\"\n ...\n","subject":"Make LoadTestShape a proper abstract class.","message":"Make LoadTestShape a proper abstract class.\n","lang":"Python","license":"mit","repos":"locustio\/locust,locustio\/locust,locustio\/locust,locustio\/locust"} {"commit":"4bdaf4d2e29da71a1bf00e1bfc5caad6d3647372","old_file":"search\/views.py","new_file":"search\/views.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponse\nfrom django.core.management import call_command\n\nfrom django.shortcuts import render\n\n@login_required(login_url='\/accounts\/login\/')\ndef search_index(request):\n call_command('search-index', 'all')\n return HttpResponse(\"Index process done.\")","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\nfrom django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponse\nfrom django.core.management import call_command\n\nfrom django.shortcuts import render\n\n@login_required(login_url='\/accounts\/login\/')\ndef search_index(request):\n call_command('search-index', '_all')\n return HttpResponse(\"Index process done.\")","subject":"Fix error in search index not updating on view call.","message":"Fix error in search index not updating on view call.\n","lang":"Python","license":"apache-2.0","repos":"toladata\/TolaActivity,toladata\/TolaActivity,toladata\/TolaActivity,toladata\/TolaActivity"} {"commit":"1f0e5b7e65914ec5c3fb0a6617f72ea2f466bbdc","old_file":"server\/admin.py","new_file":"server\/admin.py","old_contents":"from django.contrib import admin\nfrom server.models import *\n\n\nclass MachineGroupAdmin(admin.ModelAdmin):\n readonly_fields = ('key',)\n\n\nclass MachineAdmin(admin.ModelAdmin):\n list_display = ('hostname', 'serial')\n\n\nadmin.site.register(UserProfile)\nadmin.site.register(BusinessUnit)\nadmin.site.register(MachineGroup, MachineGroupAdmin)\nadmin.site.register(Machine, MachineAdmin)\nadmin.site.register(Fact)\nadmin.site.register(PluginScriptSubmission)\nadmin.site.register(PluginScriptRow)\nadmin.site.register(HistoricalFact)\nadmin.site.register(Condition)\nadmin.site.register(PendingUpdate)\nadmin.site.register(InstalledUpdate)\nadmin.site.register(PendingAppleUpdate)\nadmin.site.register(ApiKey)\nadmin.site.register(Plugin)\nadmin.site.register(Report)\n# admin.site.register(OSQueryResult)\n# admin.site.register(OSQueryColumn)\nadmin.site.register(SalSetting)\nadmin.site.register(UpdateHistory)\nadmin.site.register(UpdateHistoryItem)\nadmin.site.register(MachineDetailPlugin)\n","new_contents":"from django.contrib import admin\n\nfrom server.models import *\n\n\nclass ApiKeyAdmin(admin.ModelAdmin):\n list_display = ('name', 'public_key', 'private_key')\n\n\nclass MachineAdmin(admin.ModelAdmin):\n list_display = ('hostname', 'serial')\n\n\nclass MachineGroupAdmin(admin.ModelAdmin):\n readonly_fields = ('key',)\n\n\nadmin.site.register(ApiKey, ApiKeyAdmin)\nadmin.site.register(BusinessUnit)\nadmin.site.register(Condition)\nadmin.site.register(Fact)\nadmin.site.register(HistoricalFact)\nadmin.site.register(InstalledUpdate)\nadmin.site.register(Machine, MachineAdmin)\nadmin.site.register(MachineDetailPlugin)\nadmin.site.register(MachineGroup, MachineGroupAdmin)\n# admin.site.register(OSQueryColumn)\n# admin.site.register(OSQueryResult)\nadmin.site.register(PendingAppleUpdate)\nadmin.site.register(PendingUpdate)\nadmin.site.register(Plugin)\nadmin.site.register(PluginScriptRow)\nadmin.site.register(PluginScriptSubmission)\nadmin.site.register(Report)\nadmin.site.register(SalSetting)\nadmin.site.register(UpdateHistory)\nadmin.site.register(UpdateHistoryItem)\nadmin.site.register(UserProfile)\n","subject":"Sort registrations. Separate classes of imports. Add API key display.","message":"Sort registrations. Separate classes of imports. Add API key display.\n","lang":"Python","license":"apache-2.0","repos":"salopensource\/sal,salopensource\/sal,sheagcraig\/sal,salopensource\/sal,sheagcraig\/sal,sheagcraig\/sal,salopensource\/sal,sheagcraig\/sal"} {"commit":"51076b9d21679b1198931e2517afbf7c6d2e573a","old_file":"src\/competition\/forms\/team_forms.py","new_file":"src\/competition\/forms\/team_forms.py","old_contents":"from django import forms\nfrom django.template.defaultfilters import slugify\n\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Layout, Fieldset, Submit\nfrom crispy_forms.bootstrap import FormActions\n\nfrom competition.models.team_model import Team\n\n\nclass TeamForm(forms.ModelForm):\n class Meta:\n model = Team\n fields = ('name', )\n\n def __init__(self, *args, **kwargs):\n self.helper = FormHelper()\n self.helper.form_class = 'form-horizontal'\n self.helper.layout = Layout(\n Fieldset(\n 'Create a new team',\n 'name',\n ),\n FormActions(\n Submit('submit', 'Submit')\n )\n )\n super(TeamForm, self).__init__(*args, **kwargs)\n\n def clean_name(self):\n c = self.instance.competition\n n = self.cleaned_data['name']\n s = slugify(n)\n if Team.objects.filter(competition=c, slug=s).exists():\n msg = \"This name is already taken for %s\" % c.name\n raise forms.ValidationError(msg)\n return n\n\n def validate_unique(self):\n exclude = self._get_validation_exclusions()\n exclude.remove('competition')\n\n try:\n self.instance.validate_unique(exclude=exclude)\n except ValidationError, e:\n self._update_errors(e.message_dict)\n","new_contents":"from django import forms\nfrom django.template.defaultfilters import slugify\n\nfrom crispy_forms.helper import FormHelper\nfrom crispy_forms.layout import Layout, Fieldset, Submit\nfrom crispy_forms.bootstrap import FormActions\n\nfrom competition.models.team_model import Team\n\n\nclass TeamForm(forms.ModelForm):\n class Meta:\n model = Team\n fields = ('name', )\n\n def __init__(self, *args, **kwargs):\n self.helper = FormHelper()\n self.helper.form_class = 'form-horizontal'\n self.helper.label_class = 'col-lg-2'\n self.helper.field_class = 'col-lg-8'\n self.helper.layout = Layout(\n Fieldset(\n 'Create a new team',\n 'name',\n ),\n FormActions(\n Submit('submit', 'Submit')\n )\n )\n super(TeamForm, self).__init__(*args, **kwargs)\n\n def clean_name(self):\n c = self.instance.competition\n n = self.cleaned_data['name']\n s = slugify(n)\n if Team.objects.filter(competition=c, slug=s).exists():\n msg = \"This name is already taken for %s\" % c.name\n raise forms.ValidationError(msg)\n return n\n\n def validate_unique(self):\n exclude = self._get_validation_exclusions()\n exclude.remove('competition')\n\n try:\n self.instance.validate_unique(exclude=exclude)\n except ValidationError, e:\n self._update_errors(e.message_dict)\n","subject":"Update forms to bootstrap 3","message":"Update forms to bootstrap 3\n\nform-horizontal needs additional helper classes in BS3\n","lang":"Python","license":"bsd-3-clause","repos":"michaelwisely\/django-competition,michaelwisely\/django-competition,michaelwisely\/django-competition"} {"commit":"07e780a27253c4108c96e232ffbb975e88d23f8d","old_file":"src\/pygrapes\/serializer\/__init__.py","new_file":"src\/pygrapes\/serializer\/__init__.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nfrom abstract import Abstract\n\n\n__all__ = ['Abstract']\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nfrom abstract import Abstract\nfrom json import Json\n\n__all__ = ['Abstract', 'Json']\n","subject":"Load pygrapes.serializer.json.Json right inside pygrapes.serializer","message":"Load pygrapes.serializer.json.Json right inside pygrapes.serializer\n","lang":"Python","license":"bsd-3-clause","repos":"michalbachowski\/pygrapes,michalbachowski\/pygrapes,michalbachowski\/pygrapes"} {"commit":"384c2f34fcdefd26a928254e70a9ed6d15ffd069","old_file":"dimod\/reference\/samplers\/random_sampler.py","new_file":"dimod\/reference\/samplers\/random_sampler.py","old_contents":"\"\"\"\nRandomSampler\n-------------\n\nA random sampler that can be used for unit testing and debugging.\n\"\"\"\nimport numpy as np\n\nfrom dimod.core.sampler import Sampler\nfrom dimod.response import Response, SampleView\n\n__all__ = ['RandomSampler']\n\n\nclass RandomSampler(Sampler):\n \"\"\"Gives random samples.\n\n Note that this sampler is intended for testing.\n\n \"\"\"\n def __init__(self):\n Sampler.__init__(self)\n self.sample_kwargs = {'num_reads': []}\n\n def sample(self, bqm, num_reads=10):\n \"\"\"Gives random samples.\n\n Args:\n todo\n\n Returns:\n :obj:`.Response`: The vartype will match the given binary quadratic model.\n\n Notes:\n For each variable in each sample, the value is chosen by a coin flip.\n\n \"\"\"\n values = np.asarray(list(bqm.vartype.value), dtype='int8')\n samples = np.random.choice(values, (num_reads, len(bqm)))\n variable_labels = list(bqm.linear)\n label_to_idx = {v: idx for idx, v in enumerate(variable_labels)}\n\n energies = [bqm.energy(SampleView(idx, samples, label_to_idx)) for idx in range(num_reads)]\n\n return Response.from_matrix(samples, {'energy': energies},\n vartype=bqm.vartype, variable_labels=variable_labels)\n","new_contents":"\"\"\"\nRandomSampler\n-------------\n\nA random sampler that can be used for unit testing and debugging.\n\"\"\"\nimport numpy as np\n\nfrom dimod.core.sampler import Sampler\nfrom dimod.response import Response, SampleView\n\n__all__ = ['RandomSampler']\n\n\nclass RandomSampler(Sampler):\n \"\"\"Gives random samples.\n\n Note that this sampler is intended for testing.\n\n \"\"\"\n properties = None\n parameters = None\n\n def __init__(self):\n self.parameters = {'num_reads': []}\n self.properties = {}\n\n def sample(self, bqm, num_reads=10):\n \"\"\"Gives random samples.\n\n Args:\n todo\n\n Returns:\n :obj:`.Response`: The vartype will match the given binary quadratic model.\n\n Notes:\n For each variable in each sample, the value is chosen by a coin flip.\n\n \"\"\"\n values = np.asarray(list(bqm.vartype.value), dtype='int8')\n samples = np.random.choice(values, (num_reads, len(bqm)))\n variable_labels = list(bqm.linear)\n label_to_idx = {v: idx for idx, v in enumerate(variable_labels)}\n\n energies = [bqm.energy(SampleView(idx, samples, label_to_idx)) for idx in range(num_reads)]\n\n return Response.from_matrix(samples, {'energy': energies},\n vartype=bqm.vartype, variable_labels=variable_labels)\n","subject":"Update RandomSampler to use the new Sampler abc","message":"Update RandomSampler to use the new Sampler abc\n","lang":"Python","license":"apache-2.0","repos":"dwavesystems\/dimod,dwavesystems\/dimod"} {"commit":"0aa5741ce05dcd4926be9c74af18f6fe46f4aded","old_file":"etl_framework\/utilities\/DatetimeConverter.py","new_file":"etl_framework\/utilities\/DatetimeConverter.py","old_contents":"\"\"\"class to convert datetime values\"\"\"\n\nimport datetime\n\nclass DatetimeConverter(object):\n \"\"\"stuff\"\"\"\n\n _EPOCH_0 = datetime.datetime(1970, 1, 1)\n\n def __init__(self):\n \"\"\"stuff\"\"\"\n\n pass\n\n @staticmethod\n def get_tomorrow():\n \"\"\"stuff\"\"\"\n\n return datetime.datetime.today() + datetime.timedelta(days=1)\n\n @classmethod\n def get_timestamp(cls, datetime_obj):\n \"\"\"helper method to return timestamp fo datetime object\"\"\"\n\n return (datetime_obj - cls._EPOCH_0).total_seconds()\n\n @classmethod\n def get_tomorrow_timestamp(cls):\n \"\"\"stuff\"\"\"\n\n return cls.get_timestamp(cls.get_tomorrow())\n","new_contents":"\"\"\"class to convert datetime values\"\"\"\n\nimport datetime\n\nclass DatetimeConverter(object):\n \"\"\"stuff\"\"\"\n\n _EPOCH_0 = datetime.datetime(1970, 1, 1)\n\n def __init__(self):\n \"\"\"stuff\"\"\"\n\n pass\n\n @staticmethod\n def get_tomorrow():\n \"\"\"stuff\"\"\"\n\n return datetime.datetime.today() + datetime.timedelta(days=1)\n\n @staticmethod\n def get_yesterday():\n\n return datetime.datetime.today() - datetime.timedelta(days=1)\n\n @classmethod\n def get_timestamp(cls, datetime_obj):\n \"\"\"helper method to return timestamp fo datetime object\"\"\"\n\n return (datetime_obj - cls._EPOCH_0).total_seconds()\n\n @classmethod\n def get_tomorrow_timestamp(cls):\n \"\"\"stuff\"\"\"\n\n return cls.get_timestamp(cls.get_tomorrow())\n\n @classmethod\n def get_yesterday_timestamp(cls):\n\n return cls.get_timestamp(cls.get_yesterday())\n","subject":"Add utility methods for yesterday's date","message":"Add utility methods for yesterday's date\n","lang":"Python","license":"mit","repos":"pantheon-systems\/etl-framework"} {"commit":"26bb374b00d667de00a080c4b32e102ac69a0e23","old_file":"asn1crypto\/version.py","new_file":"asn1crypto\/version.py","old_contents":"# coding: utf-8\nfrom __future__ import unicode_literals, division, absolute_import, print_function\n\n\n__version__ = '0.24.0'\n__version_info__ = (0, 24, 0)\n","new_contents":"# coding: utf-8\nfrom __future__ import unicode_literals, division, absolute_import, print_function\n\n\n__version__ = '0.25.0-alpha'\n__version_info__ = (0, 25, 0, 'alpha')\n","subject":"Mark master as working towards 0.25.0","message":"Mark master as working towards 0.25.0\n","lang":"Python","license":"mit","repos":"wbond\/asn1crypto"} {"commit":"c52a959896c345b57fdd28e2ae8cbd75ab2e3c71","old_file":"fuzzinator\/call\/file_reader_decorator.py","new_file":"fuzzinator\/call\/file_reader_decorator.py","old_contents":"# Copyright (c) 2017-2018 Renata Hodovan, Akos Kiss.\n#\n# Licensed under the BSD 3-Clause License\n# <LICENSE.rst or https:\/\/opensource.org\/licenses\/BSD-3-Clause>.\n# This file may not be copied, modified, or distributed except\n# according to those terms.\n\nimport os\n\nfrom . import CallableDecorator\n\n\nclass FileReaderDecorator(CallableDecorator):\n \"\"\"\n Decorator for SUTs that take input as a file path: saves the content of\n the failing test case.\n\n Moreover, the issue (if any) is also extended with the new ``'filename'``\n property containing the name of the test case (as received in the ``test``\n argument).\n\n **Example configuration snippet:**\n\n .. code-block:: ini\n\n [sut.foo]\n call=fuzzinator.call.SubprocessCall\n call.decorate(0)=fuzzionator.call.FileReaderDecorator\n\n [sut.foo.call]\n # assuming that foo takes one file as input specified on command line\n command=\/home\/alice\/foo\/bin\/foo {test}\n \"\"\"\n\n def decorator(self, **kwargs):\n def wrapper(fn):\n def reader(*args, **kwargs):\n issue = fn(*args, **kwargs)\n\n if issue is not None:\n with open(kwargs['test'], 'rb') as f:\n issue['filename'] = os.path.basename(kwargs['test'])\n issue['test'] = f.read()\n\n return issue\n\n return reader\n return wrapper\n","new_contents":"# Copyright (c) 2017-2018 Renata Hodovan, Akos Kiss.\n#\n# Licensed under the BSD 3-Clause License\n# <LICENSE.rst or https:\/\/opensource.org\/licenses\/BSD-3-Clause>.\n# This file may not be copied, modified, or distributed except\n# according to those terms.\n\nimport os\n\nfrom . import CallableDecorator\n\n\nclass FileReaderDecorator(CallableDecorator):\n \"\"\"\n Decorator for SUTs that take input as a file path: saves the content of\n the failing test case.\n\n Moreover, the issue (if any) is also extended with the new ``'filename'``\n property containing the name of the test case (as received in the ``test``\n argument).\n\n **Example configuration snippet:**\n\n .. code-block:: ini\n\n [sut.foo]\n call=fuzzinator.call.SubprocessCall\n call.decorate(0)=fuzzinator.call.FileReaderDecorator\n\n [sut.foo.call]\n # assuming that foo takes one file as input specified on command line\n command=\/home\/alice\/foo\/bin\/foo {test}\n \"\"\"\n\n def decorator(self, **kwargs):\n def wrapper(fn):\n def reader(*args, **kwargs):\n issue = fn(*args, **kwargs)\n\n if issue is not None:\n with open(kwargs['test'], 'rb') as f:\n issue['filename'] = os.path.basename(kwargs['test'])\n issue['test'] = f.read()\n\n return issue\n\n return reader\n return wrapper\n","subject":"Fix a typo in the documentation of FileReaderDecorator.","message":"Fix a typo in the documentation of FileReaderDecorator.\n","lang":"Python","license":"bsd-3-clause","repos":"renatahodovan\/fuzzinator,akosthekiss\/fuzzinator,renatahodovan\/fuzzinator,akosthekiss\/fuzzinator,akosthekiss\/fuzzinator,renatahodovan\/fuzzinator,renatahodovan\/fuzzinator,akosthekiss\/fuzzinator"} {"commit":"cbd913af9013926ca7f08ab56023d7242e783698","old_file":"ad-hoc-scripts\/latex-adjust.py","new_file":"ad-hoc-scripts\/latex-adjust.py","old_contents":"#! \/usr\/bin\/env python3\n\nimport sys\nimport json\n\nfor arg in sys.argv[1:]:\n with open(arg) as f:\n equajson = json.load(f)\n\n try:\n latex = equajson[\"markup-languages\"][\"LaTeX\"][0][\"markup\"]\n except KeyError:\n continue\n\n if 'documentclass' not in latex:\n with_boilerplate = \"\\\\documentclass{article}\\n\\\\begin{document}\\n\\\\[\\n%s\\n\\\\]\\n\\\\end{document}\" % latex\n equajson[\"markup-languages\"][\"LaTeX\"][0][\"markup\"] = with_boilerplate\n\n with open(arg, 'w') as f:\n json.dump(equajson, f, indent=4, separators=(',', ': '), ensure_ascii=False, sort_keys=True)\n","new_contents":"#! \/usr\/bin\/env python3\n\nimport sys\nimport json\n\nfor arg in sys.argv[1:]:\n with open(arg) as f:\n equajson = json.load(f)\n\n try:\n latex = equajson[\"markup-languages\"][\"LaTeX\"][0][\"markup\"]\n except KeyError:\n continue\n\n if 'documentclass' not in latex:\n with_boilerplate = \"\\\\documentclass{article}\\n\\\\begin{document}\\n\\\\[\\n%s\\n\\\\]\\n\\\\end{document}\" % latex\n equajson[\"markup-languages\"][\"LaTeX\"][0][\"markup\"] = with_boilerplate\n\n with open(arg, 'w') as f:\n json.dump(equajson, f, indent=4, separators=(',', ': '), ensure_ascii=False, sort_keys=True)\n f.write('\\n') # add trailing newline\n","subject":"Add trailing newline to make round-tripping without diffs possible.","message":"Add trailing newline to make round-tripping without diffs possible.\n","lang":"Python","license":"mit","repos":"nbeaver\/equajson"} {"commit":"ace54e86e9462b25acd1636e0e9905ba6decfe9b","old_file":"admin_tools\/dashboard\/views.py","new_file":"admin_tools\/dashboard\/views.py","old_contents":"from django.contrib.auth.decorators import login_required\nfrom django.http import HttpResponse\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response\nfrom django.contrib import messages\n\ntry:\n from django.views.decorators.csrf import csrf_exempt\nexcept ImportError:\n from django.contrib.csrf.middleware import csrf_exempt\n\nfrom .forms import DashboardPreferencesForm\nfrom .models import DashboardPreferences\n\n\n@login_required\n@csrf_exempt\ndef set_preferences(request, dashboard_id):\n \"\"\"\n This view serves and validates a preferences form.\n \"\"\"\n try:\n preferences = DashboardPreferences.objects.get(\n user=request.user,\n dashboard_id=dashboard_id\n )\n except DashboardPreferences.DoesNotExist:\n preferences = None\n if request.method == \"POST\":\n form = DashboardPreferencesForm(\n user=request.user,\n dashboard_id=dashboard_id,\n data=request.POST,\n instance=preferences\n )\n if form.is_valid():\n preferences = form.save()\n if request.is_ajax():\n return HttpResponse('true')\n messages.success(request, 'Preferences saved')\n elif request.is_ajax():\n return HttpResponse('false')\n else:\n form = DashboardPreferencesForm(\n user=request.user,\n dashboard_id=dashboard_id,\n instance=preferences\n )\n return render_to_response('admin_tools\/dashboard\/preferences_form.html',\n RequestContext(request, {'form': form}))\n","new_contents":"from django.contrib.admin.views.decorators import staff_member_required\nfrom django.http import HttpResponse\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response\nfrom django.contrib import messages\n\ntry:\n from django.views.decorators.csrf import csrf_exempt\nexcept ImportError:\n from django.contrib.csrf.middleware import csrf_exempt\n\nfrom .forms import DashboardPreferencesForm\nfrom .models import DashboardPreferences\n\n\n@staff_member_required\n@csrf_exempt\ndef set_preferences(request, dashboard_id):\n \"\"\"\n This view serves and validates a preferences form.\n \"\"\"\n try:\n preferences = DashboardPreferences.objects.get(\n user=request.user,\n dashboard_id=dashboard_id\n )\n except DashboardPreferences.DoesNotExist:\n preferences = None\n if request.method == \"POST\":\n form = DashboardPreferencesForm(\n user=request.user,\n dashboard_id=dashboard_id,\n data=request.POST,\n instance=preferences\n )\n if form.is_valid():\n preferences = form.save()\n if request.is_ajax():\n return HttpResponse('true')\n messages.success(request, 'Preferences saved')\n elif request.is_ajax():\n return HttpResponse('false')\n else:\n form = DashboardPreferencesForm(\n user=request.user,\n dashboard_id=dashboard_id,\n instance=preferences\n )\n return render_to_response('admin_tools\/dashboard\/preferences_form.html',\n RequestContext(request, {'form': form}))\n","subject":"Use @staff_member_required decorator for the dashboard view as well","message":"Use @staff_member_required decorator for the dashboard view as well\n","lang":"Python","license":"mit","repos":"django-admin-tools\/django-admin-tools,django-admin-tools\/django-admin-tools,django-admin-tools\/django-admin-tools"} {"commit":"3219a925ecddbacb39e4adc484d94eaed6bddd0b","old_file":"yolk\/__init__.py","new_file":"yolk\/__init__.py","old_contents":"\"\"\"yolk.\n\nAuthor: Rob Cakebread <cakebread at gmail>\n\nLicense : BSD\n\n\"\"\"\n\n__version__ = '0.8.6'\n","new_contents":"\"\"\"yolk.\n\nAuthor: Rob Cakebread <cakebread at gmail>\n\nLicense : BSD\n\n\"\"\"\n\n__version__ = '0.8.7'\n","subject":"Increment patch version to 0.8.7","message":"Increment patch version to 0.8.7\n","lang":"Python","license":"bsd-3-clause","repos":"myint\/yolk,myint\/yolk"} {"commit":"badcdcc03517aaf705975676a5d37488b38c9738","old_file":"foomodules\/link_harvester\/common_handlers.py","new_file":"foomodules\/link_harvester\/common_handlers.py","old_contents":"import logging\nimport re\nimport socket\nimport urllib\nfrom bs4 import BeautifulSoup\n\nlogger = logging.getLogger(__name__)\nWURSTBALL_RE = re.compile(r\"^https?:\/\/(www\\.)?wurstball\\.de\/[0-9]+\/\")\n\n\ndef default_handler(metadata):\n return {key: getattr(metadata, key) for key in\n [\"original_url\", \"url\", \"title\", \"description\",\n \"human_readable_type\"]}\n\n\ndef wurstball_handler(metadata):\n if not WURSTBALL_RE.match(metadata.url):\n return None\n\n ret = default_handler(metadata)\n\n soup = BeautifulSoup(metadata.buf)\n img_url = soup.find(id=\"content-main\").img[\"src\"]\n\n try:\n response = urllib.request.urlopen(img_url, timeout=5)\n img_data = response.read()\n except (socket.timeout,\n urllib.error.URLError,\n urllib.error.HTTPError) as err:\n logger.warn(\"Could not download Wurstball image: {}\".format(err))\n return ret\n\n mime_type = response.getheader(\"Content-Type\")\n\n ret.update({\"image_mime_type\": mime_type,\n \"image_buffer\": img_data,\n \"image_url\": img_url,\n \"title\": None,\n \"description\": None})\n\n return ret\n","new_contents":"import logging\nimport re\nimport socket\nimport urllib\nimport http.client\nfrom bs4 import BeautifulSoup\n\nlogger = logging.getLogger(__name__)\nWURSTBALL_RE = re.compile(r\"^https?:\/\/(www\\.)?wurstball\\.de\/[0-9]+\/\")\n\n\ndef default_handler(metadata):\n return {key: getattr(metadata, key) for key in\n [\"original_url\", \"url\", \"title\", \"description\",\n \"human_readable_type\"]}\n\n\ndef wurstball_handler(metadata):\n if not WURSTBALL_RE.match(metadata.url):\n return None\n\n ret = default_handler(metadata)\n\n soup = BeautifulSoup(metadata.buf)\n img_url = soup.find(id=\"content-main\").img[\"src\"]\n\n try:\n response = urllib.request.urlopen(img_url, timeout=5)\n img_data = response.read()\n except (socket.timeout,\n urllib.error.URLError,\n urllib.error.HTTPError) as err:\n logger.warn(\"Could not download Wurstball image: {}\".format(err))\n return ret\n\n mime_type = response.getheader(\"Content-Type\")\n\n ret.update({\"image_mime_type\": mime_type,\n \"image_buffer\": img_data,\n \"image_url\": img_url,\n \"title\": None,\n \"description\": None})\n\n return ret\n\n\ndef image_handler(metadata):\n if not metadata.mime_type.startswith(\"image\/\"):\n return None\n\n ret = default_handler(metadata)\n\n try:\n img_data = metadata.buf + metadata.response.read()\n except http.client.IncompleteRead as err:\n logger.warn(\"Could not download image: {}\".format(err))\n return ret\n\n ret.update({\"image_mime_type\": metadata.mime_type,\n \"image_buffer\": img_data,\n \"image_url\": metadata.url})\n\n return ret\n","subject":"Add image_handler for link harvester","message":"Add image_handler for link harvester\n","lang":"Python","license":"mit","repos":"horazont\/xmpp-crowd"} {"commit":"ee31e6c0302c6840d522666b1f724d0ec429d562","old_file":"monasca_setup\/detection\/plugins\/neutron.py","new_file":"monasca_setup\/detection\/plugins\/neutron.py","old_contents":"import monasca_setup.detection\n\n\nclass Neutron(monasca_setup.detection.ServicePlugin):\n\n \"\"\"Detect Neutron daemons and setup configuration to monitor them.\n\n \"\"\"\n\n def __init__(self, template_dir, overwrite=True, args=None):\n service_params = {\n 'args': args,\n 'template_dir': template_dir,\n 'overwrite': overwrite,\n 'service_name': 'networking',\n 'process_names': ['neutron-server', 'neutron-openvswitch-agent',\n 'neutron-rootwrap', 'neutron-dhcp-agent',\n 'neutron-vpn-agent', 'neutron-metadata-agent',\n 'neutron-metering-agent', 'neutron-l3-agent',\n 'neutron-ns-metadata-proxy'],\n 'service_api_url': 'http:\/\/localhost:9696',\n 'search_pattern': '.*v2.0.*'\n }\n\n super(Neutron, self).__init__(service_params)\n\n def build_config(self):\n \"\"\"Build the config as a Plugins object and return.\"\"\"\n # Skip the http check if neutron-server is not on this box\n if 'neutron-server' not in self.found_processes:\n self.service_api_url = None\n self.search_pattern = None\n\n return monasca_setup.detection.ServicePlugin.build_config(self)\n","new_contents":"import monasca_setup.detection\n\n\nclass Neutron(monasca_setup.detection.ServicePlugin):\n\n \"\"\"Detect Neutron daemons and setup configuration to monitor them.\n\n \"\"\"\n\n def __init__(self, template_dir, overwrite=True, args=None):\n service_params = {\n 'args': args,\n 'template_dir': template_dir,\n 'overwrite': overwrite,\n 'service_name': 'networking',\n 'process_names': ['neutron-server', 'neutron-openvswitch-agent',\n 'neutron-rootwrap', 'neutron-dhcp-agent',\n 'neutron-vpn-agent', 'neutron-metadata-agent',\n 'neutron-metering-agent', 'neutron-l3-agent',\n 'neutron-ns-metadata-proxy',\n '\/opt\/stack\/service\/neutron\/venv\/bin\/neutron-lbaas-agent',\n '\/opt\/stack\/service\/neutron\/venv\/bin\/neutron-lbaasv2-agent'],\n 'service_api_url': 'http:\/\/localhost:9696',\n 'search_pattern': '.*v2.0.*'\n }\n\n super(Neutron, self).__init__(service_params)\n\n def build_config(self):\n \"\"\"Build the config as a Plugins object and return.\"\"\"\n # Skip the http check if neutron-server is not on this box\n if 'neutron-server' not in self.found_processes:\n self.service_api_url = None\n self.search_pattern = None\n\n return monasca_setup.detection.ServicePlugin.build_config(self)\n","subject":"Add process monitoring for LBaaS agents","message":"Add process monitoring for LBaaS agents\n\nAdd neutron-lbaas-agent (LBaaS V1) and neutron-lbaasv2-agent (LBaaS\nV2) to the neutron detection plugin. Because the string\n\"neutron-lbaas-agent\" can be both a process name and log file name,\nthe process monitor is susceptible to false positive matching on that\nstring. Use a longer part of the python path to disambiguate this\n\nChange-Id: I3081639a6f36a276bab2f9eb1b9b39a5bef452f1\n","lang":"Python","license":"bsd-3-clause","repos":"sapcc\/monasca-agent,sapcc\/monasca-agent,sapcc\/monasca-agent"} {"commit":"8769224d8dbe73e177d19012d54c9bb7e114a3fa","old_file":"recipes\/webrtc.py","new_file":"recipes\/webrtc.py","old_contents":"# Copyright (c) 2014 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nimport sys\n\nimport recipe_util # pylint: disable=F0401\n\n\n# This class doesn't need an __init__ method, so we disable the warning\n# pylint: disable=W0232\nclass WebRTC(recipe_util.Recipe):\n \"\"\"Basic Recipe class for WebRTC.\"\"\"\n\n @staticmethod\n def fetch_spec(props):\n url = 'https:\/\/chromium.googlesource.com\/external\/webrtc.git'\n spec = {\n 'solutions': [\n {\n 'name': 'src',\n 'url': url,\n 'deps_file': 'DEPS',\n 'managed': False,\n 'custom_deps': {},\n 'safesync_url': '',\n },\n ],\n 'auto': True, # Runs git auto-svn as a part of the fetch.\n 'with_branch_heads': True,\n }\n\n if props.get('target_os'):\n spec['target_os'] = props['target_os'].split(',')\n\n return {\n 'type': 'gclient_git_svn',\n 'gclient_git_svn_spec': spec,\n }\n\n @staticmethod\n def expected_root(_props):\n return 'src'\n\n\ndef main(argv=None):\n return WebRTC().handle_args(argv)\n\n\nif __name__ == '__main__':\n sys.exit(main(sys.argv))\n","new_contents":"# Copyright (c) 2014 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nimport sys\n\nimport recipe_util # pylint: disable=F0401\n\n\n# This class doesn't need an __init__ method, so we disable the warning\n# pylint: disable=W0232\nclass WebRTC(recipe_util.Recipe):\n \"\"\"Basic Recipe class for WebRTC.\"\"\"\n\n @staticmethod\n def fetch_spec(props):\n url = 'https:\/\/chromium.googlesource.com\/external\/webrtc.git'\n spec = {\n 'solutions': [\n {\n 'name': 'src',\n 'url': url,\n 'deps_file': 'DEPS',\n 'managed': False,\n 'custom_deps': {},\n 'safesync_url': '',\n },\n ],\n 'with_branch_heads': True,\n }\n\n if props.get('target_os'):\n spec['target_os'] = props['target_os'].split(',')\n\n return {\n 'type': 'gclient_git',\n 'gclient_git_spec': spec,\n }\n\n @staticmethod\n def expected_root(_props):\n return 'src'\n\n\ndef main(argv=None):\n return WebRTC().handle_args(argv)\n\n\nif __name__ == '__main__':\n sys.exit(main(sys.argv))\n","subject":"Switch WebRTC recipe to Git.","message":"Switch WebRTC recipe to Git.\n\nBUG=412012\n\nReview URL: https:\/\/codereview.chromium.org\/765373002\n\ngit-svn-id: fd409f4bdeea2bb50a5d34bb4d4bfc2046a5a3dd@294546 0039d316-1c4b-4281-b951-d872f2087c98\n","lang":"Python","license":"bsd-3-clause","repos":"sarvex\/depot-tools,fracting\/depot_tools,sarvex\/depot-tools,azunite\/chrome_build,disigma\/depot_tools,duongbaoduy\/gtools,fracting\/depot_tools,hsharsha\/depot_tools,Midrya\/chromium,hsharsha\/depot_tools,ajohnson23\/depot_tools,gcodetogit\/depot_tools,npe9\/depot_tools,mlufei\/depot_tools,primiano\/depot_tools,chinmaygarde\/depot_tools,primiano\/depot_tools,azunite\/chrome_build,SuYiling\/chrome_depot_tools,duanwujie\/depot_tools,withtone\/depot_tools,chinmaygarde\/depot_tools,duongbaoduy\/gtools,SuYiling\/chrome_depot_tools,azureplus\/chromium_depot_tools,gcodetogit\/depot_tools,kaiix\/depot_tools,gcodetogit\/depot_tools,sarvex\/depot-tools,SuYiling\/chrome_depot_tools,aleonliao\/depot_tools,liaorubei\/depot_tools,mlufei\/depot_tools,Midrya\/chromium,primiano\/depot_tools,npe9\/depot_tools,liaorubei\/depot_tools,sarvex\/depot-tools,withtone\/depot_tools,disigma\/depot_tools,azureplus\/chromium_depot_tools,disigma\/depot_tools,hsharsha\/depot_tools,cpanelli\/-git-clone-https-chromium.googlesource.com-chromium-tools-depot_tools,kaiix\/depot_tools,cpanelli\/-git-clone-https-chromium.googlesource.com-chromium-tools-depot_tools,liaorubei\/depot_tools,CoherentLabs\/depot_tools,cpanelli\/-git-clone-https-chromium.googlesource.com-chromium-tools-depot_tools,ajohnson23\/depot_tools,withtone\/depot_tools,azureplus\/chromium_depot_tools,duanwujie\/depot_tools,liaorubei\/depot_tools,aleonliao\/depot_tools,fracting\/depot_tools,kaiix\/depot_tools,npe9\/depot_tools,CoherentLabs\/depot_tools,mlufei\/depot_tools,npe9\/depot_tools,aleonliao\/depot_tools,ajohnson23\/depot_tools,duongbaoduy\/gtools,duanwujie\/depot_tools,azunite\/chrome_build,chinmaygarde\/depot_tools,Midrya\/chromium"} {"commit":"2393b066fbb0fc88d9e9a1918485cf57c40aecc2","old_file":"opps\/articles\/templatetags\/article_tags.py","new_file":"opps\/articles\/templatetags\/article_tags.py","old_contents":"# -*- coding: utf-8 -*-\nfrom django import template\nfrom django.conf import settings\nfrom opps.articles.models import ArticleBox\n\n\nregister = template.Library()\n\n\n@register.simple_tag\ndef get_articlebox(slug, channel_slug=None, template_name=None):\n if channel_slug:\n slug = slug + '-' + channel_slug\n\n try:\n box = ArticleBox.objects.get(site=settings.SITE_ID, slug=slug)\n except ArticleBox.DoesNotExist:\n box = None\n\n t = template.loader.get_template('articles\/articlebox_detail.html')\n if template_name:\n t = template.loader.get_template(template_name)\n\n return t.render(template.Context({'articlebox': box, 'slug': slug}))\n\n\n@register.simple_tag\ndef get_all_articlebox(channel_slug, template_name=None):\n boxes = ArticleBox.objects.filter(site=settings.SITE_ID,\n channel__slug=channel_slug)\n\n t = template.loader.get_template('articles\/articlebox_list.html')\n if template_name:\n t = template.loader.get_template(template_name)\n\n return t.render(template.Context({'articleboxes': boxes}))\n","new_contents":"# -*- coding: utf-8 -*-\nfrom django import template\nfrom django.conf import settings\nfrom django.utils import timezone\nfrom opps.articles.models import ArticleBox\n\n\nregister = template.Library()\n\n\n@register.simple_tag\ndef get_articlebox(slug, channel_slug=None, template_name=None):\n if channel_slug:\n slug = slug + '-' + channel_slug\n\n try:\n box = ArticleBox.objects.get(site=settings.SITE_ID, slug=slug,\n date_available__lte=timezone.now(),\n published=True)\n except ArticleBox.DoesNotExist:\n box = None\n\n t = template.loader.get_template('articles\/articlebox_detail.html')\n if template_name:\n t = template.loader.get_template(template_name)\n\n return t.render(template.Context({'articlebox': box, 'slug': slug}))\n\n\n@register.simple_tag\ndef get_all_articlebox(channel_slug, template_name=None):\n boxes = ArticleBox.objects.filter(site=settings.SITE_ID,\n channel__slug=channel_slug)\n\n t = template.loader.get_template('articles\/articlebox_list.html')\n if template_name:\n t = template.loader.get_template(template_name)\n\n return t.render(template.Context({'articleboxes': boxes}))\n","subject":"Add validate published on templatetag get articlebox","message":"Add validate published on templatetag get articlebox\n","lang":"Python","license":"mit","repos":"jeanmask\/opps,williamroot\/opps,opps\/opps,jeanmask\/opps,jeanmask\/opps,YACOWS\/opps,williamroot\/opps,YACOWS\/opps,jeanmask\/opps,opps\/opps,opps\/opps,YACOWS\/opps,opps\/opps,williamroot\/opps,williamroot\/opps,YACOWS\/opps"} {"commit":"8f9dc9a241515f9cab633f33b9d2243f76df55bd","old_file":"emencia_paste_djangocms_3\/django_buildout\/project\/utils\/templatetags\/utils_addons.py","new_file":"emencia_paste_djangocms_3\/django_buildout\/project\/utils\/templatetags\/utils_addons.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nVarious usefull tags\n\"\"\"\nfrom django import template\n\nregister = template.Library()\n\n@register.filter(name='split', is_safe=True)\ndef split_string(value, arg=None):\n \"\"\"\n A simple string splitter\n \n So you can do that : ::\n \n {% if LANGUAGE_CODE in \"fr,en-ca,en-gb,zh-hk,it,en,de\"|split:',' %}\n \"\"\"\n return value.split(arg)","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nVarious usefull tags\n\"\"\"\nfrom django import template\n\nregister = template.Library()\n\n@register.filter(name='split', is_safe=False)\ndef split_string(value, arg=None):\n \"\"\"\n A simple string splitter\n \n So you can do that : ::\n \n {% if LANGUAGE_CODE in \"fr,en-ca,en-gb,zh-hk,it,en,de\"|split:',' %}\n \"\"\"\n return value.split(arg)","subject":"Fix split filter in emencia_utils templatetags that was returning a string instead of a list","message":"Fix split filter in emencia_utils templatetags that was returning a string instead of a list\n","lang":"Python","license":"mit","repos":"emencia\/emencia_paste_djangocms_3,emencia\/emencia_paste_djangocms_3,emencia\/emencia_paste_djangocms_3,emencia\/emencia_paste_djangocms_3"} {"commit":"d7878a798d8208bcd9221babcd3ac1a5c12aa9f7","old_file":"drivnal\/object.py","new_file":"drivnal\/object.py","old_contents":"from constants import *\nimport os\nimport urllib\nimport mimetypes\nimport logging\n\nclass Object:\n def __init__(self, path):\n self.name = os.path.basename(path)\n self.path = path\n\n if os.path.isdir(self.path):\n self.type = DIR_MIME_TYPE\n self.size = None\n self.time = None\n else:\n try:\n stat = os.stat(self.path)\n self.size = stat.st_size\n self.time = stat.st_mtime\n except OSError:\n self.size = None\n self.time = None\n\n mime = mimetypes.MimeTypes()\n self.type = mime.guess_type(urllib.pathname2url(self.path))[0]\n\n @staticmethod\n def get_objects(path):\n objects = []\n\n if path:\n for name in os.listdir(path):\n object_path = os.path.join(path, name)\n objects.append(Object(object_path))\n\n return objects\n","new_contents":"from constants import *\nimport os\nimport urllib\nimport subprocess\nimport logging\n\nlogger = logging.getLogger(APP_NAME)\n\nclass Object:\n def __init__(self, path):\n self.name = os.path.basename(path)\n self.path = path\n\n if os.path.isdir(self.path):\n self.type = DIR_MIME_TYPE\n self.size = None\n self.time = None\n else:\n try:\n stat = os.stat(self.path)\n self.size = stat.st_size\n self.time = stat.st_mtime\n except OSError:\n self.size = None\n self.time = None\n self.type = None\n\n @staticmethod\n def get_objects(path):\n objects = []\n object_paths = []\n\n if path:\n for name in os.listdir(path):\n object_path = os.path.join(path, name)\n object = Object(object_path)\n objects.append(object)\n if not object.type:\n object_paths.append(object_path)\n\n try:\n # TODO Follow symlinks\n mime_types = subprocess.check_output(['file',\n '--mime-type', '--brief'] + object_paths).splitlines()\n except subprocess.CalledProcessError, error:\n logger.warning('File mime-type call failed. %r' % {\n 'return_code': error.returncode,\n 'output': error.output,\n })\n\n try:\n for object in objects:\n if not object.type:\n object.type = mime_types.pop(0)\n except IndexError:\n logger.error('File mime-type call index error.')\n\n return objects\n","subject":"Improve file mime type detection","message":"Improve file mime type detection\n","lang":"Python","license":"agpl-3.0","repos":"drivnal\/drivnal,drivnal\/drivnal,drivnal\/drivnal"} {"commit":"3ddeeccabb09f11fdfb60d9ddbddce406a054e50","old_file":"settings.py","new_file":"settings.py","old_contents":"from settings_common import *\n\nPACKAGE_VERSION = 0.5\n\nDEBUG = TEMPLATE_DEBUG = True\n\nDATABASE_ENGINE = 'postgresql_psycopg2'\nDATABASE_NAME = 'daisyproducer_dev'\nDATABASE_USER = 'eglic'\nDATABASE_PASSWORD = ''\n\nDAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', 'tmp', 'pipeline-20100301')\nDTBOOK2SBSFORM_PATH = os.path.join(PROJECT_DIR, '..', '..', 'workspace', 'LiblouisSaxonExtension')\n\n# debug toolbar\n#INSTALLED_APPS += ('debug_toolbar',)\n#MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',)\nINTERNAL_IPS = ('127.0.0.1',)\nDEBUG_TOOLBAR_CONFIG = {'INTERCEPT_REDIRECTS' : False}\nSERVE_STATIC_FILES = True\n","new_contents":"from settings_common import *\n\nPACKAGE_VERSION = 0.5\n\nDEBUG = TEMPLATE_DEBUG = True\n\nDATABASE_ENGINE = 'postgresql_psycopg2'\nDATABASE_NAME = 'daisyproducer_dev'\nDATABASE_USER = 'eglic'\nDATABASE_PASSWORD = ''\n\nDAISY_PIPELINE_PATH = os.path.join(PROJECT_DIR, '..', '..', 'tmp', 'pipeline-20110106')\nDTBOOK2SBSFORM_PATH = os.path.join(PROJECT_DIR, '..', '..', 'workspace', 'LiblouisSaxonExtension')\n\n# debug toolbar\n#INSTALLED_APPS += ('debug_toolbar',)\n#MIDDLEWARE_CLASSES += ('debug_toolbar.middleware.DebugToolbarMiddleware',)\nINTERNAL_IPS = ('127.0.0.1',)\nDEBUG_TOOLBAR_CONFIG = {'INTERCEPT_REDIRECTS' : False}\nSERVE_STATIC_FILES = True\n","subject":"Upgrade to a newer pipeline release","message":"Upgrade to a newer pipeline release\n","lang":"Python","license":"agpl-3.0","repos":"sbsdev\/daisyproducer,sbsdev\/daisyproducer,sbsdev\/daisyproducer,sbsdev\/daisyproducer"} {"commit":"109fc84cb307083f6a01317bb5b5bea0578088d3","old_file":"bloop\/__init__.py","new_file":"bloop\/__init__.py","old_contents":"from bloop.engine import Engine, ObjectsNotFound, ConstraintViolation\nfrom bloop.column import Column, GlobalSecondaryIndex, LocalSecondaryIndex\nfrom bloop.types import (\n String, Float, Integer, Binary, StringSet, FloatSet,\n IntegerSet, BinarySet, Null, Boolean, Map, List\n)\n\n__all__ = [\n \"Engine\", \"ObjectsNotFound\", \"ConstraintViolation\",\n \"Column\", \"GlobalSecondaryIndex\", \"LocalSecondaryIndex\",\n \"String\", \"Float\", \"Integer\", \"Binary\", \"StringSet\", \"FloatSet\",\n \"IntegerSet\", \"BinarySet\", \"Null\", \"Boolean\", \"Map\", \"List\"\n]\n","new_contents":"from bloop.engine import Engine, ObjectsNotFound, ConstraintViolation\nfrom bloop.column import Column, GlobalSecondaryIndex, LocalSecondaryIndex\nfrom bloop.types import (\n String, UUID, Float, Integer, Binary, StringSet, FloatSet,\n IntegerSet, BinarySet, Null, Boolean, Map, List\n)\n\n__all__ = [\n \"Engine\", \"ObjectsNotFound\", \"ConstraintViolation\",\n \"Column\", \"GlobalSecondaryIndex\", \"LocalSecondaryIndex\",\n \"String\", \"UUID\", \"Float\", \"Integer\", \"Binary\", \"StringSet\", \"FloatSet\",\n \"IntegerSet\", \"BinarySet\", \"Null\", \"Boolean\", \"Map\", \"List\"\n]\n","subject":"Add UUID to bloop __all__","message":"Add UUID to bloop __all__","lang":"Python","license":"mit","repos":"numberoverzero\/bloop,numberoverzero\/bloop"} {"commit":"5c11731b445df04e1b4ec92df4ff6b7e6681915b","old_file":"testMail.py","new_file":"testMail.py","old_contents":"#!\/usr\/local\/bin\/python\nimport smtplib, time, threading, sys\nfrom email.mime.text import MIMEText\n\nfromaddr = sys.argv[0]\ntoaddr = sys.argv[1]\n\n\ndef createMessage(fromaddr, toaddr, subject, msgtxt):\n msg = MIMEText(msgtxt)\n msg['Subject'] = subject\n msg['From'] = fromaddr\n msg['To'] = toaddr\n return msg\n\ndef sendMails(threadId):\n server = smtplib.SMTP('localhost', 8001)\n for i in xrange(25):\n server.sendmail(fromaddr, [toaddr], createMessage(fromaddr, toaddr, \"This is from thread %s\" % threadId, \"Some header\" ).as_string())\n server.quit()\n\nthreads = [threading.Thread(target=sendMails, args=(i,)) for i in range(10)]\n\nfor t in threads:\n t.start()\n\nfor t in threads:\n t.join()\n \n \n\n","new_contents":"#!\/usr\/local\/bin\/python\nimport smtplib, time, threading, sys\nfrom email.mime.text import MIMEText\n\nfromaddr = sys.argv[1]\ntoaddr = sys.argv[2]\n\n\ndef createMessage(fromaddr, toaddr, subject, msgtxt):\n msg = MIMEText(msgtxt)\n msg['Subject'] = subject\n msg['From'] = fromaddr\n msg['To'] = toaddr\n return msg\n\ndef sendMails(threadId):\n server = smtplib.SMTP('localhost', 8001)\n for i in xrange(25):\n server.sendmail(fromaddr, [toaddr], createMessage(fromaddr, toaddr, \"This is from thread %s\" % threadId, \"Some header\" ).as_string())\n server.quit()\n\nthreads = [threading.Thread(target=sendMails, args=(i,)) for i in range(10)]\n\nfor t in threads:\n t.start()\n\nfor t in threads:\n t.join()\n \n \n\n","subject":"Change the arg values so not to use the script name as the fromaddr","message":"Change the arg values so not to use the script name as the fromaddr","lang":"Python","license":"bsd-3-clause","repos":"bobbynewmark\/mailthrottler,bobbynewmark\/mailthrottler"} {"commit":"20f3ba97d9d78c076d3c6c7d5b95e6ac76ee79a4","old_file":"telepathy\/_version.py","new_file":"telepathy\/_version.py","old_contents":"__all__ = ('version', '__version__')\n\nversion = (0, 13, 9, 1)\n__version__ = '.'.join(str(x) for x in version)\n","new_contents":"__all__ = ('version', '__version__')\n\nversion = (0, 13, 10)\n__version__ = '.'.join(str(x) for x in version)\n","subject":"Bump version to 0.13.10 - the 0.13.9 release seems to be missing some of the code?","message":"Bump version to 0.13.10 - the 0.13.9 release seems to be missing some of the code?\n\n\n20070514164432-53eee-220b5156a45bf041cb773bd829c27ac9ff8932a4.gz\n","lang":"Python","license":"lgpl-2.1","repos":"epage\/telepathy-python,max-posedon\/telepathy-python,epage\/telepathy-python,PabloCastellano\/telepathy-python,freedesktop-unofficial-mirror\/telepathy__telepathy-python,detrout\/telepathy-python,PabloCastellano\/telepathy-python,detrout\/telepathy-python,max-posedon\/telepathy-python,freedesktop-unofficial-mirror\/telepathy__telepathy-python"} {"commit":"5e1b06a5005c994acd25793b5a5527430a8c72d4","old_file":"dimod\/package_info.py","new_file":"dimod\/package_info.py","old_contents":"# Copyright 2018 D-Wave Systems Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# ================================================================================================\n\n__version__ = '0.8.4'\n__author__ = 'D-Wave Systems Inc.'\n__authoremail__ = 'acondello@dwavesys.com'\n__description__ = 'A shared API for binary quadratic model samplers.'\n","new_contents":"# Copyright 2018 D-Wave Systems Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# ================================================================================================\n\n__version__ = '0.8.5'\n__author__ = 'D-Wave Systems Inc.'\n__authoremail__ = 'acondello@dwavesys.com'\n__description__ = 'A shared API for binary quadratic model samplers.'\n","subject":"Update version 0.8.4 -> 0.8.5","message":"Update version 0.8.4 -> 0.8.5\n\nNew Features\r\n------------\r\n* `SampleSet.truncate` method\r\n\r\nFixes\r\n-----\r\n* `BinaryQuadraticModel`s are now pickleable in python2.7\r\n* Some bug fixes and new parameters in `BinaryQuadraticModel.to_serializable`","lang":"Python","license":"apache-2.0","repos":"dwavesystems\/dimod,dwavesystems\/dimod"} {"commit":"c9850501365d086d0c4e8b89bcf89e802d5d8103","old_file":"gooey\/gui\/components\/widgets\/choosers.py","new_file":"gooey\/gui\/components\/widgets\/choosers.py","old_contents":"from gooey.gui.components.widgets import core\r\nfrom gooey.gui.components.widgets.bases import TextContainer, BaseChooser\r\n\r\n\r\n__ALL__ = [\r\n 'FileChooser',\r\n 'FileSaver',\r\n 'DirChooser',\r\n 'MultiDirChooser',\r\n 'DateChooser',\r\n 'ColourChooser'\r\n]\r\n\r\nclass FileChooser(BaseChooser):\r\n # todo: allow wildcard from argparse\r\n widget_class = core.FileChooser\r\n\r\n\r\nclass MultiFileChooser(BaseChooser):\r\n # todo: allow wildcard from argparse\r\n widget_class = core.MultiFileChooser\r\n\r\n\r\nclass FileSaver(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.FileSaver\r\n\r\n\r\nclass DirChooser(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.DirChooser\r\n\r\n\r\nclass MultiDirChooser(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.MultiDirChooser\r\n\r\n\r\nclass DateChooser(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.DateChooser\r\n\r\n\r\nclass ColourChooser(BaseChooser):\r\n widget_class = core.ColourChooser","new_contents":"from gooey.gui import formatters\r\nfrom gooey.gui.components.widgets import core\r\nfrom gooey.gui.components.widgets.bases import TextContainer, BaseChooser\r\n\r\n\r\n__ALL__ = [\r\n 'FileChooser',\r\n 'FileSaver',\r\n 'DirChooser',\r\n 'MultiDirChooser',\r\n 'DateChooser',\r\n 'ColourChooser'\r\n]\r\n\r\nclass FileChooser(BaseChooser):\r\n # todo: allow wildcard from argparse\r\n widget_class = core.FileChooser\r\n\r\n\r\nclass MultiFileChooser(BaseChooser):\r\n # todo: allow wildcard from argparse\r\n widget_class = core.MultiFileChooser\r\n\r\n\r\nclass FileSaver(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.FileSaver\r\n\r\n\r\nclass DirChooser(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.DirChooser\r\n\r\n\r\nclass MultiDirChooser(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.MultiDirChooser\r\n\r\n def formatOutput(self, metadata, value):\r\n return formatters.multiFileChooser(metadata, value)\r\n\r\n\r\nclass DateChooser(BaseChooser):\r\n # todo: allow wildcard\r\n widget_class = core.DateChooser\r\n\r\n\r\nclass ColourChooser(BaseChooser):\r\n widget_class = core.ColourChooser\r\n","subject":"Fix Issue 550: MultiDirChooser cannot handle paths with spaces","message":"Fix Issue 550: MultiDirChooser cannot handle paths with spaces\n\nInstead of using the general formatter, use formatters.multiFileChooser, which handles paths with spaces.\n","lang":"Python","license":"mit","repos":"chriskiehl\/Gooey"} {"commit":"6c32e39e2e51a80ebc9e31e88e22cc4aa39f7466","old_file":"chainer\/functions\/copy.py","new_file":"chainer\/functions\/copy.py","old_contents":"from chainer import cuda\nfrom chainer import function\n\n\nclass Copy(function.Function):\n\n \"\"\"Copy an input GPUArray onto another device.\"\"\"\n\n def __init__(self, out_device):\n self.out_device = out_device\n\n def forward_cpu(self, x):\n return x[0].copy(),\n\n def forward_gpu(self, x):\n return cuda.copy(x[0], out_device=self.out_device),\n\n def backward_cpu(self, x, gy):\n return gy[0].copy(),\n\n def backward_gpu(self, x, gy):\n return cuda.copy(gy[0], out_device=cuda.get_device(x[0])),\n\n\ndef copy(x, dst):\n \"\"\"Copies the input variable onto the specified device.\n\n This function copies the array of input variable onto the device specified\n by ``dst`` if the original array is on GPU, and otherwise just copies the\n array within host memory.\n\n Args:\n x (~chainer.Variable): Variable to be copied.\n dst: Target device specifier.\n\n Returns:\n ~chainer.Variable: Output variable.\n\n \"\"\"\n return Copy(dst)(x)\n","new_contents":"import numpy\n\nfrom chainer import cuda\nfrom chainer import function\nfrom chainer.utils import type_check\n\n\nclass Copy(function.Function):\n\n \"\"\"Copy an input GPUArray onto another device.\"\"\"\n\n def __init__(self, out_device):\n self.out_device = out_device\n\n def check_type_forward(self, in_types):\n type_check.expect(\n in_types.size() == 1,\n in_types[0].dtype == numpy.float32\n )\n\n def check_type_backward(self, in_types, out_types):\n type_check.expect(\n out_types.size() == 1,\n in_types[0].dtype == out_types[0].dtype,\n in_types[0].ndim == out_types[0].ndim,\n in_types[0].shape == out_types[0].shape\n )\n\n def forward_cpu(self, x):\n return x[0].copy(),\n\n def forward_gpu(self, x):\n return cuda.copy(x[0], out_device=self.out_device),\n\n def backward_cpu(self, x, gy):\n return gy[0].copy(),\n\n def backward_gpu(self, x, gy):\n return cuda.copy(gy[0], out_device=cuda.get_device(x[0])),\n\n\ndef copy(x, dst):\n \"\"\"Copies the input variable onto the specified device.\n\n This function copies the array of input variable onto the device specified\n by ``dst`` if the original array is on GPU, and otherwise just copies the\n array within host memory.\n\n Args:\n x (~chainer.Variable): Variable to be copied.\n dst: Target device specifier.\n\n Returns:\n ~chainer.Variable: Output variable.\n\n \"\"\"\n return Copy(dst)(x)\n","subject":"Add unittest(cpu-only) and typecheck for Copy","message":"Add unittest(cpu-only) and typecheck for Copy\n","lang":"Python","license":"mit","repos":"chainer\/chainer,sinhrks\/chainer,ronekko\/chainer,ktnyt\/chainer,chainer\/chainer,jnishi\/chainer,niboshi\/chainer,tkerola\/chainer,elviswf\/chainer,tscohen\/chainer,muupan\/chainer,keisuke-umezawa\/chainer,Kaisuke5\/chainer,woodshop\/chainer,jnishi\/chainer,keisuke-umezawa\/chainer,tigerneil\/chainer,cupy\/cupy,niboshi\/chainer,chainer\/chainer,hvy\/chainer,aonotas\/chainer,t-abe\/chainer,wkentaro\/chainer,hvy\/chainer,keisuke-umezawa\/chainer,umitanuki\/chainer,t-abe\/chainer,okuta\/chainer,kiyukuta\/chainer,ktnyt\/chainer,wkentaro\/chainer,ktnyt\/chainer,cupy\/cupy,okuta\/chainer,hvy\/chainer,cemoody\/chainer,woodshop\/complex-chainer,hidenori-t\/chainer,ysekky\/chainer,hvy\/chainer,ikasumi\/chainer,benob\/chainer,kashif\/chainer,sinhrks\/chainer,kikusu\/chainer,sou81821\/chainer,okuta\/chainer,pfnet\/chainer,1986ks\/chainer,cupy\/cupy,kikusu\/chainer,wkentaro\/chainer,muupan\/chainer,kuwa32\/chainer,wavelets\/chainer,keisuke-umezawa\/chainer,AlpacaDB\/chainer,ktnyt\/chainer,jnishi\/chainer,ytoyama\/yans_chainer_hackathon,bayerj\/chainer,delta2323\/chainer,niboshi\/chainer,niboshi\/chainer,anaruse\/chainer,yanweifu\/chainer,rezoo\/chainer,wkentaro\/chainer,truongdq\/chainer,okuta\/chainer,jnishi\/chainer,masia02\/chainer,jfsantos\/chainer,truongdq\/chainer,minhpqn\/chainer,benob\/chainer,laysakura\/chainer,cupy\/cupy,chainer\/chainer,AlpacaDB\/chainer"} {"commit":"6fbd752b1343c2e5085c3d060dbc7cc11a839728","old_file":"sympy\/utilities\/tests\/test_code_quality.py","new_file":"sympy\/utilities\/tests\/test_code_quality.py","old_contents":"from os import walk, sep, chdir, pardir\nfrom os.path import split, join, abspath\nfrom glob import glob\n\n# System path separator (usually slash or backslash)\nsepd = {\"sep\": sep}\n\n# Files having at least one of these in their path will be excluded\nEXCLUDE = set([\n \"%(sep)sthirdparty%(sep)s\" % sepd,\n \"%(sep)sprinting%(sep)spretty%(sep)s\" % sepd,\n])\n\ndef test_no_trailing_whitespace():\n message = \"File contains trailing whitespace: %s, line %s.\"\n base_path = split(__file__)[0]\n base_path += sep + pardir + sep + pardir # go to sympy\/\n base_path = abspath(base_path)\n for root, dirs, files in walk(base_path):\n for fname in glob(join(root, \"*.py\")):\n if filter(lambda ex: ex in fname, EXCLUDE):\n continue\n file = open(fname, \"r\")\n try:\n for idx, line in enumerate(file):\n if line.endswith(\" \\n\"):\n assert False, message % (fname, idx+1)\n finally:\n file.close()\n\n","new_contents":"from os import walk, sep, chdir, pardir\nfrom os.path import split, join, abspath\nfrom glob import glob\n\n# System path separator (usually slash or backslash)\nsepd = {\"sep\": sep}\n\n# Files having at least one of these in their path will be excluded\nEXCLUDE = set([\n \"%(sep)sthirdparty%(sep)s\" % sepd,\n# \"%(sep)sprinting%(sep)spretty%(sep)s\" % sepd,\n])\n\ndef test_no_trailing_whitespace():\n message = \"File contains trailing whitespace: %s, line %s.\"\n base_path = split(__file__)[0]\n base_path += sep + pardir + sep + pardir # go to sympy\/\n base_path = abspath(base_path)\n for root, dirs, files in walk(base_path):\n for fname in glob(join(root, \"*.py\")):\n if filter(lambda ex: ex in fname, EXCLUDE):\n continue\n file = open(fname, \"r\")\n try:\n for idx, line in enumerate(file):\n if line.endswith(\" \\n\"):\n assert False, message % (fname, idx+1)\n finally:\n file.close()\n","subject":"Test whitespace in pretty printing tests.","message":"Test whitespace in pretty printing tests.\n","lang":"Python","license":"bsd-3-clause","repos":"toolforger\/sympy,atsao72\/sympy,lindsayad\/sympy,fperez\/sympy,abhiii5459\/sympy,chaffra\/sympy,abhiii5459\/sympy,farhaanbukhsh\/sympy,wanglongqi\/sympy,bukzor\/sympy,ga7g08\/sympy,sahmed95\/sympy,sunny94\/temp,Shaswat27\/sympy,Titan-C\/sympy,lidavidm\/sympy,hargup\/sympy,ryanGT\/sympy,Curious72\/sympy,mcdaniel67\/sympy,emon10005\/sympy,rahuldan\/sympy,kumarkrishna\/sympy,maniteja123\/sympy,kevalds51\/sympy,farhaanbukhsh\/sympy,MridulS\/sympy,VaibhavAgarwalVA\/sympy,kaichogami\/sympy,sahilshekhawat\/sympy,minrk\/sympy,cccfran\/sympy,flacjacket\/sympy,debugger22\/sympy,kaichogami\/sympy,MridulS\/sympy,sunny94\/temp,toolforger\/sympy,jbaayen\/sympy,hrashk\/sympy,Davidjohnwilson\/sympy,ahhda\/sympy,wyom\/sympy,ChristinaZografou\/sympy,garvitr\/sympy,hazelnusse\/sympy-old,mattpap\/sympy-polys,postvakje\/sympy,wyom\/sympy,shipci\/sympy,jamesblunt\/sympy,postvakje\/sympy,Shaswat27\/sympy,emon10005\/sympy,asm666\/sympy,wanglongqi\/sympy,shikil\/sympy,garvitr\/sympy,vipulroxx\/sympy,moble\/sympy,pbrady\/sympy,souravsingh\/sympy,tovrstra\/sympy,sahilshekhawat\/sympy,grevutiu-gabriel\/sympy,kevalds51\/sympy,wanglongqi\/sympy,hrashk\/sympy,Arafatk\/sympy,oliverlee\/sympy,liangjiaxing\/sympy,Davidjohnwilson\/sympy,Designist\/sympy,yukoba\/sympy,oliverlee\/sympy,AkademieOlympia\/sympy,jaimahajan1997\/sympy,debugger22\/sympy,atsao72\/sympy,sahmed95\/sympy,kaushik94\/sympy,KevinGoodsell\/sympy,Vishluck\/sympy,amitjamadagni\/sympy,ga7g08\/sympy,atreyv\/sympy,asm666\/sympy,rahuldan\/sympy,pandeyadarsh\/sympy,vipulroxx\/sympy,Sumith1896\/sympy,Mitchkoens\/sympy,wyom\/sympy,amitjamadagni\/sympy,Gadal\/sympy,AunShiLord\/sympy,pbrady\/sympy,moble\/sympy,hazelnusse\/sympy-old,atreyv\/sympy,shipci\/sympy,vipulroxx\/sympy,cswiercz\/sympy,Designist\/sympy,shikil\/sympy,Arafatk\/sympy,shikil\/sympy,AkademieOlympia\/sympy,saurabhjn76\/sympy,abloomston\/sympy,saurabhjn76\/sympy,Titan-C\/sympy,pbrady\/sympy,kaushik94\/sympy,Sumith1896\/sympy,aktech\/sympy,abloomston\/sympy,meghana1995\/sympy,hrashk\/sympy,ahhda\/sympy,shipci\/sympy,sampadsaha5\/sympy,Gadal\/sympy,skidzo\/sympy,VaibhavAgarwalVA\/sympy,pandeyadarsh\/sympy,Curious72\/sympy,jbbskinny\/sympy,kmacinnis\/sympy,jamesblunt\/sympy,dqnykamp\/sympy,skidzo\/sympy,garvitr\/sympy,ChristinaZografou\/sympy,sampadsaha5\/sympy,jamesblunt\/sympy,jbbskinny\/sympy,Arafatk\/sympy,jerli\/sympy,ChristinaZografou\/sympy,kumarkrishna\/sympy,jerli\/sympy,souravsingh\/sympy,skirpichev\/omg,souravsingh\/sympy,maniteja123\/sympy,kmacinnis\/sympy,mafiya69\/sympy,jbbskinny\/sympy,lindsayad\/sympy,VaibhavAgarwalVA\/sympy,maniteja123\/sympy,chaffra\/sympy,moble\/sympy,Vishluck\/sympy,Mitchkoens\/sympy,madan96\/sympy,jerli\/sympy,debugger22\/sympy,chaffra\/sympy,mcdaniel67\/sympy,lidavidm\/sympy,srjoglekar246\/sympy,grevutiu-gabriel\/sympy,bukzor\/sympy,kevalds51\/sympy,Titan-C\/sympy,hargup\/sympy,AkademieOlympia\/sympy,Shaswat27\/sympy,Mitchkoens\/sympy,iamutkarshtiwari\/sympy,dqnykamp\/sympy,aktech\/sympy,cswiercz\/sympy,yashsharan\/sympy,AunShiLord\/sympy,beni55\/sympy,Vishluck\/sympy,beni55\/sympy,madan96\/sympy,yukoba\/sympy,atsao72\/sympy,kaichogami\/sympy,yashsharan\/sympy,pandeyadarsh\/sympy,meghana1995\/sympy,sahilshekhawat\/sympy,skidzo\/sympy,kmacinnis\/sympy,jaimahajan1997\/sympy,kumarkrishna\/sympy,sunny94\/temp,kaushik94\/sympy,diofant\/diofant,cccfran\/sympy,ahhda\/sympy,Curious72\/sympy,drufat\/sympy,toolforger\/sympy,sahmed95\/sympy,MechCoder\/sympy,MechCoder\/sympy,hargup\/sympy,emon10005\/sympy,rahuldan\/sympy,abloomston\/sympy,drufat\/sympy,mcdaniel67\/sympy,iamutkarshtiwari\/sympy,liangjiaxing\/sympy,ga7g08\/sympy,beni55\/sympy,saurabhjn76\/sympy,Davidjohnwilson\/sympy,Gadal\/sympy,mafiya69\/sympy,Designist\/sympy,atreyv\/sympy,abhiii5459\/sympy,mafiya69\/sympy,grevutiu-gabriel\/sympy,jaimahajan1997\/sympy,pernici\/sympy,dqnykamp\/sympy,postvakje\/sympy,cswiercz\/sympy,farhaanbukhsh\/sympy,bukzor\/sympy,lindsayad\/sympy,sampadsaha5\/sympy,minrk\/sympy,iamutkarshtiwari\/sympy,meghana1995\/sympy,lidavidm\/sympy,drufat\/sympy,Sumith1896\/sympy,AunShiLord\/sympy,asm666\/sympy,madan96\/sympy,cccfran\/sympy,yukoba\/sympy,MechCoder\/sympy,yashsharan\/sympy,liangjiaxing\/sympy,MridulS\/sympy,oliverlee\/sympy,aktech\/sympy"} {"commit":"f01222f021f277805492e3f539609f6b64be0b7e","old_file":"blanc_basic_news\/news\/views.py","new_file":"blanc_basic_news\/news\/views.py","old_contents":"from django.views.generic import ListView, DateDetailView\nfrom django.shortcuts import get_object_or_404\nfrom django.utils import timezone\nfrom django.conf import settings\nfrom .models import Category, Post\n\n\nclass PostListView(ListView):\n paginate_by = getattr(settings, 'NEWS_PER_PAGE', 10)\n\n def get_queryset(self):\n return Post.objects.filter(published=True, date__lte=timezone.now())\n\n\nclass PostListCategoryView(ListView):\n paginate_by = getattr(settings, 'NEWS_PER_PAGE', 10)\n template_name_suffix = '_list_category'\n\n def get_queryset(self):\n self.category = get_object_or_404(Category, slug=self.kwargs['slug'])\n return Post.objects.filter(published=True, date__lte=timezone.now(), category=self.category)\n\n def get_context_data(self, **kwargs):\n context = super(PostListCategoryView, self).get_context_data(**kwargs)\n context['category'] = self.category\n return context\n\n\nclass PostDetailView(DateDetailView):\n queryset = Post.objects.filter(published=True)\n month_format = '%m'\n date_field = 'date'\n","new_contents":"from django.views.generic import ListView, DateDetailView\nfrom django.shortcuts import get_object_or_404\nfrom django.utils import timezone\nfrom django.conf import settings\nfrom .models import Category, Post\n\n\nclass PostListView(ListView):\n paginate_by = getattr(settings, 'NEWS_PER_PAGE', 10)\n\n def get_queryset(self):\n return Post.objects.select_related().filter(\n published=True, date__lte=timezone.now())\n\n\nclass PostListCategoryView(ListView):\n paginate_by = getattr(settings, 'NEWS_PER_PAGE', 10)\n template_name_suffix = '_list_category'\n\n def get_queryset(self):\n self.category = get_object_or_404(Category, slug=self.kwargs['slug'])\n return Post.objects.select_related().filter(\n published=True,\n date__lte=timezone.now(),\n category=self.category)\n\n def get_context_data(self, **kwargs):\n context = super(PostListCategoryView, self).get_context_data(**kwargs)\n context['category'] = self.category\n return context\n\n\nclass PostDetailView(DateDetailView):\n queryset = Post.objects.filter(published=True)\n month_format = '%m'\n date_field = 'date'\n","subject":"Use select_related to help with category foreign keys","message":"Use select_related to help with category foreign keys\n","lang":"Python","license":"bsd-3-clause","repos":"blancltd\/blanc-basic-news"} {"commit":"19cd84480a739f9550258dc959637fe85f43af50","old_file":"fedora\/release.py","new_file":"fedora\/release.py","old_contents":"'''\nInformation about this python-fedora release\n'''\n\nfrom fedora import _\n\nNAME = 'python-fedora'\nVERSION = '0.3.6'\nDESCRIPTION = _('Python modules for interacting with Fedora services')\nLONG_DESCRIPTION = _('''\nThe Fedora Project runs many different services. These services help us to\npackage software, develop new programs, and generally put together the distro.\nThis package contains software that helps us do that.\n''')\nAUTHOR = 'Toshio Kuratomi, Luke Macken'\nEMAIL = 'tkuratom@redhat.com'\nCOPYRIGHT = '2007-2008 Red Hat, Inc.'\nURL = 'https:\/\/fedorahosted.org\/python-fedora'\nDOWNLOAD_URL = 'https:\/\/fedorahosted.org\/releases\/p\/y\/python-fedora\/'\nLICENSE = 'GPLv2'\n","new_contents":"'''\nInformation about this python-fedora release\n'''\n\nfrom fedora import _\n\nNAME = 'python-fedora'\nVERSION = '0.3.6'\nDESCRIPTION = _('Python modules for interacting with Fedora Services')\nLONG_DESCRIPTION = _('''\nThe Fedora Project runs many different services. These services help us to\npackage software, develop new programs, and generally put together the distro.\nThis package contains software that helps us do that.\n''')\nAUTHOR = 'Toshio Kuratomi, Luke Macken'\nEMAIL = 'tkuratom@redhat.com'\nCOPYRIGHT = '2007-2008 Red Hat, Inc.'\nURL = 'https:\/\/fedorahosted.org\/python-fedora'\nDOWNLOAD_URL = 'https:\/\/fedorahosted.org\/releases\/p\/y\/python-fedora\/'\nLICENSE = 'GPLv2'\n","subject":"Correct minor typo in a string.","message":"Correct minor typo in a string.\n","lang":"Python","license":"lgpl-2.1","repos":"fedora-infra\/python-fedora"} {"commit":"db37b195ea47cd18969ad482e1dae301903da092","old_file":"pyOutlook\/__init__.py","new_file":"pyOutlook\/__init__.py","old_contents":"from .core import *\n\n__all__ = ['OutlookAccount', 'Message', 'Contact', 'Folder']\n__version__ = '1.0.0'\n__release__ = '1.0.0'\n","new_contents":"from .core import *\n\n__all__ = ['OutlookAccount', 'Message', 'Contact', 'Folder']\n__version__ = '1.0.0dev'\n__release__ = '1.0.0dev'\n","subject":"Package development version of upcoming v1 release for testing.","message":"Package development version of upcoming v1 release for testing.\n","lang":"Python","license":"mit","repos":"JensAstrup\/pyOutlook"} {"commit":"c5f10b2e5ea10dd17c8c19f87dcdfd2584f8e431","old_file":"comics\/accounts\/models.py","new_file":"comics\/accounts\/models.py","old_contents":"import uuid\n\nfrom django.contrib.auth.models import User\nfrom django.db import models\nfrom django.dispatch import receiver\n\n\n@receiver(models.signals.post_save, sender=User)\ndef create_user_profile(sender, instance, created, **kwargs):\n UserProfile.objects.get_or_create(user=instance)\n\n\nclass UserProfile(models.Model):\n user = models.OneToOneField(User)\n secret_key = models.CharField(max_length=32, blank=False,\n help_text='Secret key for feed and API access')\n\n class Meta:\n db_table = 'comics_user_profile'\n\n def __init__(self, *args, **kwargs):\n super(UserProfile, self).__init__(*args, **kwargs)\n self.generate_new_secret_key()\n\n def __unicode__(self):\n return u'User profile for %s' % self.user\n\n def generate_new_secret_key(self):\n self.secret_key = uuid.uuid4().hex\n","new_contents":"import uuid\n\nfrom django.contrib.auth.models import User\nfrom django.db import models\nfrom django.dispatch import receiver\n\n\n@receiver(models.signals.post_save, sender=User)\ndef create_user_profile(sender, instance, created, **kwargs):\n if created:\n UserProfile.objects.create(user=instance)\n\nclass UserProfile(models.Model):\n user = models.OneToOneField(User)\n secret_key = models.CharField(max_length=32, blank=False,\n help_text='Secret key for feed and API access')\n\n class Meta:\n db_table = 'comics_user_profile'\n\n def __init__(self, *args, **kwargs):\n super(UserProfile, self).__init__(*args, **kwargs)\n self.generate_new_secret_key()\n\n def __unicode__(self):\n return u'User profile for %s' % self.user\n\n def generate_new_secret_key(self):\n self.secret_key = uuid.uuid4().hex\n","subject":"Remove conditional sql-select on new user creation","message":"Remove conditional sql-select on new user creation\n\nOnly create a user profile if a new user is actually\ncreated.\n","lang":"Python","license":"agpl-3.0","repos":"datagutten\/comics,jodal\/comics,datagutten\/comics,jodal\/comics,datagutten\/comics,datagutten\/comics,jodal\/comics,jodal\/comics"} {"commit":"fd6c7386cfdaa5fb97a428b323fc1f9b17f9f02c","old_file":"tests\/test_helpers.py","new_file":"tests\/test_helpers.py","old_contents":"import pandas\n\nfrom sharepa.helpers import pretty_print\nfrom sharepa.helpers import source_counts\n\n\ndef test_pretty_print():\n some_stuff = '{\"Dusty\": \"Rhodes\"}'\n pretty_print(some_stuff)\n\n\ndef test_source_counts():\n all_counts = source_counts()\n assert isinstance(all_counts, pandas.core.frame.DataFrame)\n","new_contents":"import vcr\nimport pandas\nimport pytest\n\nfrom sharepa.search import ShareSearch\nfrom sharepa.helpers import pretty_print\nfrom sharepa.helpers import source_counts\n\n\n@vcr.use_cassette('tests\/vcr\/simple_execute.yaml')\ndef test_pretty_print():\n my_search = ShareSearch()\n result = my_search.execute()\n the_dict = result.to_dict()\n try:\n pretty_print(the_dict)\n except:\n pytest.fail(\"Unexpected exception!!\")\n\n\ndef test_source_counts():\n all_counts = source_counts()\n assert isinstance(all_counts, pandas.core.frame.DataFrame)\n","subject":"Add pytest fail check on raising pretty print exeption","message":"Add pytest fail check on raising pretty print exeption\n","lang":"Python","license":"mit","repos":"CenterForOpenScience\/sharepa,fabianvf\/sharepa,samanehsan\/sharepa,erinspace\/sharepa"} {"commit":"0a9e3fb387c61f2c7cb32502f5c50eaa5b950169","old_file":"tests\/test_process.py","new_file":"tests\/test_process.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import print_function, unicode_literals\nimport pytest\nfrom wamopacker.process import run_command, ProcessException\nimport os\nimport uuid\n\n\ndef test_run_command():\n cwd = os.getcwd()\n output_cmd = run_command('ls -1A', working_dir = cwd)\n output_py = os.listdir(cwd)\n assert sorted(output_cmd) == sorted(output_py)\n\n\ndef test_run_command_error():\n data = uuid.uuid4().hex\n with pytest.raises(ProcessException) as e:\n run_command('cat {}'.format(data))\n\n assert e.value.log_stdout == ''\n assert e.value.log_stderr == 'cat: {}: No such file or directory\\n'.format(data)\n assert e.value.exit_code != 0\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom __future__ import print_function, unicode_literals\nimport pytest\nfrom wamopacker.process import run_command, ProcessException\nimport os\nimport uuid\n\n\ndef test_run_command():\n cwd = os.getcwd()\n output_cmd = run_command('ls -1A', working_dir = cwd)\n output_py = os.listdir(cwd)\n assert sorted(output_cmd) == sorted(output_py)\n\n\ndef test_run_command_error():\n data = uuid.uuid4().hex\n with pytest.raises(ProcessException) as e:\n run_command('cat {}'.format(data))\n\n assert e.value.log_stdout == ''\n assert e.value.log_stderr.startswith('cat: {}'.format(data))\n assert e.value.exit_code != 0\n","subject":"Fix intermittent travis build error.","message":"Fix intermittent travis build error.\n","lang":"Python","license":"mit","repos":"wamonite\/packermate"} {"commit":"e79c90db5dcda56ff9b2b154659984db9c6f7663","old_file":"src\/main.py","new_file":"src\/main.py","old_contents":"# -*- encoding: utf-8 -*-\n\nimport pygame\nfrom scenes import director\nfrom scenes import intro_scene\n\npygame.init()\n\n\ndef main():\n game_director = director.Director()\n scene = intro_scene.IntroScene(game_director)\n game_director.change_scene(scene)\n game_director.loop()\n\n\nif __name__ == '__main__':\n pygame.init()\n main()\n","new_contents":"# -*- encoding: utf-8 -*-\n\nimport pygame\nfrom scenes import director\nfrom scenes import intro_scene\nfrom game_logic import settings\n\npygame.init()\n\n\ndef main():\n initial_settings = settings.Settings(\n trials=1000, player='O', oponent='Computer')\n game_director = director.Director()\n scene = intro_scene.IntroScene(game_director)\n game_director.change_scene(scene, initial_settings)\n game_director.loop()\n\n\nif __name__ == '__main__':\n pygame.init()\n main()\n","subject":"Create initial config when starting game","message":"Create initial config when starting game\n","lang":"Python","license":"mit","repos":"juangallostra\/TicTacToe"} {"commit":"7dd228d7eaad6b1f37ff3c4d954aebe0ffa99170","old_file":"tests\/test_targets\/test_targets.py","new_file":"tests\/test_targets\/test_targets.py","old_contents":"# Copyright 2015 0xc0170\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\nimport os\n\nfrom unittest import TestCase\n\nfrom project_generator_definitions.definitions import ProGenTargets\n\nclass TestAllTargets(TestCase):\n\n \"\"\"test all targets\"\"\"\n\n def setUp(self):\n self.progen_target = ProGenTargets()\n self.targets_list = self.progen_target.get_targets()\n\n def test_targets_validity(self):\n for target in self.targets_list:\n record = self.progen_target.get_target_record(target)\n assert record['target']['name'][0]\n assert record['target']['mcu'][0]\n\n def test_targets_mcu_validity(self):\n for target in self.targets_list:\n mcu = self.progen_target.get_mcu_record(target)\n assert mcu['mcu']\n assert mcu['mcu']['name']\n assert mcu['mcu']['core']\n","new_contents":"# Copyright 2015 0xc0170\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom unittest import TestCase\n\nfrom project_generator_definitions.definitions import ProGenTargets\n\nclass TestAllTargets(TestCase):\n\n \"\"\"test all targets\"\"\"\n\n def setUp(self):\n self.progen_target = ProGenTargets()\n self.targets_list = self.progen_target.get_targets()\n\n def test_targets_validity(self):\n # Cehck for required info for targets\n for target in self.targets_list:\n record = self.progen_target.get_target_record(target)\n assert record['target']['name'][0]\n assert record['target']['mcu'][0]\n\n def test_targets_mcu_validity(self):\n # Check for required info in mcu\n for target in self.targets_list:\n mcu = self.progen_target.get_mcu_record(target)\n assert mcu['mcu'][0]\n assert mcu['mcu']['name'][0]\n assert mcu['mcu']['core'][0]\n","subject":"Test - targets test fix mcu validity indexes","message":"Test - targets test fix mcu validity indexes\n","lang":"Python","license":"apache-2.0","repos":"project-generator\/project_generator_definitions,0xc0170\/project_generator_definitions,ohagendorf\/project_generator_definitions"} {"commit":"bd2c0efa6b0205ff0d24cf335f65f755f18566f2","old_file":"modernrpc\/__init__.py","new_file":"modernrpc\/__init__.py","old_contents":"# coding: utf-8\n\n# default_app_config was deprecated in Django 3.2. Maybe set it only when detected django version is older?\ndefault_app_config = \"modernrpc.apps.ModernRpcConfig\"\n\n# Package version is now stored in pyproject.toml only. To retrieve it from code, use:\n# import pkg_resources; version = pkg_resources.get_distribution('django-modern-rpc').version\n","new_contents":"# coding: utf-8\nfrom packaging.version import Version\nimport django\n\n# Set default_app_config only with Django up to 3.1. This prevents a Warning on newer releases\n# See https:\/\/docs.djangoproject.com\/fr\/3.2\/releases\/3.2\/#automatic-appconfig-discovery\nif Version(django.get_version()) < Version(\"3.2\"):\n default_app_config = \"modernrpc.apps.ModernRpcConfig\"\n\n# Package version is now stored in pyproject.toml only. To retrieve it from code, use:\n# import pkg_resources; version = pkg_resources.get_distribution('django-modern-rpc').version\n","subject":"Stop defining default_app_config on Django 3.2+","message":"Stop defining default_app_config on Django 3.2+","lang":"Python","license":"mit","repos":"alorence\/django-modern-rpc,alorence\/django-modern-rpc"} {"commit":"77a1ee839da665fc1f97dabed1bf5639c980a17a","old_file":"src\/api\/controller\/ServerListController.py","new_file":"src\/api\/controller\/ServerListController.py","old_contents":"from BaseController import BaseController\nfrom api.util import settings\n\nclass ServerListController(BaseController):\n\n def get(self):\n servers = {\"servers\": self.read_server_config()}\n self.write(servers)\n\n def read_server_config(self):\n \"\"\"Returns a list of servers with the 'id' field added.\n \"\"\"\n # TODO: Move this into the settings module so everything benefits.\n server_list = []\n redis_servers = settings.get_redis_servers()\n\n for server in redis_servers:\n server_id = \"%(server)s:%(port)s\" % server\n s = dict(server=server['server'], port=server['port'], password=server['password'], id=server_id)\n server_list.append(s)\n\n return server_list\n","new_contents":"from BaseController import BaseController\nfrom api.util import settings\n\nclass ServerListController(BaseController):\n\n def get(self):\n servers = {\"servers\": self.read_server_config()}\n self.write(servers)\n\n def read_server_config(self):\n \"\"\"Returns a list of servers with the 'id' field added.\n \"\"\"\n # TODO: Move this into the settings module so everything benefits.\n server_list = []\n redis_servers = settings.get_redis_servers()\n\n for server in redis_servers:\n if 'password' not in server:\n server['password'] = None\n\n server_id = \"%(server)s:%(port)s\" % server\n s = dict(server=server['server'], port=server['port'], password=server['password'], id=server_id)\n server_list.append(s)\n\n return server_list\n","subject":"Allow servers command to work without a password.","message":"Allow servers command to work without a password.\n","lang":"Python","license":"mit","repos":"YongMan\/RedisLive,merlian\/RedisLive,heamon7\/RedisLive,fengshao0907\/RedisLive,heamon7\/RedisLive,udomsak\/RedisLive,merlian\/RedisLive,jacklee0810\/RedisLive,YongMan\/RedisLive,jacklee0810\/RedisLive,udomsak\/RedisLive,jiejieling\/RdsMonitor,udomsak\/RedisLive,fengshao0907\/RedisLive,nkrode\/RedisLive,jacklee0810\/RedisLive,jiejieling\/RdsMonitor,merlian\/RedisLive,nkrode\/RedisLive,fengshao0907\/RedisLive,YongMan\/RedisLive,nkrode\/RedisLive,jiejieling\/RdsMonitor,heamon7\/RedisLive"} {"commit":"271bb9de8f0f3674b1f6f47bc3519f1297c87abf","old_file":"examples\/linechannel.py","new_file":"examples\/linechannel.py","old_contents":"# -*- coding: utf-8 -*-\r\nfrom linepy import *\r\n\r\nclient = LineClient()\r\n#client = LineClient(authToken='AUTHTOKEN')\r\n\r\nclient.log(\"Auth Token : \" + str(client.authToken))\r\n\r\n# Initialize LineChannel with LineClient\r\n# This channel id is Timeline channel\r\nchannel = LineChannel(client, channel_id=\"1341209950\")\r\nclient.log(\"Channel Access Token : \" + str(channel.channelAccessToken))","new_contents":"# -*- coding: utf-8 -*-\r\nfrom linepy import *\r\n\r\nclient = LineClient()\r\n#client = LineClient(authToken='AUTHTOKEN')\r\n\r\nclient.log(\"Auth Token : \" + str(client.authToken))\r\n\r\n# Initialize LineChannel with LineClient\r\n# This channel id is Timeline channel\r\nchannel = LineChannel(client, channelId=\"1341209950\")\r\nclient.log(\"Channel Access Token : \" + str(channel.channelAccessToken))\r\n","subject":"Change channel_id to new channelId param","message":"Change channel_id to new channelId param","lang":"Python","license":"bsd-3-clause","repos":"fadhiilrachman\/line-py"} {"commit":"bc5d678937e69fe00e206b6a80c9a2f6dfb1a3a2","old_file":"examples\/worker_rush.py","new_file":"examples\/worker_rush.py","old_contents":"import sc2\nfrom sc2 import run_game, maps, Race, Difficulty\nfrom sc2.player import Bot, Computer\n\nclass WorkerRushBot(sc2.BotAI):\n async def on_step(self, state, iteration):\n if iteration == 0:\n for probe in self.workers:\n await self.do(probe.attack(self.enemy_start_locations[0]))\n\ndef main():\n run_game(maps.get(\"Abyssal Reef LE\"), [\n Bot(Race.Protoss, WorkerRushBot()),\n Computer(Race.Protoss, Difficulty.Medium)\n ], realtime=True)\n\nif __name__ == '__main__':\n main()\n","new_contents":"import sc2\nfrom sc2 import run_game, maps, Race, Difficulty\nfrom sc2.player import Bot, Computer\n\nclass WorkerRushBot(sc2.BotAI):\n async def on_step(self, state, iteration):\n if iteration == 0:\n for worker in self.workers:\n await self.do(worker.attack(self.enemy_start_locations[0]))\n\ndef main():\n run_game(maps.get(\"Abyssal Reef LE\"), [\n Bot(Race.Zerg, WorkerRushBot()),\n Computer(Race.Protoss, Difficulty.Medium)\n ], realtime=True)\n\nif __name__ == '__main__':\n main()\n","subject":"Use generic names in the worker rush example","message":"Use generic names in the worker rush example\n","lang":"Python","license":"mit","repos":"Dentosal\/python-sc2"} {"commit":"a6a2ee870840730f99ad475e02956c49fe2e7ed3","old_file":"common\/authapp.py","new_file":"common\/authapp.py","old_contents":"import ConfigParser\nfrom common.application import Application\nfrom keystonemiddleware.auth_token import filter_factory as auth_filter_factory\n\nclass KeystoneApplication(Application):\n\n \"\"\"\n An Application which uses Keystone for authorisation using RBAC\n \"\"\"\n\n def __init__(self, configuration):\n super(KeystoneApplication, self).__init__(configuration)\n self.required_role = self.config.get('authorisation', 'required_role')\n if self.required_role is None:\n raise ValueError(\"No required role supplied\")\n\n def _check_auth(self, req):\n if 'HTTP_X_ROLES' in req.environ:\n user_roles = req.environ['HTTP_X_ROLES'].split(',')\n return self.required_role in user_roles\n return False\n\ndef keystone_auth_filter_factory(global_config, **local_config):\n global_config.update(local_config)\n config_file_name = global_config.get('config_file', 'apiv1app.ini')\n config_file = ConfigParser.SafeConfigParser()\n config_file.read(config_file_name)\n global_config.update(config_file.items('keystone_authtoken'))\n return auth_filter_factory(global_config)\n","new_contents":"import ConfigParser\nfrom common.application import Application\nfrom keystonemiddleware.auth_token import filter_factory as auth_filter_factory\n\nclass KeystoneApplication(Application):\n\n \"\"\"\n An Application which uses Keystone for authorisation using RBAC\n \"\"\"\n\n INI_SECTION = 'keystone_authtoken'\n\n def __init__(self, configuration):\n super(KeystoneApplication, self).__init__(configuration)\n self.required_role = self.config.get('authorisation', 'required_role')\n if self.required_role is None:\n raise ValueError(\"No required role supplied\")\n\n def _check_auth(self, req):\n if 'HTTP_X_ROLES' in req.environ:\n user_roles = req.environ['HTTP_X_ROLES'].split(',')\n return self.required_role in user_roles\n return False\n\ndef keystone_auth_filter_factory(global_config, **local_config):\n global_config.update(local_config)\n config_file_name = global_config.get('config_file')\n if not config_file_name:\n raise ValueError('No config_file directive')\n config_file = ConfigParser.SafeConfigParser()\n if not config_file.read(config_file_name):\n raise ValueError(\"Cannot read config file '%s'\" % config_file_name)\n global_config.update(config_file.items(KeystoneApplication.INI_SECTION))\n return auth_filter_factory(global_config)\n","subject":"Remove hardcoded default filename. Raise an error if no app config file was specified, or it is unreadable, or it doesn't contain the section we need.","message":"Remove hardcoded default filename. Raise an error if no app config file was specified, or it is unreadable, or it doesn't contain the section we need.\n","lang":"Python","license":"apache-2.0","repos":"NCI-Cloud\/reporting-api,NeCTAR-RC\/reporting-api,NCI-Cloud\/reporting-api,NeCTAR-RC\/reporting-api"} {"commit":"66462c231011f6418fc246789ce4feed10a74a66","old_file":"web\/whim\/core\/time.py","new_file":"web\/whim\/core\/time.py","old_contents":"from datetime import datetime, timezone, time\n\n\ndef zero_time_with_timezone(date, tz=timezone.utc):\n return datetime.combine(date, time(tzinfo=tz))","new_contents":"from datetime import datetime, timezone, time\n\nimport dateparser\n\n\ndef zero_time_with_timezone(date, tz=timezone.utc):\n return datetime.combine(date, time(tzinfo=tz))\n\n\ndef attempt_parse_date(val):\n parsed_date = dateparser.parse(val, languages=['en'])\n if parsed_date is None:\n # try other strategies?\n pass\n return parsed_date","subject":"Use dateparser for parsing scraped dates","message":"Use dateparser for parsing scraped dates\n","lang":"Python","license":"mit","repos":"andrewgleave\/whim,andrewgleave\/whim,andrewgleave\/whim"} {"commit":"57a37c4a87e9757a109dfb5f3169fb8264d0795e","old_file":"neutron\/server\/rpc_eventlet.py","new_file":"neutron\/server\/rpc_eventlet.py","old_contents":"#!\/usr\/bin\/env python\n\n# Copyright 2011 VMware, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n# If ..\/neutron\/__init__.py exists, add ..\/ to Python search path, so that\n# it will override what happens to be installed in \/usr\/(local\/)lib\/python...\n\nfrom oslo_log import log\n\nfrom neutron._i18n import _LI\nfrom neutron import manager\nfrom neutron import service\n\nLOG = log.getLogger(__name__)\n\n\ndef eventlet_rpc_server():\n LOG.info(_LI(\"Eventlet based AMQP RPC server starting...\"))\n\n try:\n manager.init()\n workers = service._get_rpc_workers() + service._get_plugins_workers()\n rpc_workers_launcher = service._start_workers(workers)\n except NotImplementedError:\n LOG.info(_LI(\"RPC was already started in parent process by \"\n \"plugin.\"))\n else:\n rpc_workers_launcher.wait()\n","new_contents":"#!\/usr\/bin\/env python\n\n# Copyright 2011 VMware, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n# If ..\/neutron\/__init__.py exists, add ..\/ to Python search path, so that\n# it will override what happens to be installed in \/usr\/(local\/)lib\/python...\n\nfrom oslo_log import log\n\nfrom neutron._i18n import _LI\nfrom neutron import manager\nfrom neutron import service\n\nLOG = log.getLogger(__name__)\n\n\ndef eventlet_rpc_server():\n LOG.info(_LI(\"Eventlet based AMQP RPC server starting...\"))\n\n try:\n manager.init()\n rpc_workers_launcher = service.start_all_workers()\n except NotImplementedError:\n LOG.info(_LI(\"RPC was already started in parent process by \"\n \"plugin.\"))\n else:\n rpc_workers_launcher.wait()\n","subject":"Switch to start_all_workers in RPC server","message":"Switch to start_all_workers in RPC server\n\nThis does the same as the logic present but it emits\nthe registry callback event for resources.PROCESS AFTER_SPAWN\nthat some plugins may be expecting.\n\nChange-Id: I6f9aeca753a5d3c0052f553a2ac46786ca113e1e\nRelated-Bug: #1687896\n","lang":"Python","license":"apache-2.0","repos":"mahak\/neutron,noironetworks\/neutron,openstack\/neutron,openstack\/neutron,openstack\/neutron,eayunstack\/neutron,mahak\/neutron,huntxu\/neutron,eayunstack\/neutron,noironetworks\/neutron,huntxu\/neutron,mahak\/neutron"} {"commit":"bc961992afeae978e95209606e0e7b1a9b73719f","old_file":"jesusmtnez\/python\/kata\/game.py","new_file":"jesusmtnez\/python\/kata\/game.py","old_contents":"class Game():\n def __init__(self):\n self._score = 0\n\n def roll(self, pins):\n pass\n\n def score(self):\n return 0\n","new_contents":"class Game():\n def __init__(self):\n self._score = 0\n\n def roll(self, pins):\n self._score += pins\n\n def score(self):\n return self._score\n","subject":"Update score in Game class methods","message":"[Python] Update score in Game class methods\n","lang":"Python","license":"mit","repos":"JesusMtnez\/devexperto-challenge,JesusMtnez\/devexperto-challenge"} {"commit":"008e8a0bd4ae88adcf9e54b88cb8a4dedc9edd5a","old_file":"vtki\/_version.py","new_file":"vtki\/_version.py","old_contents":"\"\"\" version info for vtki \"\"\"\n# major, minor, patch\nversion_info = 0, 17, 5\n\n# Nice string for the version\n__version__ = '.'.join(map(str, version_info))\n","new_contents":"\"\"\" version info for vtki \"\"\"\n# major, minor, patch\nversion_info = 0, 18, 0\n\n# Nice string for the version\n__version__ = '.'.join(map(str, version_info))\n","subject":"Bump version: 0.17.5 → 0.18.0","message":"Bump version: 0.17.5 → 0.18.0\n","lang":"Python","license":"mit","repos":"akaszynski\/vtkInterface"} {"commit":"69c72d47ebf57932b6e20e2c22a5f1c84d07d3eb","old_file":"pyqode\/core\/api\/__init__.py","new_file":"pyqode\/core\/api\/__init__.py","old_contents":"\"\"\"\nThis package contains the bases classes of pyqode and some utility\nfunctions.\n\n\"\"\"\nfrom .code_edit import CodeEdit\nfrom .decoration import TextDecoration\nfrom .encodings import ENCODINGS_MAP, convert_to_codec_key\nfrom .manager import Manager\nfrom .mode import Mode\nfrom .panel import Panel\nfrom .syntax_highlighter import SyntaxHighlighter\nfrom .syntax_highlighter import ColorScheme\nfrom .syntax_highlighter import TextBlockUserData\nfrom .utils import TextHelper, TextBlockHelper\nfrom .utils import get_block_symbol_data\nfrom .utils import DelayJobRunner\nfrom .folding import FoldDetector\nfrom .folding import IndentFoldDetector\nfrom .folding import CharBasedFoldDetector\nfrom .folding import FoldScope\n\n\n__all__ = [\n 'convert_to_codec_key',\n 'get_block_symbol_data',\n 'CharBasedFoldDetector',\n 'CodeEdit',\n 'ColorScheme',\n 'DelayJobRunner',\n 'ENCODINGS_MAP',\n 'FoldDetector',\n 'IndentFoldDetector',\n 'FoldScope',\n 'Manager',\n 'Mode',\n 'Panel',\n 'SyntaxHighlighter',\n 'TextBlockUserData',\n 'TextDecoration',\n 'TextHelper',\n 'TextBlockHelper'\n]\n","new_contents":"\"\"\"\nThis package contains the bases classes of pyqode and some utility\nfunctions.\n\n\"\"\"\nfrom .code_edit import CodeEdit\nfrom .decoration import TextDecoration\nfrom .encodings import ENCODINGS_MAP, convert_to_codec_key\nfrom .manager import Manager\nfrom .mode import Mode\nfrom .panel import Panel\nfrom .syntax_highlighter import ColorScheme\nfrom .syntax_highlighter import PYGMENTS_STYLES\nfrom .syntax_highlighter import SyntaxHighlighter\nfrom .syntax_highlighter import TextBlockUserData\nfrom .utils import TextHelper, TextBlockHelper\nfrom .utils import get_block_symbol_data\nfrom .utils import DelayJobRunner\nfrom .folding import FoldDetector\nfrom .folding import IndentFoldDetector\nfrom .folding import CharBasedFoldDetector\nfrom .folding import FoldScope\n\n\n__all__ = [\n 'convert_to_codec_key',\n 'get_block_symbol_data',\n 'CharBasedFoldDetector',\n 'CodeEdit',\n 'ColorScheme',\n 'DelayJobRunner',\n 'ENCODINGS_MAP',\n 'FoldDetector',\n 'IndentFoldDetector',\n 'FoldScope',\n 'Manager',\n 'Mode',\n 'Panel',\n 'PYGMENTS_STYLES',\n 'SyntaxHighlighter',\n 'TextBlockUserData',\n 'TextDecoration',\n 'TextHelper',\n 'TextBlockHelper'\n]\n","subject":"Add missing PYGMENTS_STYLES list to pyqode.core.api","message":"Add missing PYGMENTS_STYLES list to pyqode.core.api\n","lang":"Python","license":"mit","repos":"zwadar\/pyqode.core,pyQode\/pyqode.core,pyQode\/pyqode.core"} {"commit":"23d8942ffeeee72e21330bd8ecc5bfb5e91bbc3b","old_file":"certidude\/push.py","new_file":"certidude\/push.py","old_contents":"\nimport click\nimport json\nimport logging\nimport requests\nfrom datetime import datetime\nfrom certidude import config\n\n\ndef publish(event_type, event_data):\n \"\"\"\n Publish event on push server\n \"\"\"\n if not isinstance(event_data, basestring):\n from certidude.decorators import MyEncoder\n event_data = json.dumps(event_data, cls=MyEncoder)\n\n url = config.PUSH_PUBLISH % config.PUSH_TOKEN\n click.echo(\"Publishing %s event '%s' on %s\" % (event_type, event_data, url))\n\n try:\n notification = requests.post(\n url,\n data=event_data,\n headers={\"X-EventSource-Event\": event_type, \"User-Agent\": \"Certidude API\"})\n if notification.status_code == requests.codes.created:\n pass # Sent to client\n elif notification.status_code == requests.codes.accepted:\n pass # Buffered in nchan\n else:\n click.echo(\"Failed to submit event to push server, server responded %d\" % (\n notification.status_code))\n except requests.exceptions.ConnectionError:\n click.echo(\"Failed to submit event to push server, connection error\")\n\n\nclass PushLogHandler(logging.Handler):\n \"\"\"\n To be used with Python log handling framework for publishing log entries\n \"\"\"\n def emit(self, record):\n from certidude.push import publish\n publish(\"log-entry\", dict(\n created = datetime.utcfromtimestamp(record.created),\n message = record.msg % record.args,\n severity = record.levelname.lower()))\n\n","new_contents":"\nimport click\nimport json\nimport logging\nimport requests\nfrom datetime import datetime\nfrom certidude import config\n\n\ndef publish(event_type, event_data):\n \"\"\"\n Publish event on push server\n \"\"\"\n if not config.PUSH_PUBLISH:\n # Push server disabled\n return\n\n if not isinstance(event_data, basestring):\n from certidude.decorators import MyEncoder\n event_data = json.dumps(event_data, cls=MyEncoder)\n\n url = config.PUSH_PUBLISH % config.PUSH_TOKEN\n click.echo(\"Publishing %s event '%s' on %s\" % (event_type, event_data, url))\n\n try:\n notification = requests.post(\n url,\n data=event_data,\n headers={\"X-EventSource-Event\": event_type, \"User-Agent\": \"Certidude API\"})\n if notification.status_code == requests.codes.created:\n pass # Sent to client\n elif notification.status_code == requests.codes.accepted:\n pass # Buffered in nchan\n else:\n click.echo(\"Failed to submit event to push server, server responded %d\" % (\n notification.status_code))\n except requests.exceptions.ConnectionError:\n click.echo(\"Failed to submit event to push server, connection error\")\n\n\nclass PushLogHandler(logging.Handler):\n \"\"\"\n To be used with Python log handling framework for publishing log entries\n \"\"\"\n def emit(self, record):\n from certidude.push import publish\n publish(\"log-entry\", dict(\n created = datetime.utcfromtimestamp(record.created),\n message = record.msg % record.args,\n severity = record.levelname.lower()))\n\n","subject":"Add fallbacks for e-mail handling if outbox is not defined","message":"Add fallbacks for e-mail handling if outbox is not defined\n","lang":"Python","license":"mit","repos":"laurivosandi\/certidude,laurivosandi\/certidude,plaes\/certidude,laurivosandi\/certidude,plaes\/certidude,plaes\/certidude,laurivosandi\/certidude,plaes\/certidude"} {"commit":"78515c7bbb81263fa339a67c2aabfa1a4f3c9af9","old_file":"thefuck\/rules\/ifconfig_device_not_found.py","new_file":"thefuck\/rules\/ifconfig_device_not_found.py","old_contents":"import subprocess\nfrom thefuck.utils import for_app, replace_command, eager\n\n\n@for_app('ifconfig')\ndef match(command):\n return 'error fetching interface information: Device not found' \\\n in command.stderr\n\n\n@eager\ndef _get_possible_interfaces():\n proc = subprocess.Popen(['ifconfig', '-a'], stdout=subprocess.PIPE)\n for line in proc.stdout.readlines():\n line = line.decode()\n if line and line != '\\n' and not line.startswith(' '):\n yield line.split(' ')[0]\n\n\ndef get_new_command(command):\n interface = command.stderr.split(' ')[0][:-1]\n possible_interfaces = _get_possible_interfaces()\n return replace_command(command, interface, possible_interfaces)\n\n\n","new_contents":"import subprocess\nfrom thefuck.utils import for_app, replace_command, eager\n\n\n@for_app('ifconfig')\ndef match(command):\n return 'error fetching interface information: Device not found' \\\n in command.stderr\n\n\n@eager\ndef _get_possible_interfaces():\n proc = subprocess.Popen(['ifconfig', '-a'], stdout=subprocess.PIPE)\n for line in proc.stdout.readlines():\n line = line.decode()\n if line and line != '\\n' and not line.startswith(' '):\n yield line.split(' ')[0]\n\n\ndef get_new_command(command):\n interface = command.stderr.split(' ')[0][:-1]\n possible_interfaces = _get_possible_interfaces()\n return replace_command(command, interface, possible_interfaces)\n","subject":"Fix flake8 errors: W391 blank line at end of file","message":"Fix flake8 errors: W391 blank line at end of file\n","lang":"Python","license":"mit","repos":"nvbn\/thefuck,scorphus\/thefuck,Clpsplug\/thefuck,mlk\/thefuck,nvbn\/thefuck,SimenB\/thefuck,mlk\/thefuck,SimenB\/thefuck,Clpsplug\/thefuck,scorphus\/thefuck"} {"commit":"e7e8972124d3336834f1c177f655e12528a49624","old_file":"cosmo\/monitors\/osm_data_models.py","new_file":"cosmo\/monitors\/osm_data_models.py","old_contents":"import pandas as pd\n\nfrom monitorframe.monitor import BaseDataModel\n\nfrom cosmo.filesystem import FileDataFinder\nfrom cosmo import FILES_SOURCE\nfrom cosmo.monitor_helpers import explode_df\n\n\nclass OSMDataModel(BaseDataModel):\n\n def get_data(self):\n header_keys = (\n 'ROOTNAME', 'EXPSTART', 'DETECTOR', 'LIFE_ADJ', 'OPT_ELEM', 'CENWAVE', 'FPPOS', 'PROPOSID', 'OBSET_ID'\n )\n header_extensions = (0, 1, 0, 0, 0, 0, 0, 0, 0)\n\n data_keys = ('TIME', 'SHIFT_DISP', 'SHIFT_XDISP', 'SEGMENT')\n data_extensions = (1, 1, 1, 1)\n\n finder = FileDataFinder(\n FILES_SOURCE,\n '*lampflash*',\n header_keys,\n header_extensions,\n data_keys=data_keys,\n data_extensions=data_extensions\n )\n\n df = pd.DataFrame(finder.data_from_files())\n\n return explode_df(df, list(data_keys))\n","new_contents":"import pandas as pd\n\nfrom monitorframe.monitor import BaseDataModel\n\nfrom cosmo.filesystem import FileDataFinder\nfrom cosmo import FILES_SOURCE\nfrom cosmo.monitor_helpers import explode_df\n\n\nclass OSMDataModel(BaseDataModel):\n \"\"\"Data model for all OSM Shift monitors.\"\"\"\n\n def get_data(self):\n header_keys = (\n 'ROOTNAME', 'EXPSTART', 'DETECTOR', 'LIFE_ADJ', 'OPT_ELEM', 'CENWAVE', 'FPPOS', 'PROPOSID', 'OBSET_ID'\n )\n header_extensions = (0, 1, 0, 0, 0, 0, 0, 0, 0)\n\n data_keys = ('TIME', 'SHIFT_DISP', 'SHIFT_XDISP', 'SEGMENT')\n data_extensions = (1, 1, 1, 1)\n\n # Find data from lampflash files\n finder = FileDataFinder(\n FILES_SOURCE,\n '*lampflash*',\n header_keys,\n header_extensions,\n data_keys=data_keys,\n data_extensions=data_extensions\n )\n\n df = pd.DataFrame(finder.data_from_files())\n\n return explode_df(df, list(data_keys))\n","subject":"Add comments and docstring to OSMDataModel","message":"Add comments and docstring to OSMDataModel\n","lang":"Python","license":"bsd-3-clause","repos":"justincely\/cos_monitoring"} {"commit":"3e67993eb17aca7571381d59b7fd65eab53dac98","old_file":"day19\/part2.py","new_file":"day19\/part2.py","old_contents":"inp = 3004953\n\nelves = list(range(1, inp + 1))\ni = 0\nwhile len(elves) > 1:\n index = (i + int(len(elves) \/ 2)) % len(elves)\n elves.pop(index)\n if index < i:\n i -= 1\n i = (i + 1) % len(elves)\n\nprint(elves[0])\ninput()\n","new_contents":"inp = 3004953\n\nclass Elf:\n def __init__(self, num):\n self.num = num\n self.prev = None\n self.next = None\n\n def remove(self):\n self.prev.next = self.next\n self.next.prev = self.prev\n\nelves = list(map(Elf, range(1, inp + 1)))\nfor i in range(inp):\n elves[i].prev = elves[(i - 1) % inp]\n elves[i].next = elves[(i + 1) % inp]\n\ncount, current, across = inp, elves[0], elves[inp \/\/ 2]\nwhile current != across:\n across.remove()\n across = across.next\n if count % 2 == 1:\n across = across.next\n count -= 1\n current = current.next\n\nprint(current.num)\ninput()\n","subject":"Replace list with a linked list for much better performance","message":"Replace list with a linked list for much better performance\n","lang":"Python","license":"unlicense","repos":"ultramega\/adventofcode2016"} {"commit":"561d98e59ea46b56d50341e06578b5c9fe95c73a","old_file":"perfbucket\/watcher.py","new_file":"perfbucket\/watcher.py","old_contents":"import os\nimport sys\nimport pyinotify\nimport analyzer\n\nwm = pyinotify.WatchManager()\n\nclass ProcessProfilerEvent(pyinotify.ProcessEvent):\n def process_IN_CLOSE_WRITE(self, event):\n if event.name.endswith(\".json\"):\n base = os.path.splitext(os.path.join(event.path, event.name))[0]\n analyzer.analyze_profiling_result(base)\n\ndef monitor(directory):\n notifier = pyinotify.Notifier(wm, ProcessProfilerEvent())\n mask = pyinotify.IN_CLOSE_WRITE # Watched events\n wdd = wm.add_watch(directory, mask)\n \n while True:\n try:\n # process the queue of events as explained above\n notifier.process_events()\n if notifier.check_events():\n # read notified events and enqeue them\n notifier.read_events()\n # you can do some tasks here...\n except KeyboardInterrupt:\n # destroy the inotify's instance on this interrupt (stop monitoring)\n notifier.stop()\n break\n\nif __name__ == '__main__':\n monitor(sys.argv[1])\n","new_contents":"import os\nimport sys\nimport pyinotify\nimport analyzer\n\nclass ProcessProfilerEvent(pyinotify.ProcessEvent):\n def process_IN_CLOSE_WRITE(self, event):\n if event.name.endswith(\".json\"):\n base = os.path.splitext(os.path.join(event.path, event.name))[0]\n analyzer.analyze_profiling_result(base)\n\ndef monitor(directory):\n wm = pyinotify.WatchManager()\n notifier = pyinotify.Notifier(wm, ProcessProfilerEvent())\n mask = pyinotify.IN_CLOSE_WRITE # Watched events\n wdd = wm.add_watch(directory, mask)\n \n while True:\n try:\n # process the queue of events as explained above\n notifier.process_events()\n if notifier.check_events():\n # read notified events and enqeue them\n notifier.read_events()\n # you can do some tasks here...\n except KeyboardInterrupt:\n # destroy the inotify's instance on this interrupt (stop monitoring)\n notifier.stop()\n break\n\nif __name__ == '__main__':\n monitor(sys.argv[1])\n","subject":"Change scope of watch manager.","message":"Change scope of watch manager.\n","lang":"Python","license":"agpl-3.0","repos":"davidstrauss\/perfbucket,davidstrauss\/perfbucket,davidstrauss\/perfbucket"} {"commit":"2497f494f0e3e7fb57aa8cb1deed0c05fd6b74b1","old_file":"handler\/FilesService.py","new_file":"handler\/FilesService.py","old_contents":"import tornado\nimport time\nfrom bson.json_util import dumps\nfrom tornado.options import options\n\n\nclass FilesServiceHandler(tornado.web.RequestHandler):\n def initialize(self, logger, mongodb):\n self.logger = logger\n self.mongodb = mongodb\n\n @tornado.web.asynchronous\n @tornado.gen.coroutine\n def post(self):\n self.logger.info('Request to file upload')\n\n for item in self.request.files.values():\n file_info = item[0]\n\n self.logger.info('File uploaded: %s with mime type %s' % (file_info['filename'], file_info['content_type']))\n name = '%s-%s' % (time.time(), file_info['filename'])\n\n with open('%s\/%s' % (options.files_dir, name), 'w') as f:\n f.write(file_info['body'])\n\n self.logger.info('File saved at %s' % name)\n\n self.write('done')\n","new_contents":"import tornado\nimport time\nfrom bson.json_util import dumps\nfrom tornado.options import options\n\n\nclass FilesServiceHandler(tornado.web.RequestHandler):\n def initialize(self, logger, mongodb):\n self.logger = logger\n self.mongodb = mongodb[options.db_name]['Files']\n\n @tornado.web.asynchronous\n @tornado.gen.coroutine\n def post(self):\n self.logger.info('Request to file upload')\n\n result = []\n\n for item in self.request.files.values():\n for file_info in item:\n timestamp = time.time()\n\n data = {\n 'name': '%s-%s' % (timestamp, file_info['filename']),\n 'location': 'TBD',\n 'context': 'context',\n 'realName': file_info['filename'],\n 'mimeType': file_info['content_type'],\n 'deleted': False,\n 'timestamp': timestamp,\n 'restrictions': {\n 'quota': False,\n 'session': False\n }\n }\n\n self.logger.info('File uploaded: %s with mime type %s' % (data['realName'], data['mimeType']))\n\n with open('%s\/%s' % (options.files_dir, data['name']), 'w') as f:\n f.write(file_info['body'])\n\n self.logger.info('File saved at %s' % data['name'])\n\n yield self.mongodb.save(data)\n result.append(data)\n\n self.write(dumps(result))\n","subject":"Save file info in DB","message":"Save file info in DB\n","lang":"Python","license":"apache-2.0","repos":"jiss-software\/jiss-file-service,jiss-software\/jiss-file-service,jiss-software\/jiss-file-service"} {"commit":"996713fc6aefe20b28c729c46532ae566d5160a1","old_file":"paratemp\/sim_setup\/__init__.py","new_file":"paratemp\/sim_setup\/__init__.py","old_contents":"\"\"\"This module has functions and classes useful for setting up simulations\"\"\"\n\n########################################################################\n# #\n# This test was written by Thomas Heavey in 2019. #\n# theavey@bu.edu thomasjheavey@gmail.com #\n# #\n# Copyright 2017-19 Thomas J. Heavey IV #\n# #\n# Licensed under the Apache License, Version 2.0 (the \"License\"); #\n# you may not use this file except in compliance with the License. #\n# You may obtain a copy of the License at #\n# #\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0 #\n# #\n# Unless required by applicable law or agreed to in writing, software #\n# distributed under the License is distributed on an \"AS IS\" BASIS, #\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or #\n# implied. #\n# See the License for the specific language governing permissions and #\n# limitations under the License. #\n# #\n########################################################################\n\nfrom __future__ import absolute_import\n\nfrom .para_temp_setup import *\nfrom .sim_setup import *\nfrom .simulation import Simulation, SimpleSimulation\nfrom .pt_simulation import PTSimulation\nfrom .molecule import Molecule\nfrom .system import System\n","new_contents":"\"\"\"This module has functions and classes useful for setting up simulations\"\"\"\n\n########################################################################\n# #\n# This test was written by Thomas Heavey in 2019. #\n# theavey@bu.edu thomasjheavey@gmail.com #\n# #\n# Copyright 2017-19 Thomas J. Heavey IV #\n# #\n# Licensed under the Apache License, Version 2.0 (the \"License\"); #\n# you may not use this file except in compliance with the License. #\n# You may obtain a copy of the License at #\n# #\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0 #\n# #\n# Unless required by applicable law or agreed to in writing, software #\n# distributed under the License is distributed on an \"AS IS\" BASIS, #\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or #\n# implied. #\n# See the License for the specific language governing permissions and #\n# limitations under the License. #\n# #\n########################################################################\n\nfrom __future__ import absolute_import\n\nfrom .para_temp_setup import *\nfrom .sim_setup import *\nfrom .molecule import Molecule\nfrom .system import System\nfrom .simulation import Simulation, SimpleSimulation\nfrom .pt_simulation import PTSimulation\n","subject":"Fix import order (some somewhat cyclic dependencies; should fix)","message":"Fix import order (some somewhat cyclic dependencies; should fix)\n","lang":"Python","license":"apache-2.0","repos":"theavey\/ParaTemp,theavey\/ParaTemp"} {"commit":"d787039a58d63cd85068da996a12fc36c1d63804","old_file":"ixxy_admin_utils\/admin_actions.py","new_file":"ixxy_admin_utils\/admin_actions.py","old_contents":"def xlsx_export_action(modeladmin, request, queryset):\n\n from django.http import HttpResponse\n from import_export.formats import base_formats\n formats = modeladmin.get_export_formats()\n file_format = base_formats.XLSX()\n\n export_data = modeladmin.get_export_data(file_format, queryset, request=request)\n content_type = file_format.get_content_type()\n # Django 1.7 uses the content_type kwarg instead of mimetype\n try:\n response = HttpResponse(export_data, content_type=content_type)\n except TypeError:\n response = HttpResponse(export_data, mimetype=content_type)\n response['Content-Disposition'] = 'attachment; filename=%s' % (\n modeladmin.get_export_filename(file_format),\n )\n return response\nxlsx_export_action.short_description = \"Export selected rows to Excel\"","new_contents":"from django.http import HttpResponse\nfrom import_export.formats import base_formats\n\n\ndef xlsx_export_action(modeladmin, request, queryset):\n\n formats = modeladmin.get_export_formats()\n file_format = base_formats.XLSX()\n\n export_data = modeladmin.get_export_data(file_format, queryset, request=request)\n content_type = file_format.get_content_type()\n # Django 1.7 uses the content_type kwarg instead of mimetype\n try:\n response = HttpResponse(export_data, content_type=content_type)\n except TypeError:\n response = HttpResponse(export_data, mimetype=content_type)\n response['Content-Disposition'] = 'attachment; filename=%s' % (\n modeladmin.get_export_filename(file_format),\n )\n return response\nxlsx_export_action.short_description = \"Export selected rows to Excel\"","subject":"Undo previous commit. It didn't work.","message":"Undo previous commit. It didn't work.\n","lang":"Python","license":"mit","repos":"DjangoAdminHackers\/ixxy-admin-utils,DjangoAdminHackers\/ixxy-admin-utils"} {"commit":"7dfe4381ecd252530cb7dc274b2dc6aaa39f81cc","old_file":"deps\/pyextensibletype\/extensibletype\/test\/test_interning.py","new_file":"deps\/pyextensibletype\/extensibletype\/test\/test_interning.py","old_contents":"from .. import intern\n\ndef test_global_interning():\n try:\n intern.global_intern(\"hello\")\n except AssertionError as e:\n pass\n else:\n raise Exception(\"Expects complaint about uninitialized table\")\n\n intern.global_intern_initialize()\n id1 = intern.global_intern(\"hello\")\n id2 = intern.global_intern(\"hello\")\n id3 = intern.global_intern(\"hallo\")\n assert id1 == id2\n assert id1 != id3\n\ndef test_interning():\n table = intern.InternTable()\n\n id1 = intern.global_intern(\"hello\")\n id2 = intern.global_intern(\"hello\")\n id3 = intern.global_intern(\"hallo\")\n assert id1 == id2\n assert id1 != id3\n\ndef test_intern_many():\n table = intern.InternTable()\n\n for i in range(1000000):\n table.intern(\"my randrom string %d\" % i)\n table.intern(\"my randrom string %d\" % (i \/\/ 2))\n table.intern(\"my randrom string %d\" % (i \/\/ 4))\n","new_contents":"from .. import intern\n\ndef test_global_interning():\n # Can't really test for this with nose...\n # try:\n # intern.global_intern(\"hello\")\n # except AssertionError as e:\n # pass\n # else:\n # raise Exception(\"Expects complaint about uninitialized table\")\n\n intern.global_intern_initialize()\n id1 = intern.global_intern(\"hello\")\n id2 = intern.global_intern(\"hello\")\n id3 = intern.global_intern(\"hallo\")\n assert id1 == id2\n assert id1 != id3\n\ndef test_interning():\n table = intern.InternTable()\n\n id1 = intern.global_intern(\"hello\")\n id2 = intern.global_intern(\"hello\")\n id3 = intern.global_intern(\"hallo\")\n assert id1 == id2\n assert id1 != id3\n\ndef test_intern_many():\n table = intern.InternTable()\n\n for i in range(1000000):\n table.intern(\"my randrom string %d\" % i)\n table.intern(\"my randrom string %d\" % (i \/\/ 2))\n table.intern(\"my randrom string %d\" % (i \/\/ 4))\n","subject":"Disable global intern exception test","message":"Disable global intern exception test\n","lang":"Python","license":"bsd-2-clause","repos":"stuartarchibald\/numba,pitrou\/numba,pitrou\/numba,shiquanwang\/numba,seibert\/numba,jriehl\/numba,shiquanwang\/numba,stefanseefeld\/numba,stuartarchibald\/numba,cpcloud\/numba,cpcloud\/numba,gdementen\/numba,seibert\/numba,numba\/numba,sklam\/numba,ssarangi\/numba,gmarkall\/numba,cpcloud\/numba,sklam\/numba,pitrou\/numba,IntelLabs\/numba,pombredanne\/numba,cpcloud\/numba,gdementen\/numba,gdementen\/numba,seibert\/numba,IntelLabs\/numba,pombredanne\/numba,seibert\/numba,GaZ3ll3\/numba,gmarkall\/numba,jriehl\/numba,stuartarchibald\/numba,pitrou\/numba,ssarangi\/numba,ssarangi\/numba,jriehl\/numba,numba\/numba,gdementen\/numba,pombredanne\/numba,stefanseefeld\/numba,sklam\/numba,gmarkall\/numba,stonebig\/numba,seibert\/numba,pitrou\/numba,stefanseefeld\/numba,jriehl\/numba,numba\/numba,shiquanwang\/numba,ssarangi\/numba,stonebig\/numba,jriehl\/numba,GaZ3ll3\/numba,GaZ3ll3\/numba,sklam\/numba,pombredanne\/numba,GaZ3ll3\/numba,numba\/numba,gmarkall\/numba,sklam\/numba,stefanseefeld\/numba,stonebig\/numba,stefanseefeld\/numba,gdementen\/numba,cpcloud\/numba,numba\/numba,IntelLabs\/numba,stuartarchibald\/numba,ssarangi\/numba,stuartarchibald\/numba,gmarkall\/numba,stonebig\/numba,stonebig\/numba,pombredanne\/numba,IntelLabs\/numba,GaZ3ll3\/numba,IntelLabs\/numba"} {"commit":"5bece700c7ebbb2c9ea3ce2781863baf189e2fc0","old_file":"cybox\/test\/objects\/__init__.py","new_file":"cybox\/test\/objects\/__init__.py","old_contents":"# Copyright (c) 2013, The MITRE Corporation. All rights reserved.\n# See LICENSE.txt for complete terms.\n\nimport cybox.utils\n\n\nclass ObjectTestCase(object):\n \"\"\"A base class for testing all subclasses of ObjectProperties.\n\n Each subclass of ObjectTestCase should subclass both unittest.TestCase\n and ObjectTestCase, and defined two class-level fields:\n - klass: the ObjectProperties subclass being tested\n - object_type: The name prefix used in the XML Schema bindings for the\n object.\n \"\"\"\n\n def test_type_exists(self):\n # Verify that the correct class has been added to the OBJECT_TYPES_DICT\n # dictionary in cybox.utils.nsparser\n\n # Skip this base class\n if type(self) == type(ObjectTestCase):\n return\n\n t = self.__class__.object_type\n\n expected_class = cybox.utils.get_class_for_object_type(t)\n actual_class = self.__class__.klass\n\n self.assertEqual(expected_class, actual_class)\n\n expected_namespace = expected_class._XSI_NS\n actual_namespace = cybox.utils.nsparser.OBJECT_TYPES_DICT.get(t).get('namespace_prefix')\n self.assertEqual(expected_namespace, actual_namespace)\n\n self.assertEqual(expected_class._XSI_TYPE, t)\n","new_contents":"# Copyright (c) 2013, The MITRE Corporation. All rights reserved.\n# See LICENSE.txt for complete terms.\n\nimport cybox.test\nimport cybox.utils\n\n\nclass ObjectTestCase(object):\n \"\"\"A base class for testing all subclasses of ObjectProperties.\n\n Each subclass of ObjectTestCase should subclass both unittest.TestCase\n and ObjectTestCase, and defined two class-level fields:\n - klass: the ObjectProperties subclass being tested\n - object_type: The name prefix used in the XML Schema bindings for the\n object.\n \"\"\"\n\n def test_type_exists(self):\n # Verify that the correct class has been added to the OBJECT_TYPES_DICT\n # dictionary in cybox.utils.nsparser\n\n # Skip this base class\n if type(self) == type(ObjectTestCase):\n return\n\n t = self.__class__.object_type\n\n expected_class = cybox.utils.get_class_for_object_type(t)\n actual_class = self.__class__.klass\n\n self.assertEqual(expected_class, actual_class)\n\n expected_namespace = expected_class._XSI_NS\n actual_namespace = cybox.utils.nsparser.OBJECT_TYPES_DICT.get(t).get('namespace_prefix')\n self.assertEqual(expected_namespace, actual_namespace)\n\n self.assertEqual(expected_class._XSI_TYPE, t)\n\n def test_object_reference(self):\n klass = self.__class__.klass\n\n ref_dict = {'object_reference': \"some:object-reference-1\",\n 'xsi:type': klass._XSI_TYPE}\n\n ref_dict2 = cybox.test.round_trip_dict(klass, ref_dict)\n self.assertEqual(ref_dict, ref_dict2)\n","subject":"Add (failing) test of object_reference on all ObjectProperties subclasses","message":"Add (failing) test of object_reference on all ObjectProperties subclasses\n","lang":"Python","license":"bsd-3-clause","repos":"CybOXProject\/python-cybox"} {"commit":"a354a4f52bce3c3063678b046ba76a694c076652","old_file":"web\/celSearch\/api\/scripts\/query_wikipedia.py","new_file":"web\/celSearch\/api\/scripts\/query_wikipedia.py","old_contents":"''' \nScript used to query Wikipedia for summary of object \n'''\n\nimport sys\nimport wikipedia \n\ndef main():\n\t# Check that we have the right number of arguments \n\tif (len(sys.argv) != 2): \n\t\tprint 'Incorrect number of arguments; please pass in only one string that contains the subject'\n\t \treturn 'Banana'\n\n\tprint wikipedia.summary(sys.argv[1])\n\t#return wikipedia.summary(sys.argv[1])\n\n\nif __name__ == '__main__':\n\tmain()","new_contents":"''' \nScript used to query Wikipedia for summary of object \n'''\n\nimport sys\nimport wikipedia \nimport nltk \n\ndef main():\n\t# Check that we have the right number of arguments \n\tif (len(sys.argv) != 2): \n\t\tprint 'Incorrect number of arguments; please pass in only one string that contains the query'\n\t \treturn 'Banana'\n\n\t# Get the noun from the query (uses the first noun it finds for now)\n\tprint sys.argv[0]\n\ttokens = nltk.word_tokenize(sys.argv[1])\n\ttagged = nltk.pos_tag(tokens)\n\n\t# Find first noun in query and provide Wikipedia summary for it \n\tfor tag in tagged: \n\t\tif tag[1][0] == 'N': \n\t\t\tprint wikipedia.summary(tag[0])\n\t\t\treturn \n\n\nif __name__ == '__main__':\n\tmain()","subject":"Add nltk part to script","message":"Add nltk part to script\n","lang":"Python","license":"apache-2.0","repos":"christopher18\/Celsearch,christopher18\/Celsearch,christopher18\/Celsearch"} {"commit":"2a7ce1ac70f8767e9d2b2a9f1d335cfcc63a92b6","old_file":"rplugin\/python3\/LanguageClient\/logger.py","new_file":"rplugin\/python3\/LanguageClient\/logger.py","old_contents":"import logging\nimport tempfile\n\nlogger = logging.getLogger(\"LanguageClient\")\nwith tempfile.NamedTemporaryFile(\n prefix=\"LanguageClient-\",\n suffix=\".log\", delete=False) as tmp:\n tmpname = tmp.name\nfileHandler = logging.FileHandler(filename=tmpname)\nfileHandler.setFormatter(\n logging.Formatter(\n \"%(asctime)s %(levelname)-8s %(message)s\",\n \"%H:%M:%S\"))\nlogger.addHandler(fileHandler)\nlogger.setLevel(logging.WARN)\n","new_contents":"import logging\n\nlogger = logging.getLogger(\"LanguageClient\")\nfileHandler = logging.FileHandler(filename=\"\/tmp\/LanguageClient.log\")\nfileHandler.setFormatter(\n logging.Formatter(\n \"%(asctime)s %(levelname)-8s %(message)s\",\n \"%H:%M:%S\"))\nlogger.addHandler(fileHandler)\nlogger.setLevel(logging.WARN)\n","subject":"Revert \"Use tempfile lib for log file\"","message":"Revert \"Use tempfile lib for log file\"\n\nThis reverts commit 6e8f35b83fc563c8349cb3be040c61a0588ca745.\n\nThe commit caused severer issue than it fixed. In case one need to check\nthe content of log file, there is no way to tell where the log file\nlocation\/name is.\n","lang":"Python","license":"mit","repos":"autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim,autozimu\/LanguageClient-neovim"} {"commit":"21a392df73324f111fa80e2fd8ce88b0e32c954c","old_file":"python\/algorithms\/fibonacci.py","new_file":"python\/algorithms\/fibonacci.py","old_contents":"def fib1(amount):\n \"\"\"\n Fibonacci generator example. The second variable is used to store\n the result.\n :param amount: Amount of numbers to produce.\n :return: Generator.\n\n >>> list(fib1(0))\n []\n >>> list(fib1(1))\n [0]\n >>> list(fib1(3))\n [0, 1, 1]\n >>> list(fib1(9))\n [0, 1, 1, 2, 3, 5, 8, 13, 21]\n \"\"\"\n first, second = 0, 1\n for _ in range(amount):\n yield first\n first, second = second + first, first\n\n\ndef fib2(amount):\n \"\"\"\n Fibonacci generator example. The first variable is used to store\n the result.\n :param amount: Amount of numbers to produce.\n :return: Generator.\n\n >>> list(fib2(0))\n []\n >>> list(fib2(1))\n [0]\n >>> list(fib2(3))\n [0, 1, 1]\n >>> list(fib2(9))\n [0, 1, 1, 2, 3, 5, 8, 13, 21]\n \"\"\"\n first, second = 1, 0\n for _ in range(amount):\n first, second = second, first + second\n yield first\n\n\nif __name__ == '__main__':\n import doctest\n\n doctest.testmod()\n","new_contents":"\"\"\"Implementations calculation of Fibonacci numbers.\"\"\"\n\n\ndef fib1(amount):\n \"\"\"\n Calculate Fibonacci numbers.\n\n The second variable is used to store the result.\n :param amount: Amount of numbers to produce.\n :return: Generator.\n\n >>> list(fib1(0))\n []\n >>> list(fib1(1))\n [0]\n >>> list(fib1(3))\n [0, 1, 1]\n >>> list(fib1(9))\n [0, 1, 1, 2, 3, 5, 8, 13, 21]\n \"\"\"\n first, second = 0, 1\n for _ in range(amount):\n yield first\n first, second = second + first, first\n\n\ndef fib2(amount):\n \"\"\"\n Calculate Fibonacci numbers.\n\n The first variable is used to store the result.\n :param amount: Amount of numbers to produce.\n :return: Generator.\n\n >>> list(fib2(0))\n []\n >>> list(fib2(1))\n [0]\n >>> list(fib2(3))\n [0, 1, 1]\n >>> list(fib2(9))\n [0, 1, 1, 2, 3, 5, 8, 13, 21]\n \"\"\"\n first, second = 1, 0\n for _ in range(amount):\n first, second = second, first + second\n yield first\n\n\nif __name__ == '__main__':\n import doctest\n\n doctest.testmod()\n","subject":"Adjust doc strings in Fibonacci numbers implementation","message":"Adjust doc strings in Fibonacci numbers implementation\n","lang":"Python","license":"mit","repos":"pesh1983\/exercises,pesh1983\/exercises"} {"commit":"9c1190133a680717850a4d0f46a96591b7be4e33","old_file":"autoencoder\/api.py","new_file":"autoencoder\/api.py","old_contents":"from .io import preprocess\nfrom .train import train\nfrom .encode import encode\n\n\ndef autoencode(count_matrix, kfold=None, reduced=False,\n censor_matrix=None, type='normal',\n learning_rate=1e-2,\n hidden_size=10,\n epochs=10):\n\n x = preprocess(count_matrix, kfold=kfold, censor=censor_matrix)\n model = train(x, hidden_size=hidden_size, learning_rate=learning_rate,\n aetype=type, epochs=epochs)\n encoded = encode(count_matrix, model, reduced=reduced)\n\n return encoded\n","new_contents":"from .io import preprocess\nfrom .train import train\nfrom .encode import encode\n\n\ndef autoencode(count_matrix, kfold=None, reduced=False,\n mask=None, type='normal',\n learning_rate=1e-2,\n hidden_size=10,\n epochs=10):\n\n x = preprocess(count_matrix, kfold=kfold, mask=mask)\n model = train(x, hidden_size=hidden_size, learning_rate=learning_rate,\n aetype=type, epochs=epochs)\n encoded = encode(count_matrix, model, reduced=reduced)\n\n return encoded\n","subject":"Change mask parameter in API.","message":"Change mask parameter in API.\n","lang":"Python","license":"apache-2.0","repos":"theislab\/dca,theislab\/dca,theislab\/dca"} {"commit":"8d7e4cf37e73c1ff9827e94a06327921f553e2f4","old_file":"learntools\/computer_vision\/ex4.py","new_file":"learntools\/computer_vision\/ex4.py","old_contents":"from learntools.core import *\nimport tensorflow as tf\n\n\nclass Q1A(ThoughtExperiment):\n _solution = \"\"\n\n\nclass Q1B(ThoughtExperiment):\n _solution = \"\"\n\nQ1 = MultipartProblem(Q1A, Q1B)\n\nclass Q2A(ThoughtExperiment):\n _hint = r\"Stacking the second layer expanded the receptive field by one neuron on each side, giving $3+1+1=5$ for each dimension. If you expanded by one neuron again, what would you get?\"\n _solution = r\"The third layer would have a $7 \\times 7$ receptive field.\"\n\nclass Q2B(ThoughtExperiment):\n _hint = r\"This pooling layer collapses a $2 \\times 2$ patch into a single pixel, effectively *doubling* the number of connections along each dimension. \"\n _solution = r\"Doubling a $7 \\times 7$ field produces a $14 \\times 14$ field for the final outputs.\"\n\nQ2 = MultipartProblem(Q2A, Q2B)\n\n\nclass Q3(CodingProblem):\n _hint = \"You just need a list of numbers, maybe three to five.\"\n _solution = CS(\"\"\"\nkernel = tf.constant([0.1, 0.2, 0.3, 0.4])\n\"\"\")\n def check(self):\n pass\n\n\nqvars = bind_exercises(globals(), [\n Q1, Q2, Q3,\n ],\n var_format='q_{n}',\n)\n__all__ = list(qvars)\n \n","new_contents":"from learntools.core import *\nimport tensorflow as tf\n\n\n# Free\nclass Q1(CodingProblem):\n _solution = \"\"\n def check(self):\n pass\n\nclass Q2A(ThoughtExperiment):\n _hint = r\"Stacking the second layer expanded the receptive field by one neuron on each side, giving $3+1+1=5$ for each dimension. If you expanded by one neuron again, what would you get?\"\n _solution = r\"The third layer would have a $7 \\times 7$ receptive field.\"\n\nclass Q2B(ThoughtExperiment):\n _hint = r\"This pooling layer collapses a $2 \\times 2$ patch into a single pixel, effectively *doubling* the number of connections along each dimension. \"\n _solution = r\"Doubling a $7 \\times 7$ field produces a $14 \\times 14$ field for the final outputs.\"\n\nQ2 = MultipartProblem(Q2A, Q2B)\n\n\nclass Q3(CodingProblem):\n _hint = \"You just need a list of numbers, maybe three to five.\"\n _solution = CS(\"\"\"\nkernel = tf.constant([0.1, 0.2, 0.3, 0.4])\n\"\"\")\n def check(self):\n pass\n\n\nqvars = bind_exercises(globals(), [\n Q1, Q2, Q3,\n ],\n var_format='q_{n}',\n)\n__all__ = list(qvars)\n \n","subject":"Change exercise 4 question 1","message":"Change exercise 4 question 1\n","lang":"Python","license":"apache-2.0","repos":"Kaggle\/learntools,Kaggle\/learntools"} {"commit":"2ee895c61f546f83f4b7fa0c6a2ba72578c378be","old_file":"problem_2\/solution.py","new_file":"problem_2\/solution.py","old_contents":"f1, f2, s, n = 0, 1, 0, 4000000\nwhile f2 < n:\n f2, f1 = f1, f1 + f2\n if f2 % 2 == 0:\n s += f2\nprint s\n","new_contents":"def sum_even_fibonacci_numbers_1():\n f1, f2, s, = 0, 1, 0,\n while f2 < 4000000:\n f2, f1 = f1, f1 + f2\n if f2 % 2 == 0:\n s += f2\n return s\ndef sum_even_fibonacci_numbers_2():\n s, a, b = 0, 1, 1\n c = a + b\n while c < 4000000:\n s += c\n a = b + c\n b = a + c\n c = a + b\n return s\n","subject":"Add a second Python implementation of problem 2","message":"Add a second Python implementation of problem 2\n","lang":"Python","license":"mit","repos":"mdsrosa\/project_euler,mdsrosa\/project_euler,mdsrosa\/project_euler,mdsrosa\/project_euler,mdsrosa\/project_euler,mdsrosa\/project_euler,mdsrosa\/project_euler,mdsrosa\/project_euler"} {"commit":"5523ae2278bb0ca055ef7a6e218ac40ed4172bf3","old_file":"webapp\/byceps\/blueprints\/ticket\/service.py","new_file":"webapp\/byceps\/blueprints\/ticket\/service.py","old_contents":"# -*- coding: utf-8 -*-\n\n\"\"\"\nbyceps.blueprints.ticket.service\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n:Copyright: 2006-2015 Jochen Kupperschmidt\n\"\"\"\n\nfrom ..party.models import Party\nfrom ..seating.models import Category\n\nfrom .models import Ticket\n\n\ndef find_ticket_for_user(user, party):\n \"\"\"Return the ticket used by the user for the party, or `None` if not\n found.\n \"\"\"\n if user.is_anonymous:\n return None\n\n return Ticket.query \\\n .filter(Ticket.used_by == user) \\\n .for_party(party) \\\n .first()\n\n\ndef get_attended_parties(user):\n \"\"\"Return the parties the user has attended.\"\"\"\n return Party.query \\\n .join(Category).join(Ticket).filter(Ticket.used_by == user) \\\n .all()\n","new_contents":"# -*- coding: utf-8 -*-\n\n\"\"\"\nbyceps.blueprints.ticket.service\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n:Copyright: 2006-2015 Jochen Kupperschmidt\n\"\"\"\n\nfrom ...database import db\n\nfrom ..party.models import Party\nfrom ..seating.models import Category\n\nfrom .models import Ticket\n\n\ndef find_ticket_for_user(user, party):\n \"\"\"Return the ticket used by the user for the party, or `None` if not\n found.\n \"\"\"\n if user.is_anonymous:\n return None\n\n return Ticket.query \\\n .filter(Ticket.used_by == user) \\\n .options(\n db.joinedload_all('occupied_seat.area'),\n ) \\\n .for_party(party) \\\n .first()\n\n\ndef get_attended_parties(user):\n \"\"\"Return the parties the user has attended.\"\"\"\n return Party.query \\\n .join(Category).join(Ticket).filter(Ticket.used_by == user) \\\n .all()\n","subject":"Save a few SQL queries.","message":"Save a few SQL queries.\n","lang":"Python","license":"bsd-3-clause","repos":"m-ober\/byceps,homeworkprod\/byceps,homeworkprod\/byceps,homeworkprod\/byceps,m-ober\/byceps,m-ober\/byceps"} {"commit":"7dc08364cbe513ce4b81483d9330789f5893fcee","old_file":"Challenges\/chall_03.py","new_file":"Challenges\/chall_03.py","old_contents":"#!\/usr\/local\/bin\/python3\n# Python challenge - 3\n# http:\/\/www.pythonchallenge.com\/pc\/def\/equality.html\n\nimport re\n\n\n'''\nHint:\nOne small letter surrounded by EXACTLY three big bodyguards on each of\nits sides.\n'''\n\n\ndef main():\n with open('bodyguard.txt', 'r') as bodyguard:\n pattern = re.compile(r'[^A-Z][A-Z]{3}([a-z])[A-Z]{3}[^A-Z]')\n littles = ''\n text = bodyguard.read()\n littles = re.findall(pattern, text)\n\n print(''.join(littles))\n return 0\n\n# Keyword: linkedlist\n\n\nif __name__ == '__main__':\n main()\n","new_contents":"#!\/usr\/local\/bin\/python3\n# Python challenge - 3\n# http:\/\/www.pythonchallenge.com\/pc\/def\/equality.html\n# Keyword: linkedlist\n\nimport re\n\n\ndef main():\n '''\n Hint:\n One small letter surrounded by EXACTLY three big bodyguards on each of\n its sides.\n Page source text saved in bodyguard.txt\n '''\n with open('bodyguard.txt', 'r') as bodyguard:\n pattern = re.compile(r'[^A-Z][A-Z]{3}([a-z])[A-Z]{3}[^A-Z]')\n littles = ''\n text = bodyguard.read()\n littles = re.findall(pattern, text)\n\n print(''.join(littles))\n return 0\n\n\nif __name__ == '__main__':\n main()\n","subject":"Refactor code, add hints from page","message":"Refactor code, add hints from page\n","lang":"Python","license":"mit","repos":"HKuz\/PythonChallenge"} {"commit":"da85d9660166f67133b10953104ccd81b89d0b92","old_file":"micawber\/cache.py","new_file":"micawber\/cache.py","old_contents":"from __future__ import with_statement\nimport os\nimport pickle\nfrom contextlib import closing\ntry:\n from redis import Redis\nexcept ImportError:\n Redis = None\n\n\nclass Cache(object):\n def __init__(self):\n self._cache = {}\n\n def get(self, k):\n return self._cache.get(k)\n\n def set(self, k, v):\n self._cache[k] = v\n\n\nclass PickleCache(Cache):\n def __init__(self, filename='cache.db'):\n self.filename = filename\n self._cache = self.load()\n \n def load(self):\n if os.path.exists(self.filename):\n with closing(open(self.filename, 'rb')) as fh:\n return pickle.load(fh)\n return {}\n\n def save(self):\n with closing(open(self.filename, 'wb')) as fh:\n pickle.dump(self._cache, fh)\n\n\nif Redis:\n class RedisCache(Cache):\n def __init__(self, namespace='micawber', **conn):\n self.namespace = namespace\n self.key_fn = lambda self, k: '%s.%s' % (self.namespace, k)\n self.conn = Redis(**conn)\n \n def get(self, k):\n cached = self.conn.get(self.key_fn(k))\n if cached:\n return pickle.loads(cached)\n \n def set(self, k, v):\n self.conn.set(self.key_fn(k), pickle.dumps(v))\n","new_contents":"from __future__ import with_statement\nimport os\nimport pickle\ntry:\n from redis import Redis\nexcept ImportError:\n Redis = None\n\n\nclass Cache(object):\n def __init__(self):\n self._cache = {}\n\n def get(self, k):\n return self._cache.get(k)\n\n def set(self, k, v):\n self._cache[k] = v\n\n\nclass PickleCache(Cache):\n def __init__(self, filename='cache.db'):\n self.filename = filename\n self._cache = self.load()\n \n def load(self):\n if os.path.exists(self.filename):\n with open(self.filename, 'rb') as fh:\n return pickle.load(fh)\n return {}\n\n def save(self):\n with open(self.filename, 'wb') as fh:\n pickle.dump(self._cache, fh)\n\n\nif Redis:\n class RedisCache(Cache):\n def __init__(self, namespace='micawber', **conn):\n self.namespace = namespace\n self.key_fn = lambda self, k: '%s.%s' % (self.namespace, k)\n self.conn = Redis(**conn)\n \n def get(self, k):\n cached = self.conn.get(self.key_fn(k))\n if cached:\n return pickle.loads(cached)\n \n def set(self, k, v):\n self.conn.set(self.key_fn(k), pickle.dumps(v))\n","subject":"Remove a redundant use of contextlib.closing() decorator","message":"Remove a redundant use of contextlib.closing() decorator\n\nRemove the unnecessary contextlib.closing() decorators from open()\ncalls. The file objects returned by open() provide context manager API\nthemselves and closing() is only necessary for external file-like\nobjects that do not support it.\n\nThis should work even in Python 2.6, see:\nhttps:\/\/docs.python.org\/2.6\/library\/stdtypes.html#file.close\n","lang":"Python","license":"mit","repos":"coleifer\/micawber,coleifer\/micawber"} {"commit":"2c7dc769874766b230bc11c7ec6f67d3c1157005","old_file":"duplicatefiledir\/__init__.py","new_file":"duplicatefiledir\/__init__.py","old_contents":"from fman import DirectoryPaneCommand, show_alert\r\nimport distutils\r\nfrom distutils import dir_util, file_util\r\nimport os.path\r\n\r\nclass DuplicateFileDir(DirectoryPaneCommand):\r\n def __call__(self):\r\n selected_files = self.pane.get_selected_files()\r\n if len(selected_files) >= 1 or (len(selected_files) == 0 and self.get_chosen_files()):\r\n if len(selected_files) == 0 and self.get_chosen_files():\r\n selected_files.append(self.get_chosen_files()[0])\r\n #\r\n # Loop through each file\/directory selected.\r\n #\r\n for filedir in selected_files:\r\n if os.path.isdir(filedir):\r\n #\r\n # It is a directory. Process as a directory.\r\n #\r\n newDir = filedir + \"-copy\"\r\n distutils.dir_util.copy_tree(filedir,newDir)\r\n else:\r\n #\r\n # It is a file. Process as a file.\r\n #\r\n dirPath, ofilenmc = os.path.split(filedir)\r\n ofilenm, ext = os.path.splitext(ofilenmc)\r\n nfilenm = os.path.join(dirPath,ofilenm + \"-copy\" + ext)\r\n distutils.file_util.copy_file(filedir,nfilenm)\r\n","new_contents":"from fman import DirectoryPaneCommand, show_alert\r\nfrom urllib.parse import urlparse\r\nimport os.path\r\nfrom shutil import copytree, copyfile\r\n\r\nclass DuplicateFileDir(DirectoryPaneCommand):\r\n def __call__(self):\r\n selected_files = self.pane.get_selected_files()\r\n if len(selected_files) >= 1 or (len(selected_files) == 0 and self.get_chosen_files()):\r\n if len(selected_files) == 0 and self.get_chosen_files():\r\n selected_files.append(self.get_chosen_files()[0])\r\n #\r\n # Loop through each file\/directory selected.\r\n #\r\n for filedir in selected_files:\r\n p = urlparse(filedir)\r\n filepath = os.path.abspath(os.path.join(p.netloc, p.path))\r\n if os.path.isdir(filepath):\r\n #\r\n # It is a directory. Process as a directory.\r\n #\r\n newDir = filepath + \"-copy\"\r\n copytree(filepath, newDir)\r\n else:\r\n if os.path.isfile(filepath):\r\n #\r\n # It is a file. Process as a file.\r\n #\r\n dirPath, ofilenmc = os.path.split(filepath)\r\n ofilenm, ext = os.path.splitext(ofilenmc)\r\n nfilenm = os.path.join(dirPath,ofilenm + \"-copy\" + ext)\r\n copyfile(filepath, nfilenm)\r\n else:\r\n show_alert('Bad file path : {0}'.format(filepath))\r\n","subject":"Make it work with last fman version (0.7) on linux","message":"Make it work with last fman version (0.7) on linux\n","lang":"Python","license":"mit","repos":"raguay\/DuplicateFileDir"} {"commit":"2f80f786be8e0d235dcb98c4fa562bfe2b9e783f","old_file":"jobs\/spiders\/visir.py","new_file":"jobs\/spiders\/visir.py","old_contents":"import dateutil.parser\nimport scrapy\n\nfrom jobs.items import JobsItem\n\n\nclass VisirSpider(scrapy.Spider):\n name = \"visir\"\n start_urls = ['https:\/\/job.visir.is\/search-results-jobs\/']\n\n def parse(self, response):\n for job in response.css('.thebox'):\n info = job.css('a')[1]\n\n item = JobsItem()\n item['spider'] = self.name\n item['url'] = url = info.css('a::attr(href)').extract_first()\n item['posted'] = dateutil.parser.parse(job.css('td::text').re(r'[\\d.]+')[0]).isoformat()\n\n request = scrapy.Request(url, callback=self.parse_specific_job)\n request.meta['item'] = item\n yield request\n\n next_page = response.urljoin(response.css('.nextBtn a::attr(href)').extract_first())\n if next_page != response.url:\n yield scrapy.Request(next_page, callback=self.parse)\n\n def parse_specific_job(self, response):\n item = response.meta['item']\n item['company'] = response.css('.company-name::text').extract_first()\n item['title'] = response.css('h2::text').extract_first()\n yield item\n","new_contents":"import dateutil.parser\nimport scrapy\n\nfrom jobs.items import JobsItem\n\n\nclass VisirSpider(scrapy.Spider):\n name = \"visir\"\n start_urls = ['https:\/\/job.visir.is\/search-results-jobs\/']\n\n def parse(self, response):\n for job in response.css('.thebox'):\n info = job.css('a')[1]\n\n item = JobsItem()\n item['spider'] = self.name\n item['url'] = url = info.css('a::attr(href)').extract_first()\n item['posted'] = dateutil.parser.parse(job.css('td::text').re(r'[\\d.]+')[0], dayfirst=False).isoformat()\n\n request = scrapy.Request(url, callback=self.parse_specific_job)\n request.meta['item'] = item\n yield request\n\n next_page = response.urljoin(response.css('.nextBtn a::attr(href)').extract_first())\n if next_page != response.url:\n yield scrapy.Request(next_page, callback=self.parse)\n\n def parse_specific_job(self, response):\n item = response.meta['item']\n item['company'] = response.css('.company-name::text').extract_first()\n item['title'] = response.css('h2::text').extract_first()\n yield item\n","subject":"Fix parsing of dates for Visir.","message":"Fix parsing of dates for Visir.\n\nSome dates are being wrongly parsed, so we need to specify some information about the order of things.\n","lang":"Python","license":"apache-2.0","repos":"multiplechoice\/workplace"} {"commit":"a24d6a25cb7ee5101e8131a9719744f79b23c11b","old_file":"examples\/quotes\/quotes.py","new_file":"examples\/quotes\/quotes.py","old_contents":"import sys\nprint(sys.version_info)\nimport random\nimport time\n\nimport networkzero as nw0\n\nquotes = [\n \"Humpty Dumpty sat on a wall\",\n \"Hickory Dickory Dock\",\n \"Baa Baa Black Sheep\",\n \"Old King Cole was a merry old sould\",\n]\n\ndef main(address_pattern=None):\n my_name = input(\"Name: \")\n my_address = nw0.advertise(my_name, address_pattern)\n print(\"Advertising %s on %s\" % (my_name, my_address))\n\n while True:\n services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]\n \n for name, address in services:\n topic, message = nw0.wait_for_notification(address, \"quote\", wait_for_s=0)\n if topic:\n print(\"%s says: %s\" % (name, message))\n quote = random.choice(quotes)\n nw0.send_notification(address, \"quote\", quote)\n \n time.sleep(0.5)\n\nif __name__ == '__main__':\n main(*sys.argv[1:])\n","new_contents":"import sys\nprint(sys.version_info)\nimport random\nimport time\n\nimport networkzero as nw0\n\nquotes = [\n \"Humpty Dumpty sat on a wall\",\n \"Hickory Dickory Dock\",\n \"Baa Baa Black Sheep\",\n \"Old King Cole was a merry old sould\",\n]\n\ndef main(address_pattern=None):\n my_name = input(\"Name: \")\n my_address = nw0.advertise(my_name, address_pattern)\n print(\"Advertising %s on %s\" % (my_name, my_address))\n\n while True:\n services = [(name, address) for (name, address) in nw0.discover_all() if name != my_name]\n\n for name, address in services:\n topic, message = nw0.wait_for_notification(address, \"quote\", wait_for_s=0)\n if topic:\n print(\"%s says: %s\" % (name, message))\n \n quote = random.choice(quotes)\n nw0.send_notification(my_address, \"quote\", quote)\n \n time.sleep(1)\n\nif __name__ == '__main__':\n main(*sys.argv[1:])\n","subject":"Send notification to the correct address","message":"Send notification to the correct address\n","lang":"Python","license":"mit","repos":"tjguk\/networkzero,tjguk\/networkzero,tjguk\/networkzero"} {"commit":"0f95070880f40456fbb6d7b7ccd6e999cc6fb95a","old_file":"dropbox_conflict_resolver.py","new_file":"dropbox_conflict_resolver.py","old_contents":"import os\nimport re\n\n'''\n This is used to revert back a Dropbox conflict. So in this case I want to keep all the files that where\n converted to conflict copies. So I just strip out the conflict string ie (some computer names's conflict copy some date) .ext\n and remove that conflict part of the string, and overate the original file by that name.\n'''\nfor root, dirs, files, in os.walk(r\"path to your drop box file with conflicts\"):\n for file in files:\n file_matcher = re.search(r\"(.+) (\\(.+'s conflicted copy [0-9][0-9][0-9][0-9]-[0-9][0-9]-[0-9][0-9]*\\))(.+)?\", file)\n if file_matcher:\n full_path = os.path.join(root, file)\n conflict_file_name = file_matcher.group(0)\n clean_file_name = file_matcher.group(1)\n conflict_string = file_matcher.group(2)\n file_ext = file_matcher.group(3)\n\n new_name_file_name = clean_file_name\n\n if file_ext:\n new_name_file_name += file_ext\n\n new_path = os.path.join(root, new_name_file_name)\n\n print(\"from: \" + full_path + \" to: \" + new_path)\n os.replace(full_path, new_path)\n\n","new_contents":"import os\nimport re\n\n'''\n This is used to revert back a Dropbox conflict. So in this case I want to keep all the files that were\n converted to conflict copies. So I just strip out the conflict string ie (some computer names's conflict copy some date) .ext\n and remove that conflict part of the string, and override the original file by that name.\n'''\nfor root, dirs, files, in os.walk(r\"path to your drop box file with conflicts\"):\n for file in files:\n file_matcher = re.search(r\"(.+) (\\(.+'s conflicted copy [0-9][0-9][0-9][0-9]-[0-9][0-9]-[0-9][0-9]*\\))(.+)?\", file)\n if file_matcher:\n full_path = os.path.join(root, file)\n conflict_file_name = file_matcher.group(0)\n clean_file_name = file_matcher.group(1)\n conflict_string = file_matcher.group(2)\n file_ext = file_matcher.group(3)\n\n new_name_file_name = clean_file_name\n\n if file_ext:\n new_name_file_name += file_ext\n\n new_path = os.path.join(root, new_name_file_name)\n\n print(\"from: \" + full_path + \" to: \" + new_path)\n os.replace(full_path, new_path)\n\n","subject":"Fix a couple of typos in the program description","message":"Fix a couple of typos in the program description\n","lang":"Python","license":"apache-2.0","repos":"alexwhb\/Dropbox-bulk-conflict-resolver"} {"commit":"be07a935d041a6c2d1c641f9beebe1bb49891682","old_file":"cooler\/cli\/__init__.py","new_file":"cooler\/cli\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import division, print_function\nimport logging\nimport sys\n\nimport click\nfrom .. import __version__, get_logger\n\nlogging.basicConfig(stream=sys.stderr)\nlogger = get_logger()\nlogger.setLevel(logging.INFO)\n\n\n# Monkey patch\nclick.core._verify_python3_env = lambda: None\n\n\nCONTEXT_SETTINGS = {\n 'help_option_names': ['-h', '--help'],\n}\n\n\n@click.version_option(version=__version__)\n@click.group(context_settings=CONTEXT_SETTINGS)\n@click.option(\n '--debug\/--no-debug', \n help=\"Verbose logging\", \n default=False)\ndef cli(debug):\n if debug:\n logger.setLevel(logging.DEBUG)\n\n\nfrom . import (\n makebins,\n digest,\n csort,\n cload,\n load,\n merge,\n copy,\n list_,\n info,\n dump,\n balance,\n aggregate,\n show,\n)\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import division, print_function\nimport logging\nimport sys\n\nimport click\nfrom .. import __version__, get_logger\n\nlogging.basicConfig(stream=sys.stderr)\nlogger = get_logger()\nlogger.setLevel(logging.INFO)\n\n\n# Monkey patch\nclick.core._verify_python3_env = lambda: None\n\n\nCONTEXT_SETTINGS = {\n 'help_option_names': ['-h', '--help'],\n}\n\n\n@click.version_option(version=__version__)\n@click.group(context_settings=CONTEXT_SETTINGS)\n@click.option(\n '--debug\/--no-debug', \n help=\"Verbose logging\", \n default=False)\n@click.option(\n '-pm', '--post-mortem', \n help=\"Post mortem debugging\", \n is_flag=True,\n default=False)\ndef cli(debug, post_mortem):\n if debug:\n logger.setLevel(logging.DEBUG)\n\n if post_mortem:\n import traceback\n try:\n import ipdb as pdb\n except ImportError:\n import pdb\n def _excepthook(exc_type, value, tb):\n traceback.print_exception(exc_type, value, tb)\n print()\n pdb.pm()\n sys.excepthook = _excepthook\n\n\nfrom . import (\n makebins,\n digest,\n csort,\n cload,\n load,\n merge,\n copy,\n list_,\n info,\n dump,\n balance,\n aggregate,\n show,\n)\n","subject":"Add postmortem debugging option to CLI","message":"Add postmortem debugging option to CLI\n","lang":"Python","license":"bsd-3-clause","repos":"mirnylab\/cooler"} {"commit":"723ae54f260284aad442f076772189cb5820d62e","old_file":"devtools\/ci\/push-docs-to-s3.py","new_file":"devtools\/ci\/push-docs-to-s3.py","old_contents":"import os\nimport pip\nimport tempfile\nimport subprocess\nimport opentis.version\n\n\nBUCKET_NAME = 'openpathsampling.org'\nif not opentis.version.release:\n PREFIX = 'latest'\nelse:\n PREFIX = opentis.version.short_version\n\nPREFIX = ''\n\nif not any(d.project_name == 's3cmd' for d in pip.get_installed_distributions()):\n raise ImportError('The s3cmd pacakge is required. try $ pip install s3cmd')\n# The secret key is available as a secure environment variable\n# on travis-ci to push the build documentation to Amazon S3.\nwith tempfile.NamedTemporaryFile('w') as f:\n f.write('''[default]\naccess_key = {AWS_ACCESS_KEY_ID}\nsecret_key = {AWS_SECRET_ACCESS_KEY}\n'''.format(**os.environ))\n f.flush()\n\n template = ('s3cmd --config {config} '\n 'sync docs\/_build\/ s3:\/\/{bucket}\/{prefix}\/')\n cmd = template.format(\n config=f.name,\n bucket=BUCKET_NAME\n )\n return_val = subprocess.call(cmd.split())\n\n # Sync index file.\n template = ('s3cmd --config {config} '\n 'sync devtools\/ci\/index.html s3:\/\/{bucket}\/')\n cmd = template.format(\n config=f.name,\n bucket=BUCKET_NAME\n )\n return_val = subprocess.call(cmd.split())\n\n","new_contents":"import os\nimport pip\nimport tempfile\nimport subprocess\nimport opentis.version\n\n\nBUCKET_NAME = 'openpathsampling.org'\nif not opentis.version.release:\n PREFIX = 'latest'\nelse:\n PREFIX = opentis.version.short_version\n\nPREFIX = ''\n\nif not any(d.project_name == 's3cmd' for d in pip.get_installed_distributions()):\n raise ImportError('The s3cmd pacakge is required. try $ pip install s3cmd')\n# The secret key is available as a secure environment variable\n# on travis-ci to push the build documentation to Amazon S3.\nwith tempfile.NamedTemporaryFile('w') as f:\n f.write('''[default]\naccess_key = {AWS_ACCESS_KEY_ID}\nsecret_key = {AWS_SECRET_ACCESS_KEY}\n'''.format(**os.environ))\n f.flush()\n\n template = ('s3cmd --config {config} '\n 'sync docs\/_build\/ s3:\/\/{bucket}\/{prefix}\/')\n cmd = template.format(\n config=f.name,\n bucket=BUCKET_NAME,\n prefix=PREFIX)\n return_val = subprocess.call(cmd.split())\n\n # Sync index file.\n template = ('s3cmd --config {config} '\n 'sync devtools\/ci\/index.html s3:\/\/{bucket}\/')\n cmd = template.format(\n config=f.name,\n bucket=BUCKET_NAME)\n return_val = subprocess.call(cmd.split())\n\n","subject":"Fix for PREFIX omission in S3 push","message":"Fix for PREFIX omission in S3 push\n","lang":"Python","license":"mit","repos":"dwhswenson\/openpathsampling,jhprinz\/openpathsampling,choderalab\/openpathsampling,openpathsampling\/openpathsampling,choderalab\/openpathsampling,dwhswenson\/openpathsampling,jhprinz\/openpathsampling,openpathsampling\/openpathsampling,openpathsampling\/openpathsampling,dwhswenson\/openpathsampling,openpathsampling\/openpathsampling,dwhswenson\/openpathsampling,choderalab\/openpathsampling,jhprinz\/openpathsampling"} {"commit":"eb391dde8a157252a98fc9bb9b617bc821f7285a","old_file":"email_from_template\/utils.py","new_file":"email_from_template\/utils.py","old_contents":"from django.utils.functional import memoize\n\nfrom . import app_settings\n\ndef get_render_method():\n return from_dotted_path(app_settings.EMAIL_RENDER_METHOD)\nget_render_method = memoize(get_render_method, {}, 0)\n\ndef get_context_processors():\n return [from_dotted_path(x) for x in app_settings.EMAIL_CONTEXT_PROCESSORS]\nget_context_processors = memoize(get_context_processors, {}, 0)\n\ndef from_dotted_path(fullpath):\n \"\"\"\n Returns the specified attribute of a module, specified by a string.\n\n ``from_dotted_path('a.b.c.d')`` is roughly equivalent to::\n\n from a.b.c import d\n\n except that ``d`` is returned and not entered into the current namespace.\n \"\"\"\n\n module, attr = fullpath.rsplit('.', 1)\n\n return getattr(__import__(module, {}, {}, (attr,)), attr)\n","new_contents":"from django.utils.lru_cache import lru_cache\n\nfrom . import app_settings\n\n@lru_cache\ndef get_render_method():\n return from_dotted_path(app_settings.EMAIL_RENDER_METHOD)\n\n@lru_cache\ndef get_context_processors():\n return [from_dotted_path(x) for x in app_settings.EMAIL_CONTEXT_PROCESSORS]\n\ndef from_dotted_path(fullpath):\n \"\"\"\n Returns the specified attribute of a module, specified by a string.\n\n ``from_dotted_path('a.b.c.d')`` is roughly equivalent to::\n\n from a.b.c import d\n\n except that ``d`` is returned and not entered into the current namespace.\n \"\"\"\n\n module, attr = fullpath.rsplit('.', 1)\n\n return getattr(__import__(module, {}, {}, (attr,)), attr)\n","subject":"Use @lru_cache now that memoize is gone.","message":"Use @lru_cache now that memoize is gone.\n","lang":"Python","license":"bsd-3-clause","repos":"lamby\/django-email-from-template"} {"commit":"75af7171d0245b528018c8e0d0d581916a9dc67d","old_file":"examples\/profilealignment.py","new_file":"examples\/profilealignment.py","old_contents":"# Create sequences to be aligned.\nfrom alignment.sequence import Sequence\na = Sequence(\"what a beautiful day\".split())\nb = Sequence(\"what a disappointingly bad day\".split())\nprint \"Sequence A:\", a\nprint \"Sequence B:\", b\nprint\n\n# Create a vocabulary and encode the sequences.\nfrom alignment.vocabulary import Vocabulary\nv = Vocabulary()\naEncoded = v.encodeSequence(a)\nbEncoded = v.encodeSequence(b)\nprint \"Encoded A:\", aEncoded\nprint \"Encoded B:\", bEncoded\nprint\n\n# Create a scoring and align the sequences using global aligner.\nfrom alignment.sequencealigner import SimpleScoring, GlobalSequenceAligner\nscoring = SimpleScoring(2, -1)\naligner = GlobalSequenceAligner(scoring, -2)\nscore, alignments = aligner.align(aEncoded, bEncoded, backtrace=True)\n\n# Create sequence profiles out of alignments.\nfrom alignment.profile import Profile\nprofiles = [Profile.fromSequenceAlignment(a) for a in alignments]\nfor encoded in profiles:\n\tprofile = v.decodeProfile(encoded)\n\tprint profile\nprint\n\n# Create a soft scoring and align the first profile against sequence A.\nfrom alignment.profilealigner import SoftScoring, GlobalProfileAligner\nscoring = SoftScoring(scoring)\naligner = GlobalProfileAligner(scoring, -2)\nscore, alignments = aligner.align(profiles[0], Profile.fromSequence(aEncoded), backtrace=True)\nfor encoded in alignments:\n\talignment = v.decodeProfileAlignment(encoded)\n\tprint alignment\n\n","new_contents":"from alignment.sequence import Sequence\nfrom alignment.vocabulary import Vocabulary\nfrom alignment.sequencealigner import SimpleScoring, GlobalSequenceAligner\nfrom alignment.profile import Profile\nfrom alignment.profilealigner import SoftScoring, GlobalProfileAligner\n\n\n# Create sequences to be aligned.\na = Sequence('what a beautiful day'.split())\nb = Sequence('what a disappointingly bad day'.split())\nprint 'Sequence A:', a\nprint 'Sequence B:', b\nprint\n\n# Create a vocabulary and encode the sequences.\nv = Vocabulary()\naEncoded = v.encodeSequence(a)\nbEncoded = v.encodeSequence(b)\nprint 'Encoded A:', aEncoded\nprint 'Encoded B:', bEncoded\nprint\n\n# Create a scoring and align the sequences using global aligner.\nscoring = SimpleScoring(2, -1)\naligner = GlobalSequenceAligner(scoring, -2)\nscore, alignments = aligner.align(aEncoded, bEncoded, backtrace=True)\n\n# Create sequence profiles out of alignments.\nprofiles = [Profile.fromSequenceAlignment(a) for a in alignments]\nfor encoded in profiles:\n profile = v.decodeProfile(encoded)\n print profile\nprint\n\n# Create a soft scoring and align the first profile against sequence A.\nscoring = SoftScoring(scoring)\naligner = GlobalProfileAligner(scoring, -2)\nscore, alignments = aligner.align(profiles[0], Profile.fromSequence(aEncoded),\n backtrace=True)\nfor encoded in alignments:\n alignment = v.decodeProfileAlignment(encoded)\n print alignment\n","subject":"Update the profile alignment example.","message":"Update the profile alignment example.\n","lang":"Python","license":"bsd-3-clause","repos":"eseraygun\/python-entities,eseraygun\/python-alignment"} {"commit":"4bef46ef98591d47d653eeb4f74bf00a8a1d5d69","old_file":"correios\/utils.py","new_file":"correios\/utils.py","old_contents":"from itertools import chain\nfrom typing import Sized, Iterable, Container, Set\n\n\nclass RangeSet(Sized, Iterable, Container):\n def __init__(self, *ranges):\n self.ranges = []\n\n for r in ranges:\n if isinstance(r, range):\n r = [r]\n elif isinstance(r, RangeSet):\n r = list(r.ranges)\n elif isinstance(r, Iterable) and not isinstance(r, Set):\n r = [range(*r)]\n else:\n msg = \"RangeSet argument must be a range, RangeSet or an Iterable, not {}\"\n raise ValueError(msg.format(type(r)))\n\n self.ranges.extend(r)\n\n def __iter__(self):\n return chain.from_iterable(r for r in self.ranges)\n\n def __contains__(self, elem):\n return any(elem in r for r in self.ranges)\n\n def __len__(self):\n return sum(len(r) for r in self.ranges)\n","new_contents":"from itertools import chain\nfrom typing import Container, Iterable, Sized\n\n\nclass RangeSet(Sized, Iterable, Container):\n def __init__(self, *ranges):\n self.ranges = []\n\n for r in ranges:\n if isinstance(r, range):\n self.ranges.append(r)\n continue\n\n try:\n element = list(r.ranges)\n except AttributeError:\n element = None\n\n try:\n element = element or [range(*r)]\n except:\n msg = \"RangeSet argument must be a range, RangeSet or an Iterable, not {}\"\n raise ValueError(msg.format(type(r)))\n\n self.ranges.extend(element)\n\n def __iter__(self):\n return chain.from_iterable(r for r in self.ranges)\n\n def __contains__(self, elem):\n return any(elem in r for r in self.ranges)\n\n def __len__(self):\n return sum(len(r) for r in self.ranges)\n","subject":"Use duck typing when creating a RangeSet","message":"Use duck typing when creating a RangeSet\n","lang":"Python","license":"apache-2.0","repos":"osantana\/correios,solidarium\/correios,olist\/correios"} {"commit":"5c1970c8c44d9bc174febb96d8cb4ffd59cf0169","old_file":"targetcli\/__init__.py","new_file":"targetcli\/__init__.py","old_contents":"'''\nThis file is part of targetcli.\nCopyright (c) 2011 by RisingTide Systems LLC\n\nThis program is free software: you can redistribute it and\/or modify\nit under the terms of the GNU Affero General Public License as\npublished by the Free Software Foundation, version 3 (AGPLv3).\n\nThis program is distributed in the hope that it will be useful,\nbut WITHOUT ANY WARRANTY; without even the implied warranty of\nMERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\nGNU Affero General Public License for more details.\n\nYou should have received a copy of the GNU Affero General Public License\nalong with this program. If not, see <http:\/\/www.gnu.org\/licenses\/>.\n'''\n\nfrom ui_root import UIRoot\n\n__version__ = \"2.0.22\"\n__author__ = \"Jerome Martin <jxm@risingtidesystems.com>\"\n__url__ = \"http:\/\/www.risingtidesystems.com\"\n__description__ = \"An administration shell for RTS storage targets.\"\n__license__ = __doc__\n","new_contents":"'''\nThis file is part of targetcli.\nCopyright (c) 2011 by RisingTide Systems LLC\n\nThis program is free software: you can redistribute it and\/or modify\nit under the terms of the GNU Affero General Public License as\npublished by the Free Software Foundation, version 3 (AGPLv3).\n\nThis program is distributed in the hope that it will be useful,\nbut WITHOUT ANY WARRANTY; without even the implied warranty of\nMERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the\nGNU Affero General Public License for more details.\n\nYou should have received a copy of the GNU Affero General Public License\nalong with this program. If not, see <http:\/\/www.gnu.org\/licenses\/>.\n'''\n\nfrom ui_root import UIRoot\n\n__version__ = \"2.1.22\"\n__author__ = \"Jerome Martin <jxm@risingtidesystems.com>\"\n__url__ = \"http:\/\/www.risingtidesystems.com\"\n__description__ = \"An administration shell for RTS storage targets.\"\n__license__ = __doc__\n","subject":"Update version for proper pkg versioning","message":"Update version for proper pkg versioning\n\nUnfortunately I made a mistake, and did not handle version with 'rc' in it\nas specified here:\n\nhttp:\/\/fedoraproject.org\/wiki\/Packaging:NamingGuidelines#Pre-Release_packages\n\nI'm incrementing the version to 2.1.<fb-ver> to get version sequencing\nright. It's a little uncool for me to be mixing my developer and packager\nroles here, but I think I can rationalize a version bump, given the\ntag and WWN improvements recently.\n\nSigned-off-by: Andy Grover <b7d524d2f5cc5aebadb6b92b08d3ab26911cde33@redhat.com>\n","lang":"Python","license":"apache-2.0","repos":"cvubrugier\/targetcli-fb,cloud4life\/targetcli-fb,agrover\/targetcli-fb"} {"commit":"55a4680bb07896f0bab06d836ade056d115f004f","old_file":"dsub\/_dsub_version.py","new_file":"dsub\/_dsub_version.py","old_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Single source of truth for dsub's version.\n\nThis must remain small and dependency-free so that any dsub module may\nimport it without creating circular dependencies. Note that this module\nis parsed as a text file by setup.py and changes to the format of this\nfile could break setup.py.\n\nThe version should follow formatting requirements specified in PEP-440.\n - https:\/\/www.python.org\/dev\/peps\/pep-0440\n\nA typical release sequence will be versioned as:\n 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...\n\"\"\"\n\nDSUB_VERSION = '0.1.9'\n","new_contents":"# Copyright 2017 Google Inc. All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Single source of truth for dsub's version.\n\nThis must remain small and dependency-free so that any dsub module may\nimport it without creating circular dependencies. Note that this module\nis parsed as a text file by setup.py and changes to the format of this\nfile could break setup.py.\n\nThe version should follow formatting requirements specified in PEP-440.\n - https:\/\/www.python.org\/dev\/peps\/pep-0440\n\nA typical release sequence will be versioned as:\n 0.1.3.dev0 -> 0.1.3 -> 0.1.4.dev0 -> ...\n\"\"\"\n\nDSUB_VERSION = '0.1.10.dev0'\n","subject":"Update version number to 0.1.10.dev0.","message":"Update version number to 0.1.10.dev0.\n\nPiperOrigin-RevId: 202663603\n","lang":"Python","license":"apache-2.0","repos":"DataBiosphere\/dsub,DataBiosphere\/dsub"} {"commit":"3fbbdec51cfd93217705adcae37b1bf22d5661fa","old_file":"backend\/playlist\/serializers.py","new_file":"backend\/playlist\/serializers.py","old_contents":"from rest_framework import serializers\nfrom .models import Cd, Cdtrack, Show, Playlist, PlaylistEntry\n\n\nclass TrackSerializer(serializers.ModelSerializer):\n album = serializers.StringRelatedField(\n read_only=True\n )\n\n class Meta:\n model = Cdtrack\n fields = ('trackid', 'url', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'album')\n\n\nclass ReleaseSerializer(serializers.HyperlinkedModelSerializer):\n tracks = serializers.PrimaryKeyRelatedField(\n many=True,\n read_only=True\n )\n\n class Meta:\n model = Cd\n fields = ('id', 'url', 'arrivaldate', 'artist', 'title', 'year', 'local', 'compilation', 'female', 'tracks')\n\nclass ShowSerializer(serializers.ModelSerializer):\n class Meta:\n model = Show\n fields = ('id', 'name', 'startTime', 'endTime', 'defaultHost') \n\nclass PlaylistEntrySerializer(serializers.ModelSerializer):\n playlist = serializers.PrimaryKeyRelatedField(\n queryset = Playlist.objects.all()\n )\n\n class Meta:\n model = PlaylistEntry\n fields = ('id', 'artist','album','title','duration','local','australian','female','newRelease','playlist' )\n pass\n\nclass PlaylistSerializer(serializers.ModelSerializer):\n entries =PlaylistEntrySerializer( \n many=True,\n read_only=True\n )\n\n class Meta:\n model = Playlist\n fields = ('id', 'show', 'host', 'date', 'notes', 'entries')\n","new_contents":"from rest_framework import serializers\nfrom .models import Cd, Cdtrack, Show, Playlist, PlaylistEntry\n\n\nclass TrackSerializer(serializers.ModelSerializer):\n album = serializers.StringRelatedField(\n read_only=True\n )\n\n class Meta:\n model = Cdtrack\n fields = ('trackid', 'url', 'tracknum', 'trackartist', 'tracktitle', 'tracklength', 'album')\n\n\nclass ReleaseSerializer(serializers.HyperlinkedModelSerializer):\n tracks = serializers.PrimaryKeyRelatedField(\n many=True,\n read_only=True\n )\n\n class Meta:\n model = Cd\n fields = ('id', 'url', 'arrivaldate', 'artist', 'title', 'year', 'local', 'compilation', 'female', 'tracks')\n\nclass ShowSerializer(serializers.ModelSerializer):\n class Meta:\n model = Show\n fields = ('id', 'name', 'startTime', 'endTime', 'defaultHost') \n\nclass PlaylistEntrySerializer(serializers.ModelSerializer):\n playlist = serializers.PrimaryKeyRelatedField(\n queryset = Playlist.objects.all()\n )\n\n class Meta:\n model = PlaylistEntry\n fields = ('id', 'artist','album','title','duration','local','australian','female','newRelease','playlist' )\n pass\n\nclass PlaylistSerializer(serializers.ModelSerializer):\n entries =PlaylistEntrySerializer( \n many=True,\n read_only=True\n )\n\n class Meta:\n model = Playlist\n fields = ('id', 'show','showname', 'host', 'date', 'notes', 'entries')\n","subject":"Add showname to playlist API view.","message":"Add showname to playlist API view.\n\n* Even though it's obsolete now, we need it for old shows.\n","lang":"Python","license":"mit","repos":"ThreeDRadio\/playlists,ThreeDRadio\/playlists,ThreeDRadio\/playlists"} {"commit":"56aa0448fb3cd1df1a0fd43abc9a0e37e8ddf55b","old_file":"trans_sync\/management\/commands\/save_trans.py","new_file":"trans_sync\/management\/commands\/save_trans.py","old_contents":"# coding: utf-8\nfrom __future__ import unicode_literals\n\nfrom optparse import make_option\nfrom django.core.management.base import NoArgsCommand\n\n\nclass Command(NoArgsCommand):\n\n option_list = NoArgsCommand.option_list + (\n make_option(\n '--dry-run',\n action='store_true',\n dest='dry_run',\n default=False,\n help='Do not actually send signals (and all connected stuff).'\n ),\n )\n\n def handle(self, *args, **options):\n if not options['dry_run']:\n pass","new_contents":"# coding: utf-8\nfrom __future__ import unicode_literals\nimport os\nfrom os.path import join, isdir\nfrom optparse import make_option\nfrom django.core.management.base import NoArgsCommand\nfrom django.conf import settings\nfrom modeltranslation.translator import translator\n\nfrom babel.messages.catalog import Catalog\nfrom babel.messages.pofile import write_po\n\n\nclass Command(NoArgsCommand):\n\n option_list = NoArgsCommand.option_list + (\n make_option(\n '--dry-run',\n action='store_true',\n dest='dry_run',\n default=False,\n help='Do not actually save files.'\n ),\n )\n\n def handle(self, *args, **options):\n if not options['dry_run']:\n pass\n\n locale_path = settings.LOCALE_MODEL_TRANS\n if not isdir(locale_path):\n os.mkdir(locale_path)\n\n for lang in [l[0] for l in list(settings.LANGUAGES)]:\n\n catalog = Catalog(locale=lang)\n\n for model in translator.get_registered_models():\n opts = translator.get_options_for_model(model)\n\n for field in opts.get_field_names():\n tr_field = \"%s_%s\" % (field, lang)\n for item in model.objects.all():\n msgid = \"%s.%s.%s\" % (item._meta, item.pk, field)\n msgstr = \"%s\" % getattr(item, tr_field)\n catalog.add(id=msgid, string=msgstr)\n\n # write catalog to file\n lang_path = os.path.join(locale_path, lang)\n if not isdir(lang_path):\n os.mkdir(lang_path)\n f = open(join(lang_path, \"LC_MESSAGES\", \"modeltranslation.po\"), \"w\")\n write_po(f, catalog)\n f.close()","subject":"Save trans to .po files","message":"Save trans to .po files\n","lang":"Python","license":"mit","repos":"djentlemen\/django-modeltranslation-sync"} {"commit":"e2495040277fafdac4c0e060517cf667baa27c02","old_file":"chinup\/__init__.py","new_file":"chinup\/__init__.py","old_contents":"try:\n from .allauth import *\nexcept ImportError:\n from .chinup import *\n\nfrom .exceptions import *\n\n\n__version__ = '0.1'\n","new_contents":"from __future__ import absolute_import, unicode_literals\n\ntry:\n from .allauth import *\nexcept ImportError:\n from .chinup import *\n\nfrom .exceptions import *\n\n\n__version__ = '0.1'\n\n\n# Configure logging to avoid warning.\n# https:\/\/docs.python.org\/2\/howto\/logging.html#configuring-logging-for-a-library\nimport logging\nif hasattr(logging, 'NullHandler'):\n logging.getLogger('chinup').addHandler(logging.NullHandler())\n","subject":"Configure package-level logging to avoid warning.","message":"Configure package-level logging to avoid warning.\n","lang":"Python","license":"mit","repos":"pagepart\/chinup"} {"commit":"fc36b9bc2970c611a4fb5063463f27cfd96df21d","old_file":"moksha\/hub\/messaging.py","new_file":"moksha\/hub\/messaging.py","old_contents":"# This file is part of Moksha.\n# Copyright (C) 2008-2010 Red Hat, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# Authors: Luke Macken <lmacken@redhat.com>\n\nclass MessagingHub(object):\n \"\"\"\n A generic messaging hub.\n\n This class represents the base functionality of the protocol-level hubs.\n \"\"\"\n\n def send_message(self, topic, message):\n raise NotImplementedError\n\n def subscribe(self, topic):\n raise NotImplementedError\n","new_contents":"# This file is part of Moksha.\n# Copyright (C) 2008-2010 Red Hat, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# Authors: Luke Macken <lmacken@redhat.com>\n\nclass MessagingHub(object):\n \"\"\"\n A generic messaging hub.\n\n This class represents the base functionality of the protocol-level hubs.\n \"\"\"\n\n def send_message(self, topic, message):\n raise NotImplementedError\n\n def subscribe(self, topic, callback):\n raise NotImplementedError\n","subject":"Update our MessagingHub.subscribe method arguments","message":"Update our MessagingHub.subscribe method arguments\n","lang":"Python","license":"apache-2.0","repos":"ralphbean\/moksha,pombredanne\/moksha,mokshaproject\/moksha,lmacken\/moksha,mokshaproject\/moksha,lmacken\/moksha,ralphbean\/moksha,pombredanne\/moksha,pombredanne\/moksha,mokshaproject\/moksha,mokshaproject\/moksha,ralphbean\/moksha,pombredanne\/moksha,lmacken\/moksha"} {"commit":"a7fcc89755e01bf3dbe7090e2bf7f1211ce9af84","old_file":"test\/test_property.py","new_file":"test\/test_property.py","old_contents":"import unittest\nfrom odml import Property, Section, Document\n\n\nclass TestProperty(unittest.TestCase):\n\n def setUp(self):\n pass\n\n def test_value(self):\n p = Property(\"property\", 100)\n assert(p.value[0] == 100)\n\n def test_name(self):\n pass\n\n def test_parent(self):\n pass\n\n def test_dtype(self):\n pass\n\n def test_path(self):\n pass\n\nif __name__ == \"__main__\":\n print(\"TestProperty\")\n tp = TestProperty()\n tp.test_value()\n","new_contents":"import unittest\nfrom odml import Property, Section, Document\nfrom odml.doc import BaseDocument\nfrom odml.section import BaseSection\n\n\nclass TestProperty(unittest.TestCase):\n\n def setUp(self):\n pass\n\n def test_value(self):\n p = Property(\"property\", 100)\n self.assertEqual(p.value[0], 100)\n\n def test_name(self):\n pass\n\n def test_parent(self):\n p = Property(\"property_section\", parent=Section(\"S\"))\n self.assertIsInstance(p.parent, BaseSection)\n self.assertEqual(len(p.parent._props), 1)\n with self.assertRaises(ValueError):\n Property(\"property_prop\", parent=Property(\"P\"))\n with self.assertRaises(ValueError):\n Property(\"property_doc\", parent=Document())\n\n def test_dtype(self):\n pass\n\n def test_path(self):\n pass\n\n\nclass TestSection(unittest.TestCase):\n def setUp(self):\n pass\n\n def test_value(self):\n pass\n\n def test_name(self):\n pass\n\n def test_parent(self):\n s = Section(\"Section\")\n self.assertIsNone(s.parent)\n\n s = Section(\"section_doc\", parent=Document())\n self.assertIsInstance(s.parent, BaseDocument)\n self.assertEqual(len(s.parent._sections), 1)\n\n s = Section(\"section_sec\", parent=Section(\"S\"))\n self.assertIsInstance(s.parent, BaseSection)\n self.assertEqual(len(s.parent._sections), 1)\n\n with self.assertRaises(ValueError):\n Section(\"section_property\", parent=Property(\"P\"))\n\n def test_dtype(self):\n pass\n\n def test_path(self):\n pass\n\nif __name__ == \"__main__\":\n print(\"TestProperty\")\n tp = TestProperty()\n tp.test_value()\n tp.test_parent()\n\n print(\"TestSection\")\n ts = TestSection()\n ts.test_parent()\n\n\n","subject":"Add tests to cover update parent functionality.","message":"Add tests to cover update parent functionality.\n","lang":"Python","license":"bsd-3-clause","repos":"lzehl\/python-odml"} {"commit":"bee93012144e033b02c05a1e586620dfa7f4c883","old_file":"words\/models.py","new_file":"words\/models.py","old_contents":"from django.db import models\n\n\nclass Word(models.Model):\n word = models.CharField(max_length=255)\n date_retired = models.DateTimeField(null=True, blank=True)\n date_active = models.DateTimeField(null=True, blank=True)\n views = models.IntegerField(default=0)\n\n @property\n def is_active(self):\n if self.date_retired:\n return False\n return bool(self.date_active)\n","new_contents":"from django.db import models\n\n\nclass Word(models.Model):\n word = models.CharField(max_length=255)\n date_retired = models.DateTimeField(null=True, blank=True)\n date_active = models.DateTimeField(null=True, blank=True)\n views = models.IntegerField(default=0)\n\n @property\n def is_active(self):\n if self.date_retired:\n return False\n return bool(self.date_active)\n\n def __unicode__(self):\n return self.word\n","subject":"Make the word display nice","message":"Make the word display nice\n","lang":"Python","license":"bsd-2-clause","repos":"kylegibson\/how_to_teach_your_baby_tracker"} {"commit":"6765cefc1a5a928b3cff16c0f1014096f82c3d3b","old_file":"test\/test_services.py","new_file":"test\/test_services.py","old_contents":"import pytest\n\n@pytest.mark.parametrize(\"name, enabled, running\", [\n (\"cron\", \"enabled\", \"running\"),\n (\"docker\", \"enabled\", \"running\"),\n (\"firewalld\", \"enabled\", \"running\"),\n (\"haveged\", \"enabled\", \"running\"),\n (\"ssh\", \"enabled\", \"running\"),\n])\n\ndef test_services(Service, name, enabled, running):\n is_enabled = Service(name).is_enabled\n print(is_enabled)\n if enabled == \"enabled\":\n assert is_enabled\n else:\n assert not is_enabled\n\n is_running = Service(name).is_running\n print(is_running)\n if running == \"running\":\n assert is_running\n else:\n assert not is_running\n\n","new_contents":"import pytest\n\n@pytest.mark.parametrize(\"name, enabled, running\", [\n (\"cron\", \"enabled\", \"running\"),\n (\"docker\", \"enabled\", \"running\"),\n (\"firewalld\", \"enabled\", \"running\"),\n (\"haveged\", \"enabled\", \"running\"),\n (\"ssh\", \"enabled\", \"running\"),\n])\n\ndef test_services(host, name, enabled, running):\n\n svc = host.service(name)\n\n is_enabled = svc.is_enabled\n print(is_enabled)\n if enabled == \"enabled\":\n assert is_enabled\n else:\n assert not is_enabled\n\n is_running = svc.is_running\n print(is_running)\n if running == \"running\":\n assert is_running\n else:\n assert not is_running\n","subject":"Change test function as existing method deprecated","message":"Change test function as existing method deprecated\n","lang":"Python","license":"mit","repos":"wicksy\/laptop-build,wicksy\/laptop-build,wicksy\/laptop-build,wicksy\/laptop-build"} {"commit":"eea647cf05d7143d800f834dd77aeafc32522100","old_file":"groundstation\/settings.py","new_file":"groundstation\/settings.py","old_contents":"PORT=1248\nBEACON_TIMEOUT=5\nDEFAULT_BUFSIZE=8192\n","new_contents":"PORT=1248\nBEACON_TIMEOUT=5\nDEFAULT_BUFSIZE=8192\nDEFAULT_CACHE_LIFETIME=900\n","subject":"Add config key for default cache lifetime","message":"Add config key for default cache lifetime\n","lang":"Python","license":"mit","repos":"richo\/groundstation,richo\/groundstation,richo\/groundstation,richo\/groundstation,richo\/groundstation"} {"commit":"81f3e4f10243cb31b600666a19112acee7c13f55","old_file":"signac\/db\/__init__.py","new_file":"signac\/db\/__init__.py","old_contents":"import warnings\ntry:\n import pymongo # noqa\nexcept ImportError:\n warnings.warn(\"Failed to import pymongo. \"\n \"get_database will not be available.\", ImportWarning)\n\n def get_database(*args, **kwargs):\n \"\"\"Get a database handle.\n\n This function is only available if pymongo is installed.\"\"\"\n raise ImportError(\n \"You need to install pymongo to use `get_database()`.\")\nelse:\n from .database import get_database\n\n\n__all__ = ['get_database']\n","new_contents":"import logging\nimport warnings\ntry:\n import pymongo # noqa\nexcept ImportError:\n warnings.warn(\"Failed to import pymongo. \"\n \"get_database will not be available.\", ImportWarning)\n\n def get_database(*args, **kwargs):\n \"\"\"Get a database handle.\n\n This function is only available if pymongo is installed.\"\"\"\n raise ImportError(\n \"You need to install pymongo to use `get_database()`.\")\nelse:\n if pymongo.version_tuple[0] < 3:\n logging.getLogger(__name__).warn(\n \"Your pymongo installation (version {}) is no longer \"\n \"supported by signac. Consider updating.\".format(pymongo.version))\n from .database import get_database\n\n\n__all__ = ['get_database']\n","subject":"Add warning about outdated pymongo versions.","message":"Add warning about outdated pymongo versions.\n\nsignac currently only supports pymongo versions 3.x.\n","lang":"Python","license":"bsd-3-clause","repos":"csadorf\/signac,csadorf\/signac"} {"commit":"ee4f8264d942d7af5f5b71ff6cd162f3ae1fe515","old_file":"django_hash_filter\/templatetags\/hash_filter.py","new_file":"django_hash_filter\/templatetags\/hash_filter.py","old_contents":"from django import template\nfrom django.template.defaultfilters import stringfilter\nfrom django.template.base import TemplateSyntaxError\nimport hashlib\nfrom django_hash_filter.templatetags import get_available_hashes\n\nregister = template.Library()\n\n@register.filter\n@stringfilter\ndef hash(value, arg):\n \"\"\"\n Returns a hex-digest of the passed in value for the hash algorithm given.\n \"\"\"\n arg = str(arg).lower()\n if not arg in get_available_hashes():\n raise TemplateSyntaxError(\"The %s hash algorithm does not exist.\" % arg)\n try:\n f = getattr(hashlib, arg)\n hashed = f(value).hexdigest()\n except Exception:\n raise ValueError(\"The %s hash algorithm cannot produce a hex digest. Ensure that OpenSSL is properly installed.\" % arg)\n return hashed","new_contents":"import hashlib\nimport sys\n\nfrom django import template\nfrom django.template.defaultfilters import stringfilter\nfrom django.template.base import TemplateSyntaxError\nfrom django_hash_filter.templatetags import get_available_hashes\n\nregister = template.Library()\n\n@register.filter\n@stringfilter\ndef hash(value, arg):\n \"\"\"\n Returns a hex-digest of the passed in value for the hash algorithm given.\n \"\"\"\n arg = str(arg).lower()\n if sys.version_info >= (3,0):\n value = value.encode(\"utf-8\")\n if not arg in get_available_hashes():\n raise TemplateSyntaxError(\"The %s hash algorithm does not exist.\" % arg)\n try:\n f = getattr(hashlib, arg)\n hashed = f(value).hexdigest()\n except Exception:\n raise ValueError(\"The %s hash algorithm cannot produce a hex digest. Ensure that OpenSSL is properly installed.\" % arg)\n return hashed\n","subject":"Convert unicode string to byte array on Python 3","message":"Convert unicode string to byte array on Python 3\n","lang":"Python","license":"mit","repos":"andrewjsledge\/django-hash-filter"} {"commit":"df216bdc25ef29da821f577a517ccdca61448cf4","old_file":"django_lightweight_queue\/middleware\/logging.py","new_file":"django_lightweight_queue\/middleware\/logging.py","old_contents":"from __future__ import absolute_import\n\nimport logging\nimport traceback\n\nlog = logging.getLogger(__name__)\n\nclass LoggingMiddleware(object):\n def process_job(self, job):\n log.info(\"Running job %s\", job)\n\n def process_result(self, job, result, duration):\n log.info(\"Finished job %s => %r (Time taken: %.2fs)\",\n job,\n result,\n duration,\n )\n\n def process_exception(self, job, duration, *exc_info):\n log.error(\"Exception when processing %r (duration: %.2fs): %s\",\n job,\n duration,\n ''.join(traceback.format_exception(*exc_info)),\n )\n","new_contents":"from __future__ import absolute_import\n\nimport logging\nimport traceback\n\nlog = logging.getLogger(__name__)\n\nclass LoggingMiddleware(object):\n def process_job(self, job):\n log.info(\"Running job %s\", job)\n\n def process_result(self, job, result, duration):\n log.info(\"Finished job => %r (Time taken: %.2fs)\",\n result,\n duration,\n )\n\n def process_exception(self, job, duration, *exc_info):\n log.error(\"Exception when processing job (duration: %.2fs): %s\",\n duration,\n ''.join(traceback.format_exception(*exc_info)),\n )\n","subject":"Save over 50% of logfile 'bloat' by not repeating all args on success\/failure","message":"Save over 50% of logfile 'bloat' by not repeating all args on success\/failure\n\nThe data will be right above it just before we run the job.\n","lang":"Python","license":"bsd-3-clause","repos":"prophile\/django-lightweight-queue,prophile\/django-lightweight-queue,thread\/django-lightweight-queue,lamby\/django-lightweight-queue,thread\/django-lightweight-queue"} {"commit":"802b9c2df754b3acf78e9e1facc1802a901e97a2","old_file":"furry\/furry.py","new_file":"furry\/furry.py","old_contents":"import discord\r\nfrom discord.ext import commands\r\n\r\nclass Furry:\r\n \"\"\"A cog that adds weird furry commands or something\"\"\"\r\n\r\n def __init__(self, bot):\r\n self.bot = bot\r\n \r\n @commands.command()\r\n async def owo(self):\r\n \"\"\"OwO what's this?\"\"\"\r\n\r\n await self.bot.say(\"*Notices \" + user.mention + \"'s bulge* OwO what's this?\")\r\n\r\ndef setup(bot):\r\n bot.add_cog(Furry(bot))\r\n","new_contents":"import discord\r\nfrom discord.ext import commands\r\n\r\nclass Furry:\r\n \"\"\"A cog that adds weird furry commands or something\"\"\"\r\n\r\n def __init__(self, bot):\r\n self.bot = bot\r\n \r\n @commands.command()\r\n async def owo(self, user : discord.Member):\r\n \"\"\"OwO what's this?\"\"\"\r\n\r\n await self.bot.say(\"*Notices \" + user.mention + \"'s bulge* OwO what's this?\")\r\n\r\ndef setup(bot):\r\n bot.add_cog(Furry(bot))\r\n","subject":"Fix the command and make it actually work","message":"Fix the command and make it actually work\n\nPass discord.Member as user\n","lang":"Python","license":"apache-2.0","repos":"KazroFox\/Kaz-Cogs"} {"commit":"6a508d01fa3fa0d4084406fcb2b5e41d1b614b7c","old_file":"datalogger\/__main__.py","new_file":"datalogger\/__main__.py","old_contents":"import sys\nfrom PyQt5.QtWidgets import QApplication\n\nfrom datalogger.api.workspace import Workspace\nfrom datalogger.analysis_window import AnalysisWindow\nfrom datalogger import __version__\n\ndef run_datalogger_full():\n print(\"CUED DataLogger {}\".format(__version__))\n\n app = 0\n app = QApplication(sys.argv)\n\n # Create the window\n w = AnalysisWindow()\n\n w.CurrentWorkspace = Workspace()\n #w.CurrentWorkspace.path = \"\/\/cued-fs\/users\/general\/tab53\/ts-home\/Documents\/urop\/Logger 2017\/cued_datalogger\/\"\n # Load the workspace\n #CurrentWorkspace.load(\"\/\/cued-fs\/users\/general\/tab53\/ts-home\/Documents\/urop\/Logger 2017\/cued_datalogger\/tests\/test_workspace.wsp\")\n\n w.addon_widget.discover_addons(w.CurrentWorkspace.path + \"addons\/\")\n\n # Run the program\n w.show()\n sys.exit(app.exec_())\n \nif __name__ == '__main__':\n run_datalogger_full()\n ","new_contents":"import sys\nfrom PyQt5.QtWidgets import QApplication\n\nfrom datalogger.api.workspace import Workspace\nfrom datalogger.analysis_window import AnalysisWindow\nfrom datalogger import __version__\n\ndef run_datalogger_full():\n print(\"CUED DataLogger {}\".format(__version__))\n\n app = 0\n app = QApplication(sys.argv)\n\n CurrentWorkspace = Workspace()\n\n # Create the window\n w = AnalysisWindow()\n\n w.CurrentWorkspace = CurrentWorkspace\n \n w.addon_widget.discover_addons(w.CurrentWorkspace.path + \"addons\/\")\n\n # Run the program\n w.show()\n sys.exit(app.exec_())\n \nif __name__ == '__main__':\n run_datalogger_full()\n ","subject":"Move workspace before window creation so config set for window","message":"Move workspace before window creation so config set for window\n\n","lang":"Python","license":"bsd-3-clause","repos":"torebutlin\/cued_datalogger"} {"commit":"15f0a2e67fe942760707694370cc652f17e1c6b3","old_file":"demo\/tests\/conftest.py","new_file":"demo\/tests\/conftest.py","old_contents":"\"\"\"Unit tests configuration file.\"\"\"\n\n\ndef pytest_configure(config):\n \"\"\"Disable verbose output when running tests.\"\"\"\n terminal = config.pluginmanager.getplugin('terminal')\n base = terminal.TerminalReporter\n\n class QuietReporter(base):\n \"\"\"A py.test reporting that only shows dots when running tests.\"\"\"\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.verbosity = 0\n self.showlongtestinfo = False\n self.showfspath = False\n\n terminal.TerminalReporter = QuietReporter\n","new_contents":"\"\"\"Unit tests configuration file.\"\"\"\n\nimport logging\n\n\ndef pytest_configure(config):\n \"\"\"Disable verbose output when running tests.\"\"\"\n logging.basicConfig(level=logging.DEBUG)\n\n terminal = config.pluginmanager.getplugin('terminal')\n base = terminal.TerminalReporter\n\n class QuietReporter(base):\n \"\"\"A py.test reporting that only shows dots when running tests.\"\"\"\n\n def __init__(self, *args, **kwargs):\n super().__init__(*args, **kwargs)\n self.verbosity = 0\n self.showlongtestinfo = False\n self.showfspath = False\n\n terminal.TerminalReporter = QuietReporter\n","subject":"Deploy Travis CI build 834 to GitHub","message":"Deploy Travis CI build 834 to GitHub\n","lang":"Python","license":"mit","repos":"jacebrowning\/template-python-demo"} {"commit":"70808a2243ebf04aa86d5b4539950b22cd96cc7d","old_file":"maras\/utils\/__init__.py","new_file":"maras\/utils\/__init__.py","old_contents":"'''\nMisc utilities\n'''\n\n# Import python libs\nimport os\nimport binascii\n\n\ndef rand_hex_str(size):\n '''\n Return a random string of the passed size using hex encoding\n '''\n return binascii.hexlify(os.urandom(size\/2))\n\n\ndef rand_raw_str(size):\n '''\n Return a raw byte string of the given size\n '''\n return os.urandom(size)\n","new_contents":"'''\nMisc utilities\n'''\n\n# Import python libs\nimport os\nimport time\nimport struct\nimport binascii\nimport datetime\n\n# create a standard epoch so all platforms will count revs from\n# a standard epoch of jan 1 2014\nSTD_EPOCH = time.mktime(datetime.datetime(2014, 1, 1).timetuple())\n\n\ndef rand_hex_str(size):\n '''\n Return a random string of the passed size using hex encoding\n '''\n return binascii.hexlify(os.urandom(size\/2))\n\n\ndef rand_raw_str(size):\n '''\n Return a raw byte string of the given size\n '''\n return os.urandom(size)\n\n\ndef gen_rev():\n '''\n Return a revision based on timestamp\n '''\n r_time = time.time() - STD_EPOCH\n return struct.pack('>Q', r_time * 1000000)\n","subject":"Add rev generation via normalized timestamps","message":"Add rev generation via normalized timestamps\n","lang":"Python","license":"apache-2.0","repos":"thatch45\/maras"} {"commit":"ae2d52e323ea8959caf474d23de857d59b5b6ca8","old_file":"spacy\/tests\/regression\/test_issue3625.py","new_file":"spacy\/tests\/regression\/test_issue3625.py","old_contents":"from __future__ import unicode_literals\n\nfrom spacy.lang.hi import Hindi\n\ndef test_issue3625():\n \"\"\"Test that default punctuation rules applies to hindi unicode characters\"\"\"\n nlp = Hindi()\n doc = nlp(u\"hi. how हुए. होटल, होटल\")\n assert [token.text for token in doc] == ['hi', '.', 'how', 'हुए', '.', 'होटल', ',', 'होटल']","new_contents":"# coding: utf8\nfrom __future__ import unicode_literals\n\nfrom spacy.lang.hi import Hindi\n\ndef test_issue3625():\n \"\"\"Test that default punctuation rules applies to hindi unicode characters\"\"\"\n nlp = Hindi()\n doc = nlp(u\"hi. how हुए. होटल, होटल\")\n assert [token.text for token in doc] == ['hi', '.', 'how', 'हुए', '.', 'होटल', ',', 'होटल']","subject":"Add default encoding utf-8 for test file","message":"Add default encoding utf-8 for test file\n","lang":"Python","license":"mit","repos":"honnibal\/spaCy,spacy-io\/spaCy,explosion\/spaCy,honnibal\/spaCy,explosion\/spaCy,explosion\/spaCy,honnibal\/spaCy,spacy-io\/spaCy,spacy-io\/spaCy,explosion\/spaCy,explosion\/spaCy,honnibal\/spaCy,spacy-io\/spaCy,spacy-io\/spaCy,spacy-io\/spaCy,explosion\/spaCy"} {"commit":"ce873b24318fd6493f570f370db1d2c2d244bdcc","old_file":"joby\/spiders\/data_science_jobs.py","new_file":"joby\/spiders\/data_science_jobs.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom logging import getLogger\nfrom scrapy.spiders import Rule, CrawlSpider\nfrom scrapy.linkextractors import LinkExtractor\n\n\nclass DataScienceJobsSpider(CrawlSpider):\n log = getLogger(__name__)\n name = 'data-science-jobs'\n allowed_domains = ['www.data-science-jobs.com', 'fonts.googleapis.com', 'jobs.lever.com']\n start_urls = ['http:\/\/www.data-science-jobs.com\/']\n test = Rule(LinkExtractor(allow='family'), callback='parse')\n test2 = Rule(LinkExtractor(allow='comtravo'), callback='parse')\n job_links = Rule(LinkExtractor(allow='detail\\\/'), callback='parse')\n pagination_links = Rule(LinkExtractor(allow='\\?page=\\d+'), callback='parse')\n rules = [job_links, pagination_links, test, test2]\n response = None\n\n def parse(self, response):\n self.log.info('Parsing %s', response.url)\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom logging import getLogger\nfrom scrapy.spiders import Rule, CrawlSpider\nfrom scrapy.linkextractors import LinkExtractor\n\n\nclass DataScienceJobsSpider(CrawlSpider):\n log = getLogger(__name__)\n name = 'data-science-jobs'\n allowed_domains = ['www.data-science-jobs.com']\n start_urls = ['http:\/\/www.data-science-jobs.com']\n job_links = Rule(LinkExtractor(allow='detail\\\/'), callback='parse_job')\n pagination_links = Rule(LinkExtractor(allow='page=\\d+'))\n rules = [job_links, pagination_links]\n\n def parse_job(self, response):\n self.log.info('Parsing %s', response.url)\n\n\n","subject":"Rename the parser function to parse_jobs.","message":"Rename the parser function to parse_jobs.\n","lang":"Python","license":"mit","repos":"cyberbikepunk\/job-spiders"} {"commit":"b77e8f9a081517701cccf9f177c81eaca877e8c7","old_file":"pombola\/images\/admin.py","new_file":"pombola\/images\/admin.py","old_contents":"from django.contrib import admin\nfrom django.contrib.contenttypes.generic import GenericTabularInline\n\nfrom sorl.thumbnail import get_thumbnail\nfrom sorl.thumbnail.admin import AdminImageMixin\n\nfrom pombola.images import models\n\n\n\n\n\n\nclass ImageAdmin(AdminImageMixin, admin.ModelAdmin):\n list_display = [ 'thumbnail', 'content_object', 'is_primary', 'source', ]\n\n def thumbnail(self, obj):\n im = get_thumbnail(obj.image, '100x100')\n return '<img src=\"%s\" \/>' % ( im.url )\n thumbnail.allow_tags = True\n\n\nclass ImageAdminInline(AdminImageMixin, GenericTabularInline):\n model = models.Image\n extra = 0\n can_delete = True\n\nadmin.site.register( models.Image, ImageAdmin )\n","new_contents":"from django.contrib import admin\nfrom django.contrib.contenttypes.generic import GenericTabularInline\n\nfrom sorl.thumbnail import get_thumbnail\nfrom sorl.thumbnail.admin import AdminImageMixin\n\nfrom pombola.images import models\n\n\n\n\n\n\nclass ImageAdmin(AdminImageMixin, admin.ModelAdmin):\n list_display = [ 'thumbnail', 'content_object', 'is_primary', 'source', ]\n\n def thumbnail(self, obj):\n if obj.image:\n im = get_thumbnail(obj.image, '100x100')\n return '<img src=\"%s\" \/>' % ( im.url )\n else:\n return \"NO IMAGE FOUND\"\n thumbnail.allow_tags = True\n\n\nclass ImageAdminInline(AdminImageMixin, GenericTabularInline):\n model = models.Image\n extra = 0\n can_delete = True\n\nadmin.site.register( models.Image, ImageAdmin )\n","subject":"Handle entries that have no image associated with them","message":"Handle entries that have no image associated with them\n","lang":"Python","license":"agpl-3.0","repos":"ken-muturi\/pombola,mysociety\/pombola,geoffkilpin\/pombola,hzj123\/56th,ken-muturi\/pombola,patricmutwiri\/pombola,geoffkilpin\/pombola,ken-muturi\/pombola,mysociety\/pombola,ken-muturi\/pombola,mysociety\/pombola,hzj123\/56th,mysociety\/pombola,patricmutwiri\/pombola,patricmutwiri\/pombola,geoffkilpin\/pombola,hzj123\/56th,ken-muturi\/pombola,ken-muturi\/pombola,patricmutwiri\/pombola,geoffkilpin\/pombola,hzj123\/56th,patricmutwiri\/pombola,patricmutwiri\/pombola,hzj123\/56th,geoffkilpin\/pombola,mysociety\/pombola,geoffkilpin\/pombola,mysociety\/pombola,hzj123\/56th"} {"commit":"a03b166f8297783819a43eeb78e5af4d52d11bcc","old_file":"carbonate\/list.py","new_file":"carbonate\/list.py","old_contents":"import os\nimport re\n\n# Use the built-in version of scandir\/walk if possible, otherwise\n# use the scandir module version\ntry:\n from os import scandir, walk\nexcept ImportError:\n from scandir import scandir, walk\n\n\ndef listMetrics(storage_dir, follow_sym_links=False, metric_suffix='wsp'):\n metric_regex = re.compile(\".*\\.%s$\" % metric_suffix)\n\n storage_dir = storage_dir.rstrip(os.sep)\n\n for root, dirnames, filenames in walk(storage_dir,\n followlinks=follow_sym_links):\n for filename in filenames:\n if metric_regex.match(filename):\n root_path = root[len(storage_dir) + 1:]\n m_path = os.path.join(root_path, filename)\n m_name, m_ext = os.path.splitext(m_path)\n m_name = m_name.replace('\/', '.')\n yield m_name\n","new_contents":"import os\nimport re\n\n# Use the built-in version of scandir\/walk if possible, otherwise\n# use the scandir module version\ntry:\n from os import scandir, walk # noqa # pylint: disable=unused-import\nexcept ImportError:\n from scandir import scandir, walk # noqa # pylint: disable=unused-import\n\n\ndef listMetrics(storage_dir, follow_sym_links=False, metric_suffix='wsp'):\n metric_regex = re.compile(\".*\\.%s$\" % metric_suffix)\n\n storage_dir = storage_dir.rstrip(os.sep)\n\n for root, _, filenames in walk(storage_dir, followlinks=follow_sym_links):\n for filename in filenames:\n if metric_regex.match(filename):\n root_path = root[len(storage_dir) + 1:]\n m_path = os.path.join(root_path, filename)\n m_name, m_ext = os.path.splitext(m_path)\n m_name = m_name.replace('\/', '.')\n yield m_name\n","subject":"Make pylint happy as per graphite-web example","message":"Make pylint happy as per graphite-web example\n","lang":"Python","license":"mit","repos":"criteo-forks\/carbonate,jssjr\/carbonate,deniszh\/carbonate,graphite-project\/carbonate,jssjr\/carbonate,graphite-project\/carbonate,criteo-forks\/carbonate,jssjr\/carbonate,deniszh\/carbonate,deniszh\/carbonate,criteo-forks\/carbonate,graphite-project\/carbonate"} {"commit":"8a32c26aee8df988a65f30f8864ddf2df05225c8","old_file":"lib\/file_out_gittifier.py","new_file":"lib\/file_out_gittifier.py","old_contents":"#!\/usr\/bin\/python\n\nimport sys\nimport os\n\n\nif __name__ == '__main__':\n testing = len(sys.argv) > 2 and sys.argv[1] == '--testing'\n offset = 2 if testing else 1\n\n stFilesOnly = all([f.endswith('.st') for f in sys.argv[offset:]])\n filesExist = all([os.path.isfile(f) for f in sys.argv[offset:]])\n if len(sys.argv) < 2 or not stFilesOnly or not filesExist:\n print 'This program requires existing .st files as input parameter.'\n sys.exit(1)\n\n filenames = sys.argv[offset:]\n for filename in filenames:\n content = ''\n with open(filename, 'rb') as f:\n content = f.read()\n\n if testing:\n for line in content.splitlines(True):\n if b'\\r' in line or b'\\f' in line:\n print '%s does not seem to be gittified.' % filename\n sys.exit(1)\n else:\n with open(filename, 'wb') as f:\n for line in content.splitlines():\n f.write(line.rstrip(b'\\f') + b'\\n')\n\n if testing:\n print 'Tested %s file(s).' % len(filenames)\n else:\n print 'Converted %s file(s).' % len(filenames)\n","new_contents":"#!\/usr\/bin\/python\n\nimport sys\nimport os\n\n\nif __name__ == '__main__':\n testing = len(sys.argv) > 2 and sys.argv[1] == '--testing'\n offset = 2 if testing else 1\n\n stFilesOnly = all([f.endswith('.st') for f in sys.argv[offset:]])\n filesExist = all([os.path.isfile(f) for f in sys.argv[offset:]])\n if len(sys.argv) < 2 or not stFilesOnly or not filesExist:\n print 'This program requires existing .st files as input parameter.'\n sys.exit(1)\n\n filenames = sys.argv[offset:]\n for filename in filenames:\n content = ''\n with open(filename, 'rb') as f:\n content = f.read()\n\n if testing:\n for line in content.splitlines(True):\n if b'\\r' in line or b'\\f' in line:\n print '%s does not seem to be gittified.' % filename\n sys.exit(1)\n else:\n with open(filename, 'wb') as f:\n for line in content.splitlines():\n f.write(line.rstrip(b'\\f') + b'\\n')\n\n if testing:\n print 'Tested %s file(s) for git compatibility.' % len(filenames)\n else:\n print 'Converted %s file(s).' % len(filenames)\n","subject":"Improve success message [ci skip]","message":"Improve success message [ci skip]\n","lang":"Python","license":"mit","repos":"hpi-swa\/filetreeCI,dalehenrich\/smalltalkCI,fniephaus\/filetreeCI,hpi-swa\/filetreeCI,fniephaus\/filetreeCI,hpi-swa\/smalltalkCI,hpi-swa\/filetreeCI,fniephaus\/filetreeCI,hpi-swa\/smalltalkCI,dalehenrich\/smalltalkCI,peteruhnak\/smalltalkCI,peteruhnak\/smalltalkCI,peteruhnak\/smalltalkCI,dalehenrich\/smalltalkCI"} {"commit":"119e95dedaf6633e1ca6367bfd13fa08192033bd","old_file":"pywinauto\/unittests\/testall.py","new_file":"pywinauto\/unittests\/testall.py","old_contents":"import unittest\r\n\r\nimport os.path\r\nimport os\r\nimport sys\r\nsys.path.append(\".\")\r\n\r\n#from pywinauto.timings import Timings\r\n#Timings.Fast()\r\n\r\nexcludes = ['test_sendkeys']\r\n\r\ndef run_tests():\r\n testfolder = os.path.abspath(os.path.split(__file__)[0])\r\n\r\n sys.path.append(testfolder)\r\n\r\n\r\n for root, dirs, files in os.walk(testfolder):\r\n test_modules = [\r\n file.replace('.py', '') for file in files if\r\n file.startswith('test_') and\r\n file.endswith('.py')]\r\n\r\n test_modules = [mod for mod in test_modules if mod.lower() not in excludes]\r\n for mod in test_modules:\r\n\r\n #globals().update(__import__(mod, globals(), locals()).__dict__)\r\n # import it\r\n imported_mod = __import__(mod, globals(), locals())\r\n #print imported_mod.__dict__\r\n globals().update(imported_mod.__dict__)\r\n\r\n\r\n #runner = unittest.TextTestRunner(verbosity = 2)\r\n unittest.main()#testRunner = runner)\r\n\r\nif __name__ == '__main__':\r\n run_tests()","new_contents":"import os\r\nimport sys\r\nimport unittest\r\n\r\nimport coverage\r\n\r\n# needs to be called before importing the modules\r\ncov = coverage.coverage(branch = True)\r\ncov.start()\r\n\r\ntestfolder = os.path.abspath(os.path.dirname(__file__))\r\npackage_root = os.path.abspath(os.path.join(testfolder, r\"..\\..\"))\r\nsys.path.append(package_root)\r\n\r\nimport pywinauto\r\n\r\nmodules_to_test = [pywinauto]\r\n\r\n\r\ndef run_tests():\r\n excludes = ['test_sendkeys']\r\n\r\n suite = unittest.TestSuite()\r\n\r\n sys.path.append(testfolder)\r\n\r\n for root, dirs, files in os.walk(testfolder):\r\n test_modules = [\r\n file.replace('.py', '') for file in files if\r\n file.startswith('test_') and\r\n file.endswith('.py')]\r\n\r\n test_modules = [mod for mod in test_modules if mod.lower() not in excludes]\r\n for mod in test_modules:\r\n\r\n #globals().update(__import__(mod, globals(), locals()).__dict__)\r\n # import it\r\n imported_mod = __import__(mod, globals(), locals())\r\n\r\n suite.addTests(\r\n unittest.defaultTestLoader.loadTestsFromModule(imported_mod))\r\n\r\n #unittest.main()#testRunner = runner)\r\n\r\n #runner = unittest.TextTestRunner(verbosity = 2)\r\n unittest.TextTestRunner(verbosity=1).run(suite)\r\n cov.stop()\r\n #print cov.analysis()\r\n print cov.report()\r\n cov.html_report(\r\n directory = os.path.join(package_root, \"Coverage_report\"))\r\n\r\n\r\nif __name__ == '__main__':\r\n run_tests()","subject":"Synchronize testing module with BetterBatch one - and integrate Coverage reporting","message":"Synchronize testing module with BetterBatch one - and integrate Coverage reporting\n","lang":"Python","license":"bsd-3-clause","repos":"cessor\/pywinauto,bombilee\/pywinauto,ohio813\/pywinauto,nameoffnv\/pywinauto,yongxin1029\/pywinauto,clonly\/pywinauto,vsajip\/pywinauto,cessor\/pywinauto,LogicalKnight\/pywinauto,ohio813\/pywinauto,nameoffnv\/pywinauto,ldhwin\/pywinauto,airelil\/pywinauto,drinkertea\/pywinauto,prasen-ftech\/pywinauto,vsajip\/pywinauto,wilsoc5\/pywinauto,pjquirk\/pjquirk-dotnetnames,LogicalKnight\/pywinauto,ohio813\/pywinauto,mjakop\/pywinauto,ldhwin\/pywinauto,pjquirk\/pjquirk-dotnetnames,wilsoc5\/pywinauto,nameoffnv\/pywinauto,MagazinnikIvan\/pywinauto,cessor\/pywinauto,vane\/pywinauto,moden-py\/pywinauto,ldhwin\/pywinauto,vane\/pywinauto,bombilee\/pywinauto,vsajip\/pywinauto,prasen-ftech\/pywinauto,manojklm\/pywinauto-x64,LogicalKnight\/pywinauto,vasily-v-ryabov\/pywinauto,pjquirk\/pjquirk-dotnetnames,LogicalKnight\/pywinauto,cetygamer\/pywinauto,clonly\/pywinauto,prasen-ftech\/pywinauto,manojklm\/pywinauto-x64,pywinauto\/pywinauto,vane\/pywinauto,bombilee\/pywinauto,mindw\/pywinauto,yongxin1029\/pywinauto,mjakop\/pywinauto,moden-py\/pywinauto,manojklm\/pywinauto-x64,mindw\/pywinauto,pjquirk\/pjquirk-dotnetnames"} {"commit":"29032ee9dc69b1f3226358c3a6b74a7e42d71f07","old_file":"generationkwh\/amortizations.py","new_file":"generationkwh\/amortizations.py","old_contents":"# -*- coding:utf8 -*-\n\nfrom plantmeter.isodates import isodate\nfrom dateutil.relativedelta import relativedelta\n\nwaitYears = 1\nexpirationYears = 25\n\ndef previousAmortizationDate(purchase_date, current_date):\n\n years = relativedelta(\n isodate(current_date),\n isodate(purchase_date),\n ).years\n\n if years <= waitYears:\n return None\n\n firstAmortization = (\n isodate(purchase_date)\n + relativedelta(years = min(years,expirationYears)\n )\n\n return str(amortizationDate)\n\n\ndef pendingAmortization(purchase_date, current_date, investment_amount, amortized_amount):\n\n\n years = relativedelta(\n isodate(current_date),\n isodate(purchase_date),\n ).years\n\n yearly_amortitzation = investment_amount \/ expirationYears\n\n if years <= waitYears:\n return 0\n\n if years >= expirationYears:\n return investment_amount - amortized_amount\n\n toAmortize = (years-1)*yearly_amortitzation - amortized_amount\n return max(0, toAmortize)\n\n\n# vim: et ts=4 sw=4\n","new_contents":"# -*- coding:utf8 -*-\n\nfrom plantmeter.isodates import isodate\nfrom dateutil.relativedelta import relativedelta\n\nwaitYears = 1\nexpirationYears = 25\n\ndef previousAmortizationDate(purchase_date, current_date):\n\n years = relativedelta(\n isodate(current_date),\n isodate(purchase_date),\n ).years\n\n if years <= waitYears:\n return None\n\n firstAmortization = (\n isodate(purchase_date)\n + relativedelta(years = min(years,expirationYears)\n ))\n\n return str(firstAmortization)\n\n\ndef pendingAmortization(purchase_date, current_date, investment_amount, amortized_amount):\n\n\n years = relativedelta(\n isodate(current_date),\n isodate(purchase_date),\n ).years\n\n yearly_amortitzation = investment_amount \/ expirationYears\n\n if years <= waitYears:\n return 0\n\n if years >= expirationYears:\n return investment_amount - amortized_amount\n\n toAmortize = (years-1)*yearly_amortitzation - amortized_amount\n return max(0, toAmortize)\n\n\n# vim: et ts=4 sw=4\n","subject":"Modify return variable and partenesis","message":"Modify return variable and partenesis\n","lang":"Python","license":"agpl-3.0","repos":"Som-Energia\/somenergia-generationkwh,Som-Energia\/somenergia-generationkwh"} {"commit":"9a879fb583f7f4190a4601a9a488ba61414395e0","old_file":"kivymd\/card.py","new_file":"kivymd\/card.py","old_contents":"# -*- coding: utf-8 -*-\nfrom kivy.lang import Builder\nfrom kivy.properties import BoundedNumericProperty, ReferenceListProperty\nfrom kivy.uix.boxlayout import BoxLayout\nfrom kivymd.elevationbehavior import ElevationBehavior\nfrom kivymd.theming import ThemableBehavior\nfrom kivy.metrics import dp\n\nBuilder.load_string('''\n<MDCard>\n canvas:\n Color:\n rgba: self.background_color\n RoundedRectangle:\n size: self.size\n pos: self.pos\n radius: [self.border_radius]\n background_color: self.theme_cls.bg_light\n''')\n\n\nclass MDCard(ThemableBehavior, ElevationBehavior, BoxLayout):\n r = BoundedNumericProperty(1., min=0., max=1.)\n g = BoundedNumericProperty(1., min=0., max=1.)\n b = BoundedNumericProperty(1., min=0., max=1.)\n a = BoundedNumericProperty(0., min=0., max=1.)\n \n border_radius = BoundedNumericProperty(dp(3),min=0)\n background_color = ReferenceListProperty(r, g, b, a)\n","new_contents":"# -*- coding: utf-8 -*-\nfrom kivy.lang import Builder\nfrom kivy.properties import BoundedNumericProperty, ReferenceListProperty, ListProperty,BooleanProperty\nfrom kivy.uix.boxlayout import BoxLayout\nfrom kivymd.elevationbehavior import ElevationBehavior\nfrom kivymd.theming import ThemableBehavior\nfrom kivy.metrics import dp\n\nBuilder.load_string('''\n<MDCard>\n canvas:\n Color:\n rgba: self.background_color\n RoundedRectangle:\n size: self.size\n pos: self.pos\n radius: [self.border_radius]\n Color:\n rgba: self.theme_cls.divider_color\n a: self.border_color_a\n Line:\n rounded_rectangle: (self.pos[0],self.pos[1],self.size[0],self.size[1],self.border_radius) \n background_color: self.theme_cls.bg_light\n''')\n\n\nclass MDCard(ThemableBehavior, ElevationBehavior, BoxLayout):\n r = BoundedNumericProperty(1., min=0., max=1.)\n g = BoundedNumericProperty(1., min=0., max=1.)\n b = BoundedNumericProperty(1., min=0., max=1.)\n a = BoundedNumericProperty(0., min=0., max=1.)\n \n border_radius = BoundedNumericProperty(dp(3),min=0)\n border_color_a = BoundedNumericProperty(0, min=0., max=1.)\n background_color = ReferenceListProperty(r, g, b, a)\n","subject":"Add border as option (set via alpha)","message":"Add border as option (set via alpha)","lang":"Python","license":"mit","repos":"cruor99\/KivyMD"} {"commit":"7ea053bfc1b557ce4a4df4905af4a5491517490b","old_file":"default_config.py","new_file":"default_config.py","old_contents":"# Default Config\n# Override these values in the instance\/local_config.py file, not this one.\n\nDEBUG = None\n\nDOMAIN = 'localhost'\n\nHOST = 'localhost'\nPORT = 5000\n","new_contents":"\"\"\"\\\nDefault Config\n\nOverride these values in the instance\/local_config.py file, not this one.\n\"\"\"\n\n\nDEBUG = None\n\nDOMAIN = 'localhost'\n\nHOST = 'localhost'\nPORT = 5000\n","subject":"Use doc-string, not comments in default config.","message":"Use doc-string, not comments in default config.\n","lang":"Python","license":"mit","repos":"joeyespo\/tabhouse.org,joeyespo\/tabhouse,joeyespo\/tabhouse,joeyespo\/tabhouse.org"} {"commit":"1736d7b7aed3ce3049186ce97e24941de0187caf","old_file":"oidc_provider\/lib\/utils\/common.py","new_file":"oidc_provider\/lib\/utils\/common.py","old_contents":"from django.conf import settings as django_settings\nfrom django.core.urlresolvers import reverse\n\nfrom oidc_provider import settings\n\n\ndef get_issuer():\n \"\"\"\n Construct the issuer full url. Basically is the site url with some path\n appended.\n \"\"\"\n site_url = settings.get('SITE_URL')\n path = reverse('oidc_provider:provider_info') \\\n .split('\/.well-known\/openid-configuration\/')[0]\n issuer = site_url + path\n\n return issuer\n\n\ndef get_rsa_key():\n \"\"\"\n Load the rsa key previously created with `creatersakey` command.\n \"\"\"\n file_path = settings.get('OIDC_RSA_KEY_FOLDER') + '\/OIDC_RSA_KEY.pem'\n with open(file_path, 'r') as f:\n key = f.read()\n\n return key\n","new_contents":"from django.conf import settings as django_settings\nfrom django.core.urlresolvers import reverse\n\nfrom oidc_provider import settings\n\n\ndef get_issuer():\n \"\"\"\n Construct the issuer full url. Basically is the site url with some path\n appended.\n \"\"\"\n site_url = settings.get('SITE_URL')\n path = reverse('oidc_provider:provider_info') \\\n .split('\/.well-known\/openid-configuration\/')[0]\n issuer = site_url + path\n\n return issuer\n\n\ndef get_rsa_key():\n \"\"\"\n Load the rsa key previously created with `creatersakey` command.\n \"\"\"\n file_path = settings.get('OIDC_RSA_KEY_FOLDER') + '\/OIDC_RSA_KEY.pem'\n try:\n with open(file_path, 'r') as f:\n key = f.read()\n except IOError:\n raise IOError('We could not find your key file on: ' + file_path)\n\n return key\n","subject":"Add IOError custom message when rsa key file is missing.","message":"Add IOError custom message when rsa key file is missing.\n","lang":"Python","license":"mit","repos":"ByteInternet\/django-oidc-provider,torreco\/django-oidc-provider,juanifioren\/django-oidc-provider,bunnyinc\/django-oidc-provider,wayward710\/django-oidc-provider,bunnyinc\/django-oidc-provider,wayward710\/django-oidc-provider,wojtek-fliposports\/django-oidc-provider,nmohoric\/django-oidc-provider,nmohoric\/django-oidc-provider,ByteInternet\/django-oidc-provider,torreco\/django-oidc-provider,wojtek-fliposports\/django-oidc-provider,juanifioren\/django-oidc-provider"} {"commit":"dc0129224dc01f4e9cdaa57ee2aff307a4f5d7d3","old_file":"project\/utils\/logger.py","new_file":"project\/utils\/logger.py","old_contents":"# -*- coding: utf-8 -*-\n\nimport datetime\nimport logging\n\nimport os\n\n\ndef set_up_logging():\n \"\"\"\n Main logger for usual bot needs\n \"\"\"\n logs_directory = os.path.join(os.path.dirname(os.path.realpath(__file__)), '..', 'logs')\n if not os.path.exists(logs_directory):\n os.mkdir(logs_directory)\n\n logger = logging.getLogger('tenhou')\n logger.setLevel(logging.DEBUG)\n\n ch = logging.StreamHandler()\n ch.setLevel(logging.DEBUG)\n\n file_name = datetime.datetime.now().strftime('%Y-%m-%d %H_%M_%S') + '.log'\n fh = logging.FileHandler(os.path.join(logs_directory, file_name))\n fh.setLevel(logging.DEBUG)\n\n formatter = logging.Formatter('%(asctime)s %(levelname)s: %(message)s', datefmt='%Y-%m-%d %H:%M:%S')\n ch.setFormatter(formatter)\n fh.setFormatter(formatter)\n\n logger.addHandler(ch)\n logger.addHandler(fh)\n\n logger = logging.getLogger('ai')\n logger.setLevel(logging.DEBUG)\n logger.addHandler(ch)\n logger.addHandler(fh)\n","new_contents":"# -*- coding: utf-8 -*-\n\nimport datetime\nimport logging\nimport os\nimport hashlib\n\nfrom utils.settings_handler import settings\n\n\ndef set_up_logging():\n \"\"\"\n Logger for tenhou communication and AI output\n \"\"\"\n logs_directory = os.path.join(os.path.dirname(os.path.realpath(__file__)), '..', 'logs')\n if not os.path.exists(logs_directory):\n os.mkdir(logs_directory)\n\n # we shouldn't be afraid about collision\n # also, we need it to distinguish different bots logs (if they were run in the same time)\n name_hash = hashlib.sha1(settings.USER_ID.encode('utf-8')).hexdigest()[:5]\n\n logger = logging.getLogger('tenhou')\n logger.setLevel(logging.DEBUG)\n\n ch = logging.StreamHandler()\n ch.setLevel(logging.DEBUG)\n\n file_name = '{}_{}.log'.format(name_hash, datetime.datetime.now().strftime('%Y-%m-%d %H_%M_%S'))\n fh = logging.FileHandler(os.path.join(logs_directory, file_name))\n fh.setLevel(logging.DEBUG)\n\n formatter = logging.Formatter('%(asctime)s %(levelname)s: %(message)s', datefmt='%Y-%m-%d %H:%M:%S')\n ch.setFormatter(formatter)\n fh.setFormatter(formatter)\n\n logger.addHandler(ch)\n logger.addHandler(fh)\n\n logger = logging.getLogger('ai')\n logger.setLevel(logging.DEBUG)\n logger.addHandler(ch)\n logger.addHandler(fh)\n","subject":"Add hash from the bot name to the log name","message":"Add hash from the bot name to the log name\n","lang":"Python","license":"mit","repos":"huangenyan\/Lattish,MahjongRepository\/tenhou-python-bot,MahjongRepository\/tenhou-python-bot,huangenyan\/Lattish"} {"commit":"a174fbd637bf9ccc7b8a97a251c016495f92f6a9","old_file":"eliot\/__init__.py","new_file":"eliot\/__init__.py","old_contents":"\"\"\"\nEliot: Logging as Storytelling\n\n Suppose we turn from outside estimates of a man, to wonder, with keener\n interest, what is the report of his own consciousness about his doings or\n capacity: with what hindrances he is carrying on his daily labors; what\n fading of hopes, or what deeper fixity of self-delusion the years are\n marking off within him; and with what spirit he wrestles against universal\n pressure, which will one day be too heavy for him, and bring his heart to\n its final pause.\n\n -- George Eliot, \"Middlemarch\"\n\"\"\"\n\nfrom ._version import __version__\n\n# Expose the public API:\nfrom ._message import Message\nfrom ._action import startAction, startTask, Action\nfrom ._output import ILogger, Logger, MemoryLogger\nfrom ._validation import Field, MessageType, ActionType\nfrom ._traceback import writeTraceback, writeFailure\naddDestination = Logger._destinations.add\nremoveDestination = Logger._destinations.remove\n\n\n__all__ = [\"Message\", \"writeTraceback\", \"writeFailure\",\n \"startAction\", \"startTask\", \"Action\",\n \"Field\", \"MessageType\", \"ActionType\",\n \"ILogger\", \"Logger\", \"MemoryLogger\", \"addDestination\",\n \"removeDestination\",\n\n \"__version__\",\n ]\n","new_contents":"\"\"\"\nEliot: Logging as Storytelling\n\n Suppose we turn from outside estimates of a man, to wonder, with keener\n interest, what is the report of his own consciousness about his doings or\n capacity: with what hindrances he is carrying on his daily labors; what\n fading of hopes, or what deeper fixity of self-delusion the years are\n marking off within him; and with what spirit he wrestles against universal\n pressure, which will one day be too heavy for him, and bring his heart to\n its final pause.\n\n -- George Eliot, \"Middlemarch\"\n\"\"\"\n\nfrom ._version import __version__\n\n# Expose the public API:\nfrom ._message import Message\nfrom ._action import startAction, startTask, Action\nfrom ._output import ILogger, Logger, MemoryLogger\nfrom ._validation import Field, fields, MessageType, ActionType\nfrom ._traceback import writeTraceback, writeFailure\naddDestination = Logger._destinations.add\nremoveDestination = Logger._destinations.remove\n\n\n__all__ = [\"Message\", \"writeTraceback\", \"writeFailure\",\n \"startAction\", \"startTask\", \"Action\",\n \"Field\", \"fields\", \"MessageType\", \"ActionType\",\n \"ILogger\", \"Logger\", \"MemoryLogger\", \"addDestination\",\n \"removeDestination\",\n\n \"__version__\",\n ]\n","subject":"Add fields to the public API.","message":"Add fields to the public API.\n","lang":"Python","license":"apache-2.0","repos":"ClusterHQ\/eliot,ScatterHQ\/eliot,iffy\/eliot,ScatterHQ\/eliot,ScatterHQ\/eliot"} {"commit":"fb1f6f30fc7ba2d3dcce357168a05669c934c234","old_file":"build\/oggm\/run_test.py","new_file":"build\/oggm\/run_test.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\nos.environ[\"MPLBACKEND\"] = 'agg'\n\nimport matplotlib\nmatplotlib.use('agg')\n\nimport pytest\nimport oggm\nimport sys\n\nimport ssl\nssl._create_default_https_context = ssl._create_unverified_context\n\ninitial_dir = os.getcwd()\noggm_file = os.path.abspath(oggm.__file__)\noggm_dir = os.path.dirname(oggm_file)\n\nsys.exit(pytest.main([oggm_dir, '--mpl']))\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os\nos.environ[\"MPLBACKEND\"] = 'agg'\n\nimport matplotlib\nmatplotlib.use('agg')\n\nimport pytest\nimport oggm\nimport sys\n\nimport ssl\nssl._create_default_https_context = ssl._create_unverified_context\n\nif os.name == 'nt':\n sys.exit(0)\n\ninitial_dir = os.getcwd()\noggm_file = os.path.abspath(oggm.__file__)\noggm_dir = os.path.dirname(oggm_file)\n\nsys.exit(pytest.main([oggm_dir, '--mpl']))\n","subject":"Disable testing on Windows for now, it just takes too long for any CI service","message":"Disable testing on Windows for now, it just takes too long for any CI service\n","lang":"Python","license":"mit","repos":"OGGM\/OGGM-Anaconda"} {"commit":"b1cc99458d22b8ed54326de6b4eafececb3a8093","old_file":"jobs\/telemetry_aggregator.py","new_file":"jobs\/telemetry_aggregator.py","old_contents":"#!\/home\/hadoop\/anaconda2\/bin\/ipython\n\nimport logging\nfrom os import environ\nfrom mozaggregator.aggregator import aggregate_metrics\nfrom mozaggregator.db import submit_aggregates\n\nlogger = logging.getLogger(__name__)\nlogger.setLevel(logging.INFO)\nlogger.addHandler(logging.StreamHandler())\n\ndate = environ['date']\n\nlogger.info(\"Running job for {}\".format(date))\naggregates = aggregate_metrics(sc, (\"nightly\", \"aurora\", \"beta\", \"release\"), date)\nlogger.info(\"Number of build-id aggregates: {}\".format(aggregates[0].count()))\nlogger.info(\"Number of submission date aggregates: {}\".format(aggregates[1].count()))\nsubmit_aggregates(aggregates)\n","new_contents":"#!\/home\/hadoop\/anaconda2\/bin\/ipython\n\nimport logging\nfrom os import environ\nfrom mozaggregator.aggregator import aggregate_metrics\nfrom mozaggregator.db import submit_aggregates\n\ndate = environ['date']\nprint \"Running job for {}\".format(date)\naggregates = aggregate_metrics(sc, (\"nightly\", \"aurora\", \"beta\", \"release\"), date)\nprint \"Number of build-id aggregates: {}\".format(aggregates[0].count())\nprint \"Number of submission date aggregates: {}\".format(aggregates[1].count())\nsubmit_aggregates(aggregates)\n","subject":"Use simple prints for logging.","message":"Use simple prints for logging.\n","lang":"Python","license":"mpl-2.0","repos":"opentrials\/opentrials-airflow,opentrials\/opentrials-airflow"} {"commit":"1983885acfccfe4ffa010401fd9ef0971bb6c12c","old_file":"etcd3\/__init__.py","new_file":"etcd3\/__init__.py","old_contents":"from __future__ import absolute_import\n\nfrom etcd3.client import Etcd3Client\nfrom etcd3.client import client\nfrom etcd3.client import Transactions\n\n__author__ = 'Louis Taylor'\n__email__ = 'louis@kragniz.eu'\n__version__ = '0.1.0'\n\n__all__ = ['Etcd3Client', 'client', 'etcdrpc', 'utils', 'Transactions']\n","new_contents":"from __future__ import absolute_import\n\nfrom etcd3.client import Etcd3Client\nfrom etcd3.client import client\nfrom etcd3.client import Transactions\nfrom etcd3.members import Member\n\n__author__ = 'Louis Taylor'\n__email__ = 'louis@kragniz.eu'\n__version__ = '0.1.0'\n\n__all__ = ['Etcd3Client', 'client', 'etcdrpc', 'utils', 'Transactions',\n 'Member']\n","subject":"Make Member part of the public api","message":"Make Member part of the public api\n","lang":"Python","license":"apache-2.0","repos":"kragniz\/python-etcd3"} {"commit":"fb1db28198b54b6288a9e7d499b43f6f1a51284c","old_file":"partner_deduplicate_by_website\/__manifest__.py","new_file":"partner_deduplicate_by_website\/__manifest__.py","old_contents":"# Copyright 2016 Tecnativa - Pedro M. Baeza\n# Copyright 2017 Tecnativa - Vicent Cubells\n# Copyright 2018 Tecnativa - Cristina Martin\n# License AGPL-3.0 or later (https:\/\/www.gnu.org\/licenses\/agpl.html).\n\n{\n \"name\": \"Deduplicate Contacts by Website\",\n \"version\": \"13.0.1.0.0\",\n \"category\": \"Tools\",\n \"website\": \"https:\/\/github.com\/OCA\/crm\",\n \"author\": \"Tecnativa, \" \"Odoo Community Association (OCA)\",\n \"license\": \"AGPL-3\",\n \"installable\": True,\n \"depends\": [\"contacts\"],\n \"data\": [\"wizards\/partner_merge_view.xml\"],\n}\n","new_contents":"# Copyright 2016 Tecnativa - Pedro M. Baeza\n# Copyright 2017 Tecnativa - Vicent Cubells\n# Copyright 2018 Tecnativa - Cristina Martin\n# License AGPL-3.0 or later (https:\/\/www.gnu.org\/licenses\/agpl.html).\n\n{\n \"name\": \"Deduplicate Contacts by Website\",\n \"version\": \"13.0.1.0.0\",\n \"category\": \"Tools\",\n \"website\": \"https:\/\/github.com\/OCA\/partner-contact\",\n \"author\": \"Tecnativa, \" \"Odoo Community Association (OCA)\",\n \"license\": \"AGPL-3\",\n \"installable\": True,\n \"depends\": [\"contacts\"],\n \"data\": [\"wizards\/partner_merge_view.xml\"],\n}\n","subject":"Fix website attribute in manifest","message":"Fix website attribute in manifest\n","lang":"Python","license":"agpl-3.0","repos":"OCA\/partner-contact,OCA\/partner-contact"} {"commit":"db84de91e665a131ad82be3ed49eb291afd5342d","old_file":"oratioignoreparser.py","new_file":"oratioignoreparser.py","old_contents":"import os\nimport re\n\n\nclass OratioIgnoreParser():\n def __init__(self):\n self.ignored_paths = [\"oratiomodule.tar.gz\"]\n\n def load(self, oratio_ignore_path):\n with open(oratio_ignore_path, \"r\") as f:\n self.ignored_paths.extend([line.strip() for line in f])\n\n def should_be_ignored(self, filepath):\n for ig in self.ignored_paths:\n compiled_regex = re.compile('^' + re.escape(ig).replace('\\\\*', '.*') + '$')\n if compiled_regex.search(filepath) or compiled_regex.search(filepath.split('\/')[-1]):\n return True\n return False\n\n def list_files(self, directory):\n filepaths = []\n ignored_files = []\n for root, dirs, files in os.walk(\".\"):\n for name in files:\n relative_path = os.path.join(root, name)\n if relative_path.startswith(\".\/\"):\n relative_path = relative_path[2:]\n if not self.should_be_ignored(relative_path):\n filepaths.append(relative_path)\n else:\n ignored_files.append(relative_path)\n return filepaths, ignored_files\n","new_contents":"import os\nimport re\n\n\nclass OratioIgnoreParser():\n def __init__(self):\n self.ignored_paths = [\"oratiomodule.tar.gz\"]\n\n def load(self, oratio_ignore_path):\n with open(oratio_ignore_path, \"r\") as f:\n self.ignored_paths.extend([line.strip() for line in f])\n\n def should_be_ignored(self, filepath):\n for ig in self.ignored_paths:\n compiled_regex = re.compile(\n '^' + re.escape(ig).replace('\\\\*', '.*') + '$'\n )\n if compiled_regex.search(filepath) or \\\n compiled_regex.search(filepath.split('\/')[-1]):\n return True\n return False\n\n def list_files(self, directory):\n filepaths = []\n ignored_files = []\n for root, dirs, files in os.walk(\".\"):\n for name in files:\n relative_path = os.path.join(root, name)\n if relative_path.startswith(\".\/\"):\n relative_path = relative_path[2:]\n if not self.should_be_ignored(relative_path):\n filepaths.append(relative_path)\n else:\n ignored_files.append(relative_path)\n return filepaths, ignored_files\n","subject":"Make all lines shorter than 80 characters","message":"Make all lines shorter than 80 characters\n","lang":"Python","license":"mit","repos":"oratio-io\/oratio-cli,oratio-io\/oratio-cli"} {"commit":"5125bbfcf96ff0d3f2690198b43ed96059eb6745","old_file":"common\/parsableText.py","new_file":"common\/parsableText.py","old_contents":"from docutils import core\n\n\nclass ParsableText:\n \"\"\"Allow to parse a string with different parsers\"\"\"\n \n def __init__(self,content,mode=\"rst\"):\n \"\"\"Init the object. Content is the string to be parsed. Mode is the parser to be used. Currently, only rst(reStructuredText) and HTML are supported\"\"\"\n if mode not in [\"rst\",\"HTML\"]:\n raise Exception(\"Unknown text parser: \"+ mode)\n self.content = content\n self.mode = mode\n \n def parse(self):\n \"\"\"Returns parsed text\"\"\"\n if self.mode == \"HTML\":\n return self.content\n else:\n return self.rst(self.content)\n \n def __str__(self):\n \"\"\"Returns parsed text\"\"\"\n return self.parse()\n \n def rst(self,s):\n \"\"\"Parses reStructuredText\"\"\"\n parts = core.publish_parts(source=s,writer_name='html')\n return parts['body_pre_docinfo']+parts['fragment']","new_contents":"from docutils import core\n\n\nclass ParsableText:\n \"\"\"Allow to parse a string with different parsers\"\"\"\n \n def __init__(self,content,mode=\"rst\"):\n \"\"\"Init the object. Content is the string to be parsed. Mode is the parser to be used. Currently, only rst(reStructuredText) and HTML are supported\"\"\"\n if mode not in [\"rst\",\"HTML\"]:\n raise Exception(\"Unknown text parser: \"+ mode)\n self.content = content\n self.mode = mode\n \n def parse(self):\n \"\"\"Returns parsed text\"\"\"\n if self.mode == \"HTML\":\n return self.content\n else:\n return self.rst(self.content)\n \n def __str__(self):\n \"\"\"Returns parsed text\"\"\"\n return self.parse()\n \n def __unicode__(self):\n \"\"\"Returns parsed text\"\"\"\n return self.parse()\n \n def rst(self,s):\n \"\"\"Parses reStructuredText\"\"\"\n parts = core.publish_parts(source=s,writer_name='html')\n return parts['body_pre_docinfo']+parts['fragment']","subject":"Fix unicode in parsable text","message":"Fix unicode in parsable text\n","lang":"Python","license":"agpl-3.0","repos":"GuillaumeDerval\/INGInious,GuillaumeDerval\/INGInious,layus\/INGInious,layus\/INGInious,layus\/INGInious,GuillaumeDerval\/INGInious,GuillaumeDerval\/INGInious,layus\/INGInious"} {"commit":"acd376d854693cacf8ca20a9971dcd2653a22429","old_file":"rlpy\/Agents\/__init__.py","new_file":"rlpy\/Agents\/__init__.py","old_contents":"from .TDControlAgent import Q_Learning, SARSA\n# for compatibility of old scripts\nQ_LEARNING = Q_Learning\nfrom .Greedy_GQ import Greedy_GQ\nfrom .LSPI import LSPI\nfrom .LSPI_SARSA import LSPI_SARSA\nfrom .NaturalActorCritic import NaturalActorCritic\n","new_contents":"from .TDControlAgent import Q_Learning, SARSA\n# for compatibility of old scripts\nQ_LEARNING = Q_Learning\nfrom .Greedy_GQ import Greedy_GQ\nfrom .LSPI import LSPI\nfrom .LSPI_SARSA import LSPI_SARSA\nfrom .NaturalActorCritic import NaturalActorCritic\nfrom .PosteriorSampling import PosteriorSampling\nfrom .UCRL import UCRL\n","subject":"Add new agents to init file","message":"Add new agents to init file\n","lang":"Python","license":"bsd-3-clause","repos":"imanolarrieta\/RL,imanolarrieta\/RL,imanolarrieta\/RL"} {"commit":"48bc050c59d60037fa719542db8f6a0c68752ed1","old_file":"config\/flask_config.py","new_file":"config\/flask_config.py","old_contents":"# flake8: noqa: E501\n\nimport config.options\n\n\n# Flask-SQLAlchemy\nSQLALCHEMY_DATABASE_URI = 'mysql:\/\/{database_user}:{database_password}@{database_host}\/{database_name}'.format(\n database_user=config.options.DATABASE_USER,\n database_password=config.options.DATABASE_PASSWORD,\n database_host=config.options.DATABASE_HOST,\n database_name=config.options.DATABASE_NAME,\n)\nSQLALCHEMY_TEST_DATABASE_URI = 'mysql:\/\/{database_user}:{database_password}@{database_host}\/{database_name}'.format(\n database_user=config.options.DATABASE_USER,\n database_password=config.options.DATABASE_PASSWORD,\n database_host=config.options.DATABASE_HOST,\n database_name=config.options.DATABASE_NAME + '_test',\n)\nSQLALCHEMY_TRACK_MODIFICATIONS = False\n\n# Flask session secret key\nSECRET_KEY = '\\xec5\\xea\\xc9\\x9f,o\\xd7v\\xac\\x06\\xe2\\xeeK2\\xb9\\x1d\\x8a\\xdel\\xb27\\x8a\\xa8>\\x07\\n\\xd4Z\\xfeO\\xa1'\n","new_contents":"# flake8: noqa: E501\n\nimport config.options\n\n\n# Flask-SQLAlchemy\nSQLALCHEMY_DATABASE_URI = 'mysql:\/\/{database_user}:{database_password}@{database_host}\/{database_name}'.format(\n database_user=config.options.DATABASE_USER,\n database_password=config.options.DATABASE_PASSWORD,\n database_host=config.options.DATABASE_HOST,\n database_name=config.options.DATABASE_NAME,\n)\nSQLALCHEMY_TEST_DATABASE_URI = 'mysql:\/\/{database_user}:{database_password}@{database_host}\/{database_name}'.format(\n database_user=config.options.DATABASE_USER,\n database_password=config.options.DATABASE_PASSWORD,\n database_host=config.options.DATABASE_HOST,\n database_name=config.options.DATABASE_NAME + '_test',\n)\nSQLALCHEMY_TRACK_MODIFICATIONS = False\n\n# Flask session cookie name\nSESSION_COOKIE_NAME = 'linkr-session'\n\n# Flask session secret key\nSECRET_KEY = '\\xec5\\xea\\xc9\\x9f,o\\xd7v\\xac\\x06\\xe2\\xeeK2\\xb9\\x1d\\x8a\\xdel\\xb27\\x8a\\xa8>\\x07\\n\\xd4Z\\xfeO\\xa1'\n","subject":"Use Linkr-unique session cookie name","message":"Use Linkr-unique session cookie name\n","lang":"Python","license":"mit","repos":"LINKIWI\/linkr,LINKIWI\/linkr,LINKIWI\/linkr"} {"commit":"3ed02baa8ad7fcd1f6ca5cccc4f67799ec79e272","old_file":"kimi.py","new_file":"kimi.py","old_contents":"# Kimi language interpreter in Python 3\n# Anjana Vakil\n# http:\/\/www.github.com\/vakila\/kimi\n\nimport sys\n\ndef tokenize(program):\n '''Take a Kimi program as a string, return the tokenized program as a list of strings.\n\n >>> tokenize(\"(+ 1 2)\")\n ['(', '+', '1', '2', ')']\n\n >>> tokenize(\"(define square (lambda x (* x x)))\")\n ['(', 'define', 'square', '(', 'lambda', 'x', '(', '*', 'x', 'x', ')', ')', ')']\n '''\n program = program.replace(\"(\", \" ( \")\n program = program.replace(\")\", \" ) \")\n tokens = program.split()\n return tokens\n\ndef parse(tokens):\n pass\n\ndef evaluate(tree):\n pass\n\n\nif __name__ == \"__main__\":\n program = sys.argv[1]\n print(tokenize(program))\n","new_contents":"# Kimi language interpreter in Python 3\n# Anjana Vakil\n# http:\/\/www.github.com\/vakila\/kimi\n\nimport sys\n\ndef tokenize(string):\n '''Take a Kimi program as a string, return the tokenized program as a list of strings.\n\n >>> tokenize(\"(+ 1 2)\")\n ['(', '+', '1', '2', ')']\n\n >>> tokenize(\"(define square (lambda x (* x x)))\")\n ['(', 'define', 'square', '(', 'lambda', 'x', '(', '*', 'x', 'x', ')', ')', ')']\n '''\n string = string.replace(\"(\", \" ( \")\n string = string.replace(\")\", \" ) \")\n tokens = string.split()\n return tokens\n\ndef parse(tokens):\n pass\n\ndef evaluate(tree):\n pass\n\n\nif __name__ == \"__main__\":\n program = sys.argv[1]\n print(tokenize(program))\n","subject":"Rename program to string in tokenize","message":"Rename program to string in tokenize\n","lang":"Python","license":"mit","repos":"vakila\/kimi"} {"commit":"592105b9ee6a9c6f3d9bd7358bc5ab18f8ded0c8","old_file":"jfr_playoff\/remote.py","new_file":"jfr_playoff\/remote.py","old_contents":"import re\n\nimport requests\n\nfrom bs4 import BeautifulSoup as bs\nfrom jfr_playoff.logger import PlayoffLogger\n\nclass RemoteUrl:\n\n url_cache = {}\n\n @classmethod\n def fetch_raw(cls, url):\n PlayoffLogger.get('remote').info(\n 'fetching content for: %s', url)\n if url not in cls.url_cache:\n request = requests.get(url)\n encoding_match = re.search(\n 'content=\".*;( )?charset=(.*?)\"',\n request.content, re.IGNORECASE)\n PlayoffLogger.get('remote').debug(\n 'Content encoding: %s',\n encoding_match.group(2))\n if encoding_match:\n request.encoding = encoding_match.group(2)\n cls.url_cache[url] = request.text\n PlayoffLogger.get('remote').info(\n 'fetched %d bytes from remote location',\n len(cls.url_cache[url]))\n return cls.url_cache[url]\n\n @classmethod\n def fetch(cls, url):\n return bs(RemoteUrl.fetch_raw(url), 'lxml')\n\n @classmethod\n def clear_cache(cls):\n cls.url_cache = {}\n","new_contents":"import re\n\nimport requests\n\nfrom bs4 import BeautifulSoup as bs\nfrom jfr_playoff.logger import PlayoffLogger\n\nclass RemoteUrl:\n\n url_cache = {}\n\n @classmethod\n def fetch_raw(cls, url):\n PlayoffLogger.get('remote').info(\n 'fetching content for: %s', url)\n if url not in cls.url_cache:\n request = requests.get(url)\n encoding_match = re.search(\n 'content=\".*;( )?charset=(.*?)\"',\n request.content, re.IGNORECASE)\n if encoding_match:\n PlayoffLogger.get('remote').debug(\n 'Content encoding: %s',\n encoding_match.group(2))\n request.encoding = encoding_match.group(2)\n cls.url_cache[url] = request.text\n PlayoffLogger.get('remote').info(\n 'fetched %d bytes from remote location',\n len(cls.url_cache[url]))\n return cls.url_cache[url]\n\n @classmethod\n def fetch(cls, url):\n return bs(RemoteUrl.fetch_raw(url), 'lxml')\n\n @classmethod\n def clear_cache(cls):\n cls.url_cache = {}\n","subject":"Print detected content encoding info only if it's actually been detected","message":"Print detected content encoding info only if it's actually been detected\n","lang":"Python","license":"bsd-2-clause","repos":"emkael\/jfrteamy-playoff,emkael\/jfrteamy-playoff"} {"commit":"a7c40b43d90f32d0da4de1389d859865ae283180","old_file":"seleniumbase\/config\/proxy_list.py","new_file":"seleniumbase\/config\/proxy_list.py","old_contents":"\"\"\"\nProxy Server \"Phone Book\".\n\nSimplify running browser tests through a proxy server\nby adding your frequently-used proxies here.\n\nNow you can do something like this on the command line:\n\"pytest SOME_TEST.py --proxy=proxy1\"\n\nFormat of PROXY_LIST server entries:\n* \"ip_address:port\" OR \"username:password@ip_address:port\"\n* \"server:port\" OR \"username:password@server:port\"\n(Do NOT include the http:\/\/ or https:\/\/ in your proxy string!)\n\nExample proxies in PROXY_LIST below are not guaranteed to be active or secure.\nIf you don't already have a proxy server to connect to,\nyou can try finding one from one of following sites:\n* https:\/\/www.us-proxy.org\/\n* https:\/\/hidemy.name\/en\/proxy-list\/\n\"\"\"\n\nPROXY_LIST = {\n \"example1\": \"46.28.229.75:3128\", # (Example) - set your own proxy here\n \"example2\": \"82.200.233.4:3128\", # (Example) - set your own proxy here\n \"example3\": \"128.199.214.87:3128\", # (Example) - set your own proxy here\n \"proxy1\": None,\n \"proxy2\": None,\n \"proxy3\": None,\n \"proxy4\": None,\n \"proxy5\": None,\n}\n","new_contents":"\"\"\"\nProxy Server \"Phone Book\".\n\nSimplify running browser tests through a proxy server\nby adding your frequently-used proxies here.\n\nNow you can do something like this on the command line:\n\"pytest SOME_TEST.py --proxy=proxy1\"\n\nFormat of PROXY_LIST server entries:\n* \"ip_address:port\" OR \"username:password@ip_address:port\"\n* \"server:port\" OR \"username:password@server:port\"\n(Do NOT include the http:\/\/ or https:\/\/ in your proxy string!)\n\nExample proxies in PROXY_LIST below are not guaranteed to be active or secure.\nIf you don't already have a proxy server to connect to,\nyou can try finding one from one of following sites:\n* https:\/\/www.us-proxy.org\/\n* https:\/\/hidemy.name\/en\/proxy-list\/\n\"\"\"\n\nPROXY_LIST = {\n \"example1\": \"45.133.182.18:18080\", # (Example) - set your own proxy here\n \"example2\": \"95.174.67.50:18080\", # (Example) - set your own proxy here\n \"example3\": \"83.97.23.90:18080\", # (Example) - set your own proxy here\n \"example4\": \"82.200.233.4:3128\", # (Example) - set your own proxy here\n \"proxy1\": None,\n \"proxy2\": None,\n \"proxy3\": None,\n \"proxy4\": None,\n \"proxy5\": None,\n}\n","subject":"Update the proxy list examples","message":"Update the proxy list examples\n","lang":"Python","license":"mit","repos":"mdmintz\/SeleniumBase,mdmintz\/SeleniumBase,seleniumbase\/SeleniumBase,seleniumbase\/SeleniumBase,seleniumbase\/SeleniumBase,mdmintz\/SeleniumBase,seleniumbase\/SeleniumBase,mdmintz\/SeleniumBase"} {"commit":"b158e65839b9b662d56bd43dfd362ad26da70184","old_file":"__init__.py","new_file":"__init__.py","old_contents":"#Shoopdawoop\nfrom . import CuraEngineBackend\n\nfrom UM.Preferences import Preferences\n\ndef getMetaData():\n return { \"name\": \"CuraEngine Backend\", \"type\": \"Backend\" }\n\ndef register(app):\n Preferences.addPreference(\"BackendLocation\",\"..\/PinkUnicornEngine\/CuraEngine\")\n engine = CuraEngineBackend.CuraEngineBackend()\n app.setBackend(engine)\n #engine.addCommand(TransferMeshCommand())\n","new_contents":"#Shoopdawoop\nfrom . import CuraEngineBackend\n\nfrom UM.Preferences import Preferences\n\ndef getMetaData():\n return { \"name\": \"CuraEngine Backend\", \"type\": \"Backend\" }\n\ndef register(app):\n Preferences.addPreference(\"BackendLocation\",\"..\/PinkUnicornEngine\/CuraEngine\")\n return CuraEngineBackend.CuraEngineBackend()\n\n","subject":"Update plugin's register functions to return the object instance instead of performing the registration themselves","message":"Update plugin's register functions to return the object instance instead of performing the registration themselves\n","lang":"Python","license":"agpl-3.0","repos":"Curahelper\/Cura,Curahelper\/Cura,bq\/Ultimaker-Cura,DeskboxBrazil\/Cura,lo0ol\/Ultimaker-Cura,quillford\/Cura,fxtentacle\/Cura,totalretribution\/Cura,DeskboxBrazil\/Cura,hmflash\/Cura,ynotstartups\/Wanhao,markwal\/Cura,ad1217\/Cura,ynotstartups\/Wanhao,derekhe\/Cura,lo0ol\/Ultimaker-Cura,senttech\/Cura,totalretribution\/Cura,bq\/Ultimaker-Cura,fxtentacle\/Cura,fieldOfView\/Cura,quillford\/Cura,senttech\/Cura,hmflash\/Cura,ad1217\/Cura,derekhe\/Cura,fieldOfView\/Cura,markwal\/Cura"} {"commit":"c612b92847dc89bb4cd4b63502c43a7a9f63c52f","old_file":"tx_salaries\/utils\/transformers\/mixins.py","new_file":"tx_salaries\/utils\/transformers\/mixins.py","old_contents":"class OrganizationMixin(object):\n @property\n def organization(self):\n return {\n 'name': self.ORGANIZATION_NAME,\n 'children': [{\n 'name': unicode(self.department),\n }],\n }\n\n","new_contents":"class OrganizationMixin(object):\n \"\"\"\n Adds a generic ``organization`` property to the class\n\n This requires that the class mixing it in adds an\n ``ORGANIZATION_NAME`` property of the main level agency or\n department.\n \"\"\"\n @property\n def organization(self):\n return {\n 'name': self.ORGANIZATION_NAME,\n 'children': [{\n 'name': unicode(self.department),\n }],\n }\n\n","subject":"Add a docblock for this mixin","message":"Add a docblock for this mixin\n","lang":"Python","license":"apache-2.0","repos":"texastribune\/tx_salaries,texastribune\/tx_salaries"} {"commit":"a08005a03ccce63a541e8e41b0d98e9c7c30cc67","old_file":"vispy\/visuals\/graphs\/layouts\/circular.py","new_file":"vispy\/visuals\/graphs\/layouts\/circular.py","old_contents":"# -*- coding: utf-8 -*-\n# Copyright (c) 2015, Vispy Development Team.\n# Distributed under the (new) BSD License. See LICENSE.txt for more info.\n\"\"\"\nCircular Layout\n===============\n\nThis module contains several graph layouts which rely heavily on circles.\n\"\"\"\n\nimport numpy as np\n\nfrom ..util import _straight_line_vertices\n\n\ndef circular(adjacency_mat, directed=False):\n \"\"\"Places all nodes on a single circle.\n\n Parameters\n ----------\n adjacency_mat : matrix or sparse\n The graph adjacency matrix\n directed : bool\n Whether the graph is directed. If this is True, is will also\n generate the vertices for arrows, which can be passed to an\n ArrowVisual.\n\n Yields\n ------\n (node_vertices, line_vertices, arrow_vertices) : tuple\n Yields the node and line vertices in a tuple. This layout only yields a\n single time, and has no builtin animation\n \"\"\"\n\n num_nodes = adjacency_mat.shape[0]\n\n t = np.arange(0, 2.0*np.pi, 2.0*np.pi\/num_nodes, dtype=np.float32)\n\n # Visual coordinate system is between 0 and 1, so generate a circle with\n # radius 0.5 and center it at the point (0.5, 0.5).\n node_coords = (0.5 * np.array([np.cos(t), np.sin(t)]) + 0.5).T\n\n line_vertices, arrows = _straight_line_vertices(adjacency_mat,\n node_coords, directed)\n\n yield node_coords, line_vertices, arrows\n","new_contents":"# -*- coding: utf-8 -*-\n# Copyright (c) 2015, Vispy Development Team.\n# Distributed under the (new) BSD License. See LICENSE.txt for more info.\n\"\"\"\nCircular Layout\n===============\n\nThis module contains several graph layouts which rely heavily on circles.\n\"\"\"\n\nimport numpy as np\n\nfrom ..util import _straight_line_vertices\n\n\ndef circular(adjacency_mat, directed=False):\n \"\"\"Places all nodes on a single circle.\n\n Parameters\n ----------\n adjacency_mat : matrix or sparse\n The graph adjacency matrix\n directed : bool\n Whether the graph is directed. If this is True, is will also\n generate the vertices for arrows, which can be passed to an\n ArrowVisual.\n\n Yields\n ------\n (node_vertices, line_vertices, arrow_vertices) : tuple\n Yields the node and line vertices in a tuple. This layout only yields a\n single time, and has no builtin animation\n \"\"\"\n\n num_nodes = adjacency_mat.shape[0]\n\n t = np.linpsace(0, 2 * np.pi, num_nodes, endpt=False, dtype=np.float32)\n\n # Visual coordinate system is between 0 and 1, so generate a circle with\n # radius 0.5 and center it at the point (0.5, 0.5).\n node_coords = (0.5 * np.array([np.cos(t), np.sin(t)]) + 0.5).T\n\n line_vertices, arrows = _straight_line_vertices(adjacency_mat,\n node_coords, directed)\n\n yield node_coords, line_vertices, arrows\n","subject":"Use the more obvious linspace instead of arange","message":"Use the more obvious linspace instead of arange\n","lang":"Python","license":"bsd-3-clause","repos":"michaelaye\/vispy,Eric89GXL\/vispy,drufat\/vispy,drufat\/vispy,drufat\/vispy,michaelaye\/vispy,ghisvail\/vispy,ghisvail\/vispy,Eric89GXL\/vispy,ghisvail\/vispy,michaelaye\/vispy,Eric89GXL\/vispy"} {"commit":"76b0c364b8bfbc553d3eedc97e4805299b8d9974","old_file":"extensions\/ExtGameController.py","new_file":"extensions\/ExtGameController.py","old_contents":"from python_cowbull_game.GameController import GameController\nfrom python_cowbull_game.GameMode import GameMode\n\n\nclass ExtGameController(GameController):\n \"\"\"\n TBC\n \"\"\"\n\n #\n # Example of defining additional game modes:\n # ==========================================\n #\n # Replace:\n # ------------------------------------------\n # additional_modes = []\n #\n # With:\n # ------------------------------------------\n # additional_modes = [\n # GameMode(mode=\"SuperTough\", priority=6, digits=10, digit_type=0),\n # GameMode(mode=\"hexTough\", priority=5, digits=3, guesses_allowed=3, digit_type=1)\n # ]\n #\n\n additional_modes = []\n\n def __init__(self, game_modes=None, mode=None, game_json=None):\n if game_modes is not None and not isinstance(game_modes, list):\n raise TypeError(\"ExtGameController expected a list of GameMode objects\")\n\n super(ExtGameController, self).__init__(\n game_json=game_json,\n mode=mode,\n game_modes=self.additional_modes + (game_modes or [])\n )\n","new_contents":"from python_cowbull_game.GameController import GameController\nfrom python_cowbull_game.GameMode import GameMode\n\n\nclass ExtGameController(GameController):\n \"\"\"\n TBC\n \"\"\"\n\n #\n # Example of defining additional game modes:\n # ==========================================\n #\n # Replace:\n # ------------------------------------------\n # additional_modes = []\n #\n # With:\n # ------------------------------------------\n # additional_modes = [\n # GameMode(mode=\"SuperTough\", priority=6, digits=10, digit_type=0),\n # GameMode(mode=\"hexTough\", priority=5, digits=3, guesses_allowed=3, digit_type=1)\n # ]\n #\n\n additional_modes = [\n GameMode(\n mode=\"hexTough\",\n priority=5,\n digits=3,\n guesses_allowed=3,\n digit_type=1,\n help_text=\"Guess a set of 3 digits between 0 and F\",\n instruction_text=\"hexTough is a hard hexidecimal based game. You need to \"\n \"guess 3 digits, each of which needs to be a hex number \"\n \"(0, 1, 2, 3, 4, 5, 6, 7, 8, 9, A, B, C, D, E, or F). \"\n \"The numbers can be passed as hex (0x0, 0xd, 0xE) or as \"\n \"strings (A, b, C, 0, 5, etc.).\"\n )\n ]\n\n def __init__(self, game_modes=None, mode=None, game_json=None):\n if game_modes is not None and not isinstance(game_modes, list):\n raise TypeError(\"ExtGameController expected a list of GameMode objects\")\n\n super(ExtGameController, self).__init__(\n game_json=game_json,\n mode=mode,\n game_modes=self.additional_modes + (game_modes or [])\n )\n","subject":"Update to include instruction and help texts in GET response.","message":"Update to include instruction and help texts in GET response.\n","lang":"Python","license":"apache-2.0","repos":"dsandersAzure\/python_cowbull_server,dsandersAzure\/python_cowbull_server"} {"commit":"19fcd893b88fd2ac9891904af93baf76b49fd5c0","old_file":"plasmapy\/_metadata.py","new_file":"plasmapy\/_metadata.py","old_contents":"##\n# Package metadata\n##\n\nimport ah_bootstrap\n\nfrom astropy_helpers.git_helpers import get_git_devstr\n\n# Name\nname = 'plasmapy'\n\n# PlasmaPy uses Semantic Versioning of the form: MAJOR.MINOR.PATCH\n#\n# - The MAJOR version changes when there are backwards incompatible changes\n# - The MINOR version changes when backwards compatible functionality is added\n# - The PATCH version changes when the public API remains the same\n#\n# During initial development releases (with MAJOR = 0), backwards compatibility\n# does not need to be maintained when MINOR is incremented.\n#\n# While a new version is being developed, '.dev' followed by the commit number\n# will be appended to the version string.\n\nversion = '0.1.dev'\n\nrelease = 'dev' in version\nif release:\n version += get_git_devstr(False)\n\n# Long description \/ docstring\ndescription = \"\"\"\nPlasmaPy is a community-developed and community-driven core Python\npackage for plasma physics.\n\"\"\"\n\n# Author(s)\nauthor = 'The PlasmaPy Community'\n","new_contents":"##\n# Package metadata\n##\n\nimport ah_bootstrap\n\nfrom astropy_helpers.git_helpers import get_git_devstr\n\n# Name\nname = 'plasmapy'\n\n# PlasmaPy uses Semantic Versioning of the form: MAJOR.MINOR.PATCH\n#\n# - The MAJOR version changes when there are backwards incompatible changes\n# - The MINOR version changes when backwards compatible functionality is added\n# - The PATCH version changes when the public API remains the same\n#\n# During initial development releases (with MAJOR = 0), backwards compatibility\n# does not need to be maintained when MINOR is incremented.\n#\n# While a new version is being developed, '.dev' followed by the commit number\n# will be appended to the version string.\n\nversion = '0.1.0.dev'\n\nrelease = 'dev' in version\n\nif release:\n version += get_git_devstr(False)\n\n# Long description \/ docstring\ndescription = \"\"\"\nPlasmaPy is a community-developed and community-driven core Python\npackage for plasma physics.\n\"\"\"\n\n# Author(s)\nauthor = 'The PlasmaPy Community'\n","subject":"Change version from 0.1.dev* to 0.1.0.dev*","message":"Change version from 0.1.dev* to 0.1.0.dev*\n\nThe semantic versioning specification requires that the major, minor,\nand patch numbers always be present.\n","lang":"Python","license":"bsd-3-clause","repos":"StanczakDominik\/PlasmaPy"} {"commit":"2eac437b9d907fb60d53522633dd278aa277ea08","old_file":"test\/user_tests\/test_models.py","new_file":"test\/user_tests\/test_models.py","old_contents":"# coding: utf-8\n\nimport unittest\nfrom test.factories import UserFactory\n\nfrom django.contrib.auth.models import User\nfrom django.db.models.signals import post_save\n\nfrom users.models import create_user_profile, Users\n\n\nclass UserTest(unittest.TestCase):\n '''User-specific tests'''\n def setUp(self):\n self.user = UserFactory.build()\n\n def tearDown(self):\n self.user = None\n\n def test_user(self):\n self.assertNotEqual(None, self.user)\n self.assertEqual('Boy', self.user.first_name)\n self.assertEqual('Factory', self.user.last_name)\n self.assertEqual('boy_factory@example.com', self.user.email)\n\n def test_user_generator(self):\n pass\n\n\nclass UserProfileTest(unittest.TestCase):\n '''User profile test'''\n def test_post_save_signal(self):\n # Disconnect post_save signal from user model (for test purposing only)\n post_save.disconnect(create_user_profile, sender=User)\n sender = User\n user = UserFactory.create()\n create_user_profile(sender, user, True)\n cnt = Users.objects.all().count()\n self.assertEqual(1, cnt)\n","new_contents":"# coding: utf-8\n\nimport unittest\nfrom test.factories import UserFactory\n\nfrom django.contrib.auth.models import User\nfrom django.db.models.signals import post_save\n\nfrom users.models import create_new_user, Users\n\n\nclass UserTest(unittest.TestCase):\n '''User-specific tests'''\n def setUp(self):\n self.user = UserFactory.build()\n\n def tearDown(self):\n self.user = None\n\n def test_user(self):\n self.assertNotEqual(None, self.user)\n self.assertEqual('Boy', self.user.first_name)\n self.assertEqual('Factory', self.user.last_name)\n self.assertEqual('boy_factory@example.com', self.user.email)\n\n def test_user_generator(self):\n pass\n\n def test_create_new_user(self):\n self.assertEqual(0, User.objects.all().count())\n create_new_user(\n first_name = self.user.first_name,\n last_name = self.user.last_name,\n email = self.user.email,\n password='123'\n )\n self.assertEqual(1, User.objects.all().count())\n u = User.objects.get(email=self.user.email)\n self.assertEqual(u.first_name, self.user.first_name)\n self.assertEqual(u.last_name, self.user.last_name)\n self.assertTrue(u.check_password('123'))\n self.assertFalse(u.is_staff)\n self.assertFalse(u.is_active)\n\n","subject":"Test for create user in model. Remove test profile creation","message":"Test for create user in model. Remove test profile creation\n","lang":"Python","license":"mit","repos":"sarutobi\/ritmserdtsa,sarutobi\/Rynda,sarutobi\/Rynda,sarutobi\/flowofkindness,sarutobi\/ritmserdtsa,sarutobi\/ritmserdtsa,sarutobi\/flowofkindness,sarutobi\/Rynda,sarutobi\/Rynda,sarutobi\/flowofkindness,sarutobi\/flowofkindness,sarutobi\/ritmserdtsa"} {"commit":"fc01acc869969e5c0666de1065f149b3caec851d","old_file":"core\/wait_ssh_ready.py","new_file":"core\/wait_ssh_ready.py","old_contents":"from __future__ import print_function\n\nimport time\nimport sys\nimport socket\nimport logging\n\n\ndef wait_ssh_ready(host, tries=40, delay=3, port=22):\n # Wait until the SSH is actually up\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n logging.info('Waiting for SSH at %s to be ready to connect' % host, end='')\n sys.stdout.flush()\n \n for _ in xrange(tries):\n try:\n s.connect((host, port))\n assert s.recv(3) == 'SSH'\n except KeyboardInterrupt:\n logging.warn('User stopped the loop.')\n break\n except socket.error:\n time.sleep(delay)\n print('.', end='')\n sys.stdout.flush()\n except AssertionError:\n time.sleep(delay)\n print('!', end='')\n sys.stdout.flush()\n else:\n print() # A new line\n logging.info('SSH is ready to connect')\n return True\n else:\n waited = tries * delay\n logging.error('SSH is not available after %s seconds.' % waited)\n return False","new_contents":"from __future__ import print_function\n\nimport time\nimport sys\nimport socket\nimport logging\n\n\ndef wait_ssh_ready(host, tries=40, delay=3, port=22):\n # Wait until the SSH is actually up\n s = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n print('Waiting for SSH at %s to be ready to connect' % host, end='')\n sys.stdout.flush()\n \n for _ in xrange(tries):\n try:\n s.connect((host, port))\n assert s.recv(3) == 'SSH'\n except KeyboardInterrupt:\n logging.warn('User stopped the loop.')\n break\n except socket.error:\n time.sleep(delay)\n print('.', end='')\n sys.stdout.flush()\n except AssertionError:\n time.sleep(delay)\n print('!', end='')\n sys.stdout.flush()\n else:\n print() # A new line\n logging.info('SSH is ready to connect')\n return True\n else:\n waited = tries * delay\n logging.error('SSH is not available after %s seconds.' % waited)\n return False","subject":"Fix incorrect call to logging module","message":"Fix incorrect call to logging module\n","lang":"Python","license":"agpl-3.0","repos":"andresriancho\/nimbostratus-target"} {"commit":"afbcda104f9903bda2d82e34a6fdc63b6e2b52a9","old_file":"mbio\/Application\/__init__.py","new_file":"mbio\/Application\/__init__.py","old_contents":"__author__ = 'Wenzhi Mao'\n\n__all__ = []\n\n\ndef _Startup():\n '''Get _path__ and compile files.'''\n from os import path\n from mbio import _ABSpath\n global _path__\n _path__ = _ABSpath()\n\n\n_Startup()\n\n\nfrom . import sort\nfrom .sort import *\n__all__.extend(sort.__all__)\n\nfrom . import cluster\n# from .cluster import *\n# __all__.extend(cluster.__all__)\n\nfrom . import job_organization\nfrom .job_organization import *\n__all__.extend(job_organization.__all__)\n\nfrom . import math\n# from .math import *\n# __all__.extend(math.__all__)\n","new_contents":"__author__ = 'Wenzhi Mao'\n\n__all__ = []\n\n\ndef _Startup():\n '''Get _path__ and compile files.'''\n from os import path\n from mbio import _ABSpath\n global _path__\n _path__ = _ABSpath()\n\n\n_Startup()\n\n\nfrom . import sort\nfrom .sort import *\n__all__.extend(sort.__all__)\n\n# from . import cluster\n# from .cluster import *\n# __all__.extend(cluster.__all__)\n\n# from . import job_organization\n# from .job_organization import *\n# __all__.extend(job_organization.__all__)\n\nfrom . import math\n# from .math import *\n# __all__.extend(math.__all__)\n","subject":"Remove the cluster and job_organization.","message":"Remove the cluster and job_organization.\n","lang":"Python","license":"mit","repos":"wzmao\/mbio,wzmao\/mbio,wzmao\/mbio"} {"commit":"020e8db7ed28c3c6e6968d2d107b23e1fa8eb284","old_file":"pcapfile\/test\/__main__.py","new_file":"pcapfile\/test\/__main__.py","old_contents":"#!\/usr\/bin\/env python\n\"\"\"\nThis is the front end to the pcapfile test SUITE.\n\"\"\"\n\nimport unittest\n\n\nfrom pcapfile.test.linklayer_test import TestCase as LinklayerTest\nfrom pcapfile.test.savefile_test import TestCase as SavefileTest\nfrom pcapfile.test.protocols_linklayer_ethernet import TestCase as EthernetTest\nfrom pcapfile.test.protocols_linklayer_wifi import TestCase as WifiTest\nfrom pcapfile.test.protocols_network_ip import TestCase as IpTest\nfrom pcapfile.test.protocols_transport_tcp import TestCase as TcpTest\n\nif __name__ == '__main__':\n TEST_CLASSES = [SavefileTest, LinklayerTest, EthernetTest, WifiTest, IpTest, TcpTest]\n SUITE = unittest.TestSuite()\n LOADER = unittest.TestLoader()\n for test_class in TEST_CLASSES:\n SUITE.addTests(LOADER.loadTestsFromTestCase(test_class))\n unittest.TextTestRunner(verbosity=2).run(SUITE)\n","new_contents":"#!\/usr\/bin\/env python\n\"\"\"\nThis is the front end to the pcapfile test SUITE.\n\"\"\"\n\nimport unittest\nimport sys\n\n\nfrom pcapfile.test.linklayer_test import TestCase as LinklayerTest\nfrom pcapfile.test.savefile_test import TestCase as SavefileTest\nfrom pcapfile.test.protocols_linklayer_ethernet import TestCase as EthernetTest\nfrom pcapfile.test.protocols_linklayer_wifi import TestCase as WifiTest\nfrom pcapfile.test.protocols_network_ip import TestCase as IpTest\nfrom pcapfile.test.protocols_transport_tcp import TestCase as TcpTest\n\nif __name__ == '__main__':\n TEST_CLASSES = [SavefileTest, LinklayerTest, EthernetTest, WifiTest, IpTest, TcpTest]\n SUITE = unittest.TestSuite()\n LOADER = unittest.TestLoader()\n for test_class in TEST_CLASSES:\n SUITE.addTests(LOADER.loadTestsFromTestCase(test_class))\n result = unittest.TextTestRunner(verbosity=2).run(SUITE)\n if not result.wasSuccessful():\n sys.exit(1)\n","subject":"Return -1 when tests fail","message":"Return -1 when tests fail\n","lang":"Python","license":"isc","repos":"kisom\/pypcapfile"} {"commit":"d474edcdbe1d9966ad09609b87d119c60c2a38d4","old_file":"datapusher\/main.py","new_file":"datapusher\/main.py","old_contents":"import os\nimport six\nimport ckanserviceprovider.web as web\n\nfrom . import jobs\n\n# check whether jobs have been imported properly\nassert(jobs.push_to_datastore)\n\n\ndef serve():\n web.init()\n web.app.run(web.app.config.get('HOST'), web.app.config.get('PORT'))\n\n\ndef serve_test():\n web.init()\n return web.app.test_client()\n\n\ndef main():\n import argparse\n\n argparser = argparse.ArgumentParser(\n description='Service that allows automatic migration of data to the CKAN DataStore',\n epilog='''\"He reached out and pressed an invitingly large red button on a nearby panel.\n The panel lit up with the words Please do not press this button again.\"''')\n if six.PY3:\n argparser.add_argument('config', metavar='CONFIG', type=argparse.FileType('r'),\n help='configuration file')\n if six.PY2:\n argparser.add_argument('config', metavar='CONFIG', type=file,\n help='configuration file')\n args = argparser.parse_args()\n\n os.environ['JOB_CONFIG'] = os.path.abspath(args.config.name)\n serve()\n\nif __name__ == '__main__':\n main()\n","new_contents":"import os\nimport six\nimport ckanserviceprovider.web as web\n\nfrom datapusher import jobs\n\n# check whether jobs have been imported properly\nassert(jobs.push_to_datastore)\n\n\ndef serve():\n web.init()\n web.app.run(web.app.config.get('HOST'), web.app.config.get('PORT'))\n\n\ndef serve_test():\n web.init()\n return web.app.test_client()\n\n\ndef main():\n import argparse\n\n argparser = argparse.ArgumentParser(\n description='Service that allows automatic migration of data to the CKAN DataStore',\n epilog='''\"He reached out and pressed an invitingly large red button on a nearby panel.\n The panel lit up with the words Please do not press this button again.\"''')\n if six.PY3:\n argparser.add_argument('config', metavar='CONFIG', type=argparse.FileType('r'),\n help='configuration file')\n if six.PY2:\n argparser.add_argument('config', metavar='CONFIG', type=file,\n help='configuration file')\n args = argparser.parse_args()\n\n os.environ['JOB_CONFIG'] = os.path.abspath(args.config.name)\n serve()\n\nif __name__ == '__main__':\n main()\n","subject":"Fix Import Error for relative Import","message":"[x]: Fix Import Error for relative Import\n","lang":"Python","license":"agpl-3.0","repos":"ckan\/datapusher"} {"commit":"946a2bcd57ac33cca0f48d29350a8f75b2fee2cf","old_file":"sparqllib\/tests\/test_formatter.py","new_file":"sparqllib\/tests\/test_formatter.py","old_contents":"\nimport unittest\nimport sparqllib\n\nclass TestBasicFormatter(unittest.TestCase):\n def setUp(self):\n self.formatter = sparqllib.formatter.BasicFormatter()\n\n def test_newlines(self):\n self.assertEqual(self.formatter.format(\"{}\"), \"{\\n}\")\n\n def test_indentation(self):\n self.assertEqual(self.formatter.format(\"{test text}\"), \"{\\n test text\\n}\")\n self.assertEqual(self.formatter.format(\"{test\\ntext}\"), \"{\\n test\\n text\\n}\")\n self.assertEqual(self.formatter.format(\"{{text}}\"), \"{\\n {\\n text\\n }\\n}\")\n\n def test_trim_whitespace(self):\n self.assertEqual(self.formatter.format(\"text \\n\"), \"text\\n\")\n\n def test_remove_duplicate_newlines(self):\n self.assertEqual(self.formatter.format(\"\\n\\n\"), \"\\n\")\n self.assertEqual(self.formatter.format(\"\\n\"), \"\\n\")\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"\nimport unittest\nimport sparqllib\n\nclass TestBasicFormatter(unittest.TestCase):\n def setUp(self):\n self.formatter = sparqllib.formatter.BasicFormatter()\n\n def test_newlines(self):\n self.assertEqual(self.formatter.format(\"{}\"), \"{\\n}\")\n self.assertEqual(self.formatter.format(\"{\\n}\"), \"{\\n}\")\n\n def test_indentation(self):\n self.assertEqual(self.formatter.format(\"{test text}\"), \"{\\n test text\\n}\")\n self.assertEqual(self.formatter.format(\"{test\\ntext}\"), \"{\\n test\\n text\\n}\")\n self.assertEqual(self.formatter.format(\"{{text}}\"), \"{\\n {\\n text\\n }\\n}\")\n\n def test_trim_whitespace(self):\n self.assertEqual(self.formatter.format(\"text \\n\"), \"text\\n\")\n\n def test_remove_duplicate_newlines(self):\n self.assertEqual(self.formatter.format(\"\\n\\n\"), \"\\n\")\n self.assertEqual(self.formatter.format(\"\\n\"), \"\\n\")\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Add test to verify single newline is not stripped","message":"Add test to verify single newline is not stripped\n","lang":"Python","license":"mit","repos":"ALSchwalm\/sparqllib"} {"commit":"a7083c3c70142f744ace0055c537d9217ed9cbfe","old_file":"paypal\/base.py","new_file":"paypal\/base.py","old_contents":"import urlparse\n\nfrom django.db import models\n\n\nclass ResponseModel(models.Model):\n\n # Debug information\n raw_request = models.TextField(max_length=512)\n raw_response = models.TextField(max_length=512)\n\n response_time = models.FloatField(help_text=\"Response time in milliseconds\")\n\n date_created = models.DateTimeField(auto_now_add=True)\n\n class Meta:\n abstract = True\n ordering = ('-date_created',)\n app_label = 'paypal'\n\n def request(self):\n request_params = urlparse.parse_qs(self.raw_request)\n return self._as_table(request_params)\n request.allow_tags = True\n\n def response(self):\n return self._as_table(self.context)\n response.allow_tags = True\n\n def _as_table(self, params):\n rows = []\n for k, v in sorted(params.items()):\n rows.append('<tr><th>%s<\/th><td>%s<\/td><\/tr>' % (k, v[0]))\n return '<table>%s<\/table>' % ''.join(rows)\n\n @property\n def context(self):\n return urlparse.parse_qs(self.raw_response)\n\n def value(self, key):\n ctx = self.context\n return ctx[key][0] if key in ctx else None\n","new_contents":"import urlparse\n\nfrom django.db import models\n\n\nclass ResponseModel(models.Model):\n\n # Debug information\n raw_request = models.TextField(max_length=512)\n raw_response = models.TextField(max_length=512)\n\n response_time = models.FloatField(help_text=\"Response time in milliseconds\")\n\n date_created = models.DateTimeField(auto_now_add=True)\n\n class Meta:\n abstract = True\n ordering = ('-date_created',)\n app_label = 'paypal'\n\n def request(self):\n request_params = urlparse.parse_qs(self.raw_request)\n return self._as_table(request_params)\n request.allow_tags = True\n\n def response(self):\n return self._as_table(self.context)\n response.allow_tags = True\n\n def _as_table(self, params):\n rows = []\n for k, v in sorted(params.items()):\n rows.append('<tr><th>%s<\/th><td>%s<\/td><\/tr>' % (k, v[0]))\n return '<table>%s<\/table>' % ''.join(rows)\n\n @property\n def context(self):\n return urlparse.parse_qs(self.raw_response)\n\n def value(self, key):\n ctx = self.context\n return ctx[key][0].decode('utf8') if key in ctx else None\n","subject":"Fix a bad issue when PAYPAL returning utf8 encoded chars","message":"Fix a bad issue when PAYPAL returning utf8 encoded chars\n","lang":"Python","license":"bsd-3-clause","repos":"bharling\/django-oscar-worldpay,embedded1\/django-oscar-paypal,FedeDR\/django-oscar-paypal,django-oscar\/django-oscar-paypal,evonove\/django-oscar-paypal,st8st8\/django-oscar-paypal,nfletton\/django-oscar-paypal,britco\/django-oscar-paypal,britco\/django-oscar-paypal,st8st8\/django-oscar-paypal,ZachGoldberg\/django-oscar-paypal,evonove\/django-oscar-paypal,ZachGoldberg\/django-oscar-paypal,enodyt\/django-oscar-paypal,evonove\/django-oscar-paypal,embedded1\/django-oscar-paypal,vintasoftware\/django-oscar-paypal,bharling\/django-oscar-worldpay,st8st8\/django-oscar-paypal,lpakula\/django-oscar-paypal,lpakula\/django-oscar-paypal,enodyt\/django-oscar-paypal,nfletton\/django-oscar-paypal,britco\/django-oscar-paypal,FedeDR\/django-oscar-paypal,embedded1\/django-oscar-paypal,lpakula\/django-oscar-paypal,bharling\/django-oscar-worldpay,ZachGoldberg\/django-oscar-paypal,FedeDR\/django-oscar-paypal,django-oscar\/django-oscar-paypal,django-oscar\/django-oscar-paypal,vintasoftware\/django-oscar-paypal,enodyt\/django-oscar-paypal,phedoreanu\/django-oscar-paypal,phedoreanu\/django-oscar-paypal,phedoreanu\/django-oscar-paypal,nfletton\/django-oscar-paypal,bharling\/django-oscar-worldpay,vintasoftware\/django-oscar-paypal"} {"commit":"233d52247d89bb39ccc9ada3a591296baae9cff5","old_file":"notification\/backends\/web.py","new_file":"notification\/backends\/web.py","old_contents":"from notification.backends.base import NotificationBackend\n\nclass WebBackend(NotificationBackend):\n slug = u'web'\n display_name = u'E-mail'\n formats = ['short.txt', 'full.txt']\n\n def send(self, sender, recipient, notice_type, context, on_site=False,\n *args, **kwargs):\n \"\"\"Always \"sends\" (i.e. stores to the database), setting on_site\n accordingly.\n \"\"\"\n # TODO can't do this at the top or we get circular imports\n from notification.models import Notice\n Notice.objects.create(recipient=recipient,\n message=self.format_message(notice_type.label,\n 'notice.html', context),\n notice_type=notice_type,\n on_site=on_site,\n sender=sender)\n return True\n","new_contents":"from notification.backends.base import NotificationBackend\n\nclass WebBackend(NotificationBackend):\n slug = u'web'\n display_name = u'Web'\n formats = ['short.txt', 'full.txt']\n\n def send(self, sender, recipient, notice_type, context, on_site=False,\n *args, **kwargs):\n \"\"\"Always \"sends\" (i.e. stores to the database), setting on_site\n accordingly.\n \"\"\"\n # TODO can't do this at the top or we get circular imports\n from notification.models import Notice\n Notice.objects.create(recipient=recipient,\n message=self.format_message(notice_type.label,\n 'notice.html', context),\n notice_type=notice_type,\n on_site=on_site,\n sender=sender)\n return True\n","subject":"Use correct slug for Web backend.","message":"Use correct slug for Web backend.\n","lang":"Python","license":"mit","repos":"theatlantic\/django-notification,theatlantic\/django-notification"} {"commit":"39406267d31ca428dc73d721ccc19285ff7599bd","old_file":"lit\/Quit\/expect_exit_code.py","new_file":"lit\/Quit\/expect_exit_code.py","old_contents":"#!\/usr\/bin\/env python2\n\nimport subprocess\nimport sys\n\nargs = sys.argv\n\nexpected_exit_code = args[1]\n\nargs = args[2:]\nprint(\"Running \" + (\" \".join(args)))\nreal_exit_code = subprocess.call(args)\n\nif str(real_exit_code) != expected_exit_code:\n print(\"Got exit code %d but expected %s\" % (real_exit_code, expected_exit_code))\n exit(1)\n","new_contents":"#!\/usr\/bin\/env python\n\nimport subprocess\nimport sys\n\nargs = sys.argv\n\nexpected_exit_code = args[1]\n\nargs = args[2:]\nprint(\"Running \" + (\" \".join(args)))\nreal_exit_code = subprocess.call(args)\n\nif str(real_exit_code) != expected_exit_code:\n print(\"Got exit code %d but expected %s\" % (real_exit_code, expected_exit_code))\n exit(1)\n","subject":"Update shebang python2 -> python","message":"[lldb] Update shebang python2 -> python\n\ngit-svn-id: 4c4cc70b1ef44ba2b7963015e681894188cea27e@352259 91177308-0d34-0410-b5e6-96231b3b80d8\n","lang":"Python","license":"apache-2.0","repos":"apple\/swift-lldb,llvm-mirror\/lldb,llvm-mirror\/lldb,apple\/swift-lldb,apple\/swift-lldb,llvm-mirror\/lldb,apple\/swift-lldb,apple\/swift-lldb,apple\/swift-lldb,llvm-mirror\/lldb,llvm-mirror\/lldb"} {"commit":"7aa84fbcc7a3af57ef62c29008fac4036d2d28af","old_file":"django_afip\/migrations\/0021_drop_batches.py","new_file":"django_afip\/migrations\/0021_drop_batches.py","old_contents":"# -*- coding: utf-8 -*-\n# Generated by Django 1.11.2 on 2017-07-02 23:19\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('afip', '0020_backfill_receiptvalidation__processed_date'),\n ]\n\n operations = [\n migrations.RemoveField(\n model_name='receipt',\n name='batch',\n ),\n migrations.RemoveField(\n model_name='receiptvalidation',\n name='validation',\n ),\n migrations.AlterField(\n model_name='receiptvalidation',\n name='processed_date',\n field=models.DateTimeField(verbose_name='processed date'),\n ),\n migrations.DeleteModel(\n name='ReceiptBatch',\n ),\n migrations.DeleteModel(\n name='Validation',\n ),\n ]\n","new_contents":"# -*- coding: utf-8 -*-\n# Generated by Django 1.11.2 on 2017-07-02 23:19\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('afip', '0020_backfill_receiptvalidation__processed_date'),\n ]\n\n operations = [\n migrations.RemoveField(\n model_name='receipt',\n name='batch',\n ),\n migrations.RemoveField(\n model_name='receiptvalidation',\n name='validation',\n ),\n migrations.AlterField(\n model_name='receiptvalidation',\n name='processed_date',\n field=models.DateTimeField(verbose_name='processed date'),\n ),\n migrations.DeleteModel(\n name='Validation',\n ),\n migrations.DeleteModel(\n name='ReceiptBatch',\n ),\n ]\n","subject":"Tweak a migration to run on non-transactional DBs","message":"Tweak a migration to run on non-transactional DBs\n\nA single migration failed to run on databases with no support for\ntransactions because those require explicit ordering of commands that's\ngenerally implicit on modern relational DBs.\n\nSwitch the order of those queries to prevent that crash.\n\nFixes #27\n","lang":"Python","license":"isc","repos":"hobarrera\/django-afip,hobarrera\/django-afip"} {"commit":"34b19dcdf2d86ecdbca5fb5b5282f64b2f7e5153","old_file":"threebot_worker\/__init__.py","new_file":"threebot_worker\/__init__.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n__version__ = '0.1.6'\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n__version__ = '0.1.8'\n","subject":"Bump version: 0.1.7 → 0.1.8","message":"Bump version: 0.1.7 → 0.1.8\n","lang":"Python","license":"bsd-3-clause","repos":"3bot\/3bot-worker"} {"commit":"1ff616fe4f6ff0ff295eeeaa4a817851df750e51","old_file":"openslides\/utils\/validate.py","new_file":"openslides\/utils\/validate.py","old_contents":"import bleach\n\n\nallowed_tags = [\n \"a\",\n \"img\", # links and images\n \"br\",\n \"p\",\n \"span\",\n \"blockquote\", # text layout\n \"strike\",\n \"strong\",\n \"u\",\n \"em\",\n \"sup\",\n \"sub\",\n \"pre\", # text formatting\n \"h1\",\n \"h2\",\n \"h3\",\n \"h4\",\n \"h5\",\n \"h6\", # headings\n \"ol\",\n \"ul\",\n \"li\", # lists\n \"table\",\n \"caption\",\n \"thead\",\n \"tbody\",\n \"th\",\n \"tr\",\n \"td\", # tables\n]\nallowed_attributes = {\n \"*\": [\"class\", \"style\"],\n \"img\": [\"alt\", \"src\", \"title\"],\n \"a\": [\"href\", \"title\"],\n \"th\": [\"scope\"],\n \"ol\": [\"start\"],\n}\nallowed_styles = [\n \"color\",\n \"background-color\",\n \"height\",\n \"width\",\n \"text-align\",\n \"float\",\n \"padding\",\n \"text-decoration\",\n]\n\n\ndef validate_html(html: str) -> str:\n \"\"\"\n This method takes a string and escapes all non-whitelisted html entries.\n Every field of a model that is loaded trusted in the DOM should be validated.\n During copy and paste from Word maybe some tabs are spread over the html. Remove them.\n \"\"\"\n html = html.replace(\"\\t\", \"\")\n return bleach.clean(\n html, tags=allowed_tags, attributes=allowed_attributes, styles=allowed_styles\n )\n","new_contents":"import bleach\n\n\nallowed_tags = [\n \"a\",\n \"img\", # links and images\n \"br\",\n \"p\",\n \"span\",\n \"blockquote\", # text layout\n \"strike\",\n \"del\",\n \"ins\",\n \"strong\",\n \"u\",\n \"em\",\n \"sup\",\n \"sub\",\n \"pre\", # text formatting\n \"h1\",\n \"h2\",\n \"h3\",\n \"h4\",\n \"h5\",\n \"h6\", # headings\n \"ol\",\n \"ul\",\n \"li\", # lists\n \"table\",\n \"caption\",\n \"thead\",\n \"tbody\",\n \"th\",\n \"tr\",\n \"td\", # tables\n]\nallowed_attributes = {\n \"*\": [\"class\", \"style\"],\n \"img\": [\"alt\", \"src\", \"title\"],\n \"a\": [\"href\", \"title\"],\n \"th\": [\"scope\"],\n \"ol\": [\"start\"],\n}\nallowed_styles = [\n \"color\",\n \"background-color\",\n \"height\",\n \"width\",\n \"text-align\",\n \"float\",\n \"padding\",\n \"text-decoration\",\n]\n\n\ndef validate_html(html: str) -> str:\n \"\"\"\n This method takes a string and escapes all non-whitelisted html entries.\n Every field of a model that is loaded trusted in the DOM should be validated.\n During copy and paste from Word maybe some tabs are spread over the html. Remove them.\n \"\"\"\n html = html.replace(\"\\t\", \"\")\n return bleach.clean(\n html, tags=allowed_tags, attributes=allowed_attributes, styles=allowed_styles\n )\n","subject":"Allow <del> and <ins> html tags.","message":"Allow <del> and <ins> html tags.\n","lang":"Python","license":"mit","repos":"tsiegleauq\/OpenSlides,ostcar\/OpenSlides,FinnStutzenstein\/OpenSlides,normanjaeckel\/OpenSlides,jwinzer\/OpenSlides,OpenSlides\/OpenSlides,tsiegleauq\/OpenSlides,jwinzer\/OpenSlides,FinnStutzenstein\/OpenSlides,CatoTH\/OpenSlides,CatoTH\/OpenSlides,OpenSlides\/OpenSlides,tsiegleauq\/OpenSlides,CatoTH\/OpenSlides,normanjaeckel\/OpenSlides,ostcar\/OpenSlides,ostcar\/OpenSlides,FinnStutzenstein\/OpenSlides,normanjaeckel\/OpenSlides,jwinzer\/OpenSlides,normanjaeckel\/OpenSlides,FinnStutzenstein\/OpenSlides,jwinzer\/OpenSlides,CatoTH\/OpenSlides,jwinzer\/OpenSlides"} {"commit":"e0ebd4cb41d3ed9168e819f7017dd98c2fbb599a","old_file":"insertion_sort.py","new_file":"insertion_sort.py","old_contents":"def insertion_sort(un_list):\n for idx in range(1, len(un_list)):\n current = un_list[idx]\n position = idx\n\n while position > 0 and un_list[position-1] > current:\n un_list[position] = un_list[position-1]\n position = position - 1\n\n un_list[position] = current\n\nif __name__ == '__main__':\n BEST_CASE = range(1000)\n WORST_CASE = BEST_CASE[::-1]\n\n from timeit import Timer\n\n best = Timer(\n 'insertion_sort({})'.format(BEST_CASE),\n 'from __main__ import BEST_CASE, insertion_sort').timeit(1000)\n\n worst = Timer(\n 'insertion_sort({})'.format(WORST_CASE),\n 'from __main__ import WORST_CASE, insertion_sort').timeit(1000)\n\n print(\"\"\"Best case represented as a list that is already sorted\\n\n Worst case represented as a list that is absolute reverse of sorted\"\"\")\n print('Best Case: {}'.format(best))\n print('Worst Case: {}'.format(worst))\n","new_contents":"def insertion_sort(un_list):\n if type(un_list) is not list:\n return \"You must pass a valid list as argument. Do it.\"\n\n for idx in range(1, len(un_list)):\n current = un_list[idx]\n position = idx\n\n while position > 0 and un_list[position-1] > current:\n un_list[position] = un_list[position-1]\n position = position - 1\n\n un_list[position] = current\n\nif __name__ == '__main__':\n BEST_CASE = range(1000)\n WORST_CASE = BEST_CASE[::-1]\n\n from timeit import Timer\n\n best = Timer(\n 'insertion_sort({})'.format(BEST_CASE),\n 'from __main__ import BEST_CASE, insertion_sort').timeit(1000)\n\n worst = Timer(\n 'insertion_sort({})'.format(WORST_CASE),\n 'from __main__ import WORST_CASE, insertion_sort').timeit(1000)\n\n print(\"\"\"Best case represented as a list that is already sorted\\n\n Worst case represented as a list that is absolute reverse of sorted\"\"\")\n print('Best Case: {}'.format(best))\n print('Worst Case: {}'.format(worst))\n","subject":"Update insertion sort with list validation","message":"Update insertion sort with list validation\n","lang":"Python","license":"mit","repos":"jonathanstallings\/data-structures"} {"commit":"a5697ddd595b929ef7261d62fd333c2cd2f56dd0","old_file":"plots\/views.py","new_file":"plots\/views.py","old_contents":"# Create your views here.\nimport json\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response\nfrom django.utils import simplejson\nfrom django.http import HttpResponse, Http404\nfrom .models import BenchmarkLogs, MachineInfo\nimport data\n\n\ndef rawdata(request, plotname):\n \"\"\"\n Based on the ajax request from the template, it calls data.py to fetch the appropriate data.\n It returns a JSON dump of the dictionary that is returned from data.py\n \"\"\"\n if request.is_ajax():\n try:\n data_dict = {}\n data_dict = getattr(data, plotname).__call__()\n print data_dict\n return HttpResponse(simplejson.dumps(data_dict), content_type=\"application\/json\")\n except AttributeError:\n raise 404\n raise 404\n\n\ndef draw(request, plotname):\n \"\"\"\n The draw view is responsible for drawing the chart. It renders a template chart.html which sends ajax request for the JSON data.\n It also provides the template the name of the plot to draw via the <name_dict> dictionary.\n \"\"\"\n name_dict = {'plotname': plotname}\n return render_to_response(\"flot-chart.html\", name_dict, context_instance=RequestContext(request))\n","new_contents":"# Create your views here.\nimport json\nfrom django.template import RequestContext\nfrom django.shortcuts import render_to_response\nfrom django.utils import simplejson\nfrom django.http import HttpResponse, Http404\nfrom .models import BenchmarkLogs, MachineInfo\nimport data\n\n\ndef rawdata(request, plotname):\n \"\"\"\n Based on the ajax request from the template, it calls data.py to fetch the appropriate data.\n It returns a JSON dump of the dictionary that is returned from data.py\n \"\"\"\n if request.is_ajax():\n try:\n data_dict = {}\n data_dict = getattr(data, plotname).__call__()\n print data_dict\n return HttpResponse(simplejson.dumps(data_dict), content_type=\"application\/json\")\n except AttributeError:\n raise Http404\n raise Http404\n\n\ndef draw(request, plotname):\n \"\"\"\n The draw view is responsible for drawing the chart. It renders a template chart.html which sends ajax request for the JSON data.\n It also provides the template the name of the plot to draw via the <name_dict> dictionary.\n \"\"\"\n name_dict = {'plotname': plotname}\n return render_to_response(\"flot-chart.html\", name_dict, context_instance=RequestContext(request))\n","subject":"Raise Http404 instead of 404","message":"Raise Http404 instead of 404\n","lang":"Python","license":"bsd-2-clause","repos":"ankeshanand\/benchmark,ankeshanand\/benchmark,ankeshanand\/benchmark,ankeshanand\/benchmark"} {"commit":"e394467dcfea0a5141441112b3dce14ad7832328","old_file":"spyder_terminal\/__init__.py","new_file":"spyder_terminal\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n# -----------------------------------------------------------------------------\n# Copyright (c) Spyder Project Contributors\n#\n# Licensed under the terms of the MIT License\n# (see LICENSE.txt for details)\n# -----------------------------------------------------------------------------\n\"\"\"Spyder Terminal Plugin.\"\"\"\n\nfrom qtpy import PYQT5\n\nif PYQT5:\n from .terminalplugin import TerminalPlugin as PLUGIN_CLASS\n\nVERSION_INFO = (0, 2, 'dev0')\n__version__ = '.'.join(map(str, VERSION_INFO))\n","new_contents":"# -*- coding: utf-8 -*-\n# -----------------------------------------------------------------------------\n# Copyright (c) Spyder Project Contributors\n#\n# Licensed under the terms of the MIT License\n# (see LICENSE.txt for details)\n# -----------------------------------------------------------------------------\n\"\"\"Spyder Terminal Plugin.\"\"\"\n\nfrom qtpy import PYQT5\n\nif PYQT5:\n from .terminalplugin import TerminalPlugin as PLUGIN_CLASS\n\nVERSION_INFO = (0, 2, 0, 'dev0')\n__version__ = '.'.join(map(str, VERSION_INFO))\n","subject":"Bump version number to 0.2.0.dev0","message":"Bump version number to 0.2.0.dev0\n","lang":"Python","license":"mit","repos":"andfoy\/spyder-terminal,spyder-ide\/spyder-terminal,spyder-ide\/spyder-terminal,andfoy\/spyder-terminal,spyder-ide\/spyder-terminal,spyder-ide\/spyder-terminal,andfoy\/spyder-terminal"} {"commit":"5a2212746bfabcfd64cf27846770b35f767d57a6","old_file":"polls\/views.py","new_file":"polls\/views.py","old_contents":"from django.shortcuts import render\nfrom django.core.urlresolvers import reverse_lazy\n\nfrom singleurlcrud.views import CRUDView\nfrom .models import *\n\n# Create your views here.\nclass AuthorCRUDView(CRUDView):\n model = Author\n list_display = ('name',)\n\nclass QuestionCRUDView(CRUDView):\n model = Question\n list_display = ('question_text', 'pub_date', 'author')\n related_field_crud_urls = {\n 'author': reverse_lazy(\"polls:authors\")\n }\n\n def get_actions(self):\n self.related_field_crud_urls = {\n 'author': reverse_lazy('polls:authors') +\"?o=add\",\n }\n return [\n ('Delete', self.delete_multiple_items)\n ]\n\n def delete_multiple_items(self, request, items):\n pass\n\n class VoteItemAction(object):\n title = 'Vote'\n key = 'vote1'\n css = 'glyphicon glyphicon-envelope'\n\n def doAction(self, item):\n import logging\n logging.getLogger('general').info(\"VoteItemAction invoked!\")\n pass\n\n def get_item_actions(self):\n return [self.VoteItemAction()]\n\n","new_contents":"from django.shortcuts import render\nfrom django.core.urlresolvers import reverse_lazy\n\nfrom singleurlcrud.views import CRUDView\nfrom .models import *\n\n# Create your views here.\nclass AuthorCRUDView(CRUDView):\n model = Author\n list_display = ('name',)\n\nclass QuestionCRUDView(CRUDView):\n model = Question\n list_display = ('question_text', 'pub_date', 'author')\n related_field_crud_urls = {\n 'author': reverse_lazy(\"polls:authors\")\n }\n\n def get_actions(self):\n self.related_field_crud_urls = {\n 'author': reverse_lazy('polls:authors') +\"?o=add\",\n }\n return [\n ('Delete', self.delete_multiple_items)\n ]\n\n def delete_multiple_items(self, request, items):\n Question.objects.filter(pk__in=items).delete()\n\n class VoteItemAction(object):\n title = 'Vote'\n key = 'vote1'\n css = 'glyphicon glyphicon-envelope'\n\n def doAction(self, item):\n import logging\n logging.getLogger('general').info(\"VoteItemAction invoked!\")\n pass\n\n def get_item_actions(self):\n return [self.VoteItemAction()]\n\n","subject":"Implement 'Delete' action for polls sample app","message":"Implement 'Delete' action for polls sample app\n","lang":"Python","license":"bsd-3-clause","repos":"harikvpy\/crud,harikvpy\/crud,harikvpy\/crud"} {"commit":"14123d5e3544ab9dbee813e26536e43cbfb9f783","old_file":"pycroscopy\/__version__.py","new_file":"pycroscopy\/__version__.py","old_contents":"version = '0.59.8'\ntime = '2018-04-18 08:12:59'\n","new_contents":"version = '0.60.0rc1'\ntime = '2018-04-18 08:12:59'\n","subject":"Mark as release candidate version","message":"Mark as release candidate version\n\n","lang":"Python","license":"mit","repos":"pycroscopy\/pycroscopy"} {"commit":"9851430922f9c14583c9eb17062629f6ea99c258","old_file":"turbustat\/tests\/test_vcs.py","new_file":"turbustat\/tests\/test_vcs.py","old_contents":"# Licensed under an MIT open source license - see LICENSE\n\n'''\nTest functions for VCS\n'''\n\nfrom unittest import TestCase\n\nimport numpy as np\nimport numpy.testing as npt\n\nfrom ..statistics import VCS, VCS_Distance\nfrom ._testing_data import \\\n dataset1, dataset2, computed_data, computed_distances\n\n\nclass testVCS(TestCase):\n\n def setUp(self):\n self.dataset1 = dataset1\n self.dataset2 = dataset2\n\n def test_VCS_method(self):\n self.tester = VCS(dataset1[\"cube\"]).run()\n\n npt.assert_allclose(self.tester.ps1D, computed_data['vcs_val'])\n\n def test_VCS_distance(self):\n self.tester_dist = \\\n VCS_Distance(dataset1[\"cube\"], dataset2[\"cube\"])\n self.tester_dist = self.tester_dist.distance_metric()\n\n npt.assert_almost_equal(self.tester_dist.distance,\n computed_distances['vcs_distance'])\n","new_contents":"# Licensed under an MIT open source license - see LICENSE\n\n'''\nTest functions for VCS\n'''\n\nimport pytest\n\nimport numpy as np\nimport numpy.testing as npt\n\nfrom ..statistics import VCS, VCS_Distance\nfrom ._testing_data import \\\n dataset1, dataset2, computed_data, computed_distances\n\n\ndef test_VCS_method():\n tester = VCS(dataset1[\"cube\"]).run()\n\n npt.assert_allclose(tester.ps1D, computed_data['vcs_val'])\n\n npt.assert_allclose(tester.slope, computed_data['vcs_slopes_val'])\n\n\ndef test_VCS_distance():\n tester_dist = \\\n VCS_Distance(dataset1[\"cube\"], dataset2[\"cube\"])\n tester_dist = tester_dist.distance_metric()\n\n npt.assert_almost_equal(tester_dist.distance,\n computed_distances['vcs_distance'])\n\n# Add tests for: VCS changing the spectral width, pixel and spectral units,\n","subject":"Reformat VCS tests; need updated unit test values!","message":"Reformat VCS tests; need updated unit test values!\n","lang":"Python","license":"mit","repos":"e-koch\/TurbuStat,Astroua\/TurbuStat"} {"commit":"75b221fa63b0f81b94ffbbe9f5cdc39a0adb848a","old_file":"dmrg101\/core\/braket.py","new_file":"dmrg101\/core\/braket.py","old_contents":"'''\nFile: braket.py\nAuthor: Ivan Gonzalez\nDescription: A function to implement quantum-mechanics brakets\n'''\nfrom numpy import inner\nfrom core.exceptions import DMRGException\n\ndef braket(bra, ket):\n \"\"\"Takes a bra and a ket and return their braket\n\n You use this function to calculate the quantum mechanical braket, i.e.\n the inner product in the wavefunction Hilbert space of two\n wavefunction.\n\n The wavefunction in the bra is hermitian conjugated by the braket\n function.\n\n Attributes:\n bra: a Wavefunction with the bra part of the braket.\n ket: a Wavefunction with the ket part of the braket.\n\n Returns:\n a double\/complex with value of the braket.\n \n Raises:\n DMRGException: if the wavefunction don't belong to the same\n\tHilbert space, i.e. they have a different number of elements.\n \"\"\"\n # use wf.as_matrix to access the matrix elements of wf\n if bra.as_matrix.shape() != ket.as_matrix.shape():\n\traise DMRGException(\"Wavefunctions in braket are not in the same\n\t \t\t Hilbert space\")\n\n hermitian_conjugated_bra=conjugate(bra.as_matrix).transpose()\n return inner(hermitian_conjugated_bra, ket.as_matrix)\n","new_contents":"'''\nFile: braket.py\nAuthor: Ivan Gonzalez\nDescription: A function to implement quantum-mechanics brakets\n'''\nfrom numpy import inner, conjugate\nfrom dmrg_exceptions import DMRGException\n\ndef braket(bra, ket):\n \"\"\"Takes a bra and a ket and return their braket\n\n You use this function to calculate the quantum mechanical braket, i.e.\n the inner product in the wavefunction Hilbert space of two\n wavefunction.\n\n The wavefunction in the bra is hermitian conjugated by the braket\n function.\n\n Parameters\n ----------\n bra: a Wavefunction with the bra part of the braket.\n ket: a Wavefunction with the ket part of the braket.\n\n Returns\n -------\n a double\/complex with value of the braket.\n \n Raises\n ------\n DMRGException: if the wavefunction don't belong to the same\n\t Hilbert space, i.e. they have a different number of elements.\n \"\"\"\n # use wf.as_matrix to access the matrix elements of wf\n if bra.as_matrix.shape() != ket.as_matrix.shape():\n\traise DMRGException(\"Wavefunctions in braket are not in the same\n\t \t\t Hilbert space\")\n\n hermitian_conjugated_bra=conjugate(bra.as_matrix).transpose()\n return inner(hermitian_conjugated_bra, ket.as_matrix)\n","subject":"Clean up comments, fixing imports.","message":"Clean up comments, fixing imports.\n","lang":"Python","license":"mit","repos":"iglpdc\/dmrg101"} {"commit":"b1bb9e86b51bf0d1c57fa10ac9b8297f0bc078db","old_file":"flow_workflow\/petri_net\/future_nets\/base.py","new_file":"flow_workflow\/petri_net\/future_nets\/base.py","old_contents":"from flow.petri_net import future\nfrom flow.petri_net import success_failure_net\n\n\n# XXX Maybe this turns into a historian mixin?\nclass GenomeNetBase(success_failure_net.SuccessFailureNet):\n def __init__(self, name, operation_id, parent_operation_id=None):\n success_failure_net.SuccessFailureNet.__init__(self, name=name)\n self.operation_id = operation_id\n self.parent_operation_id = parent_operation_id\n\n\n def historian_action(self, status, **kwargs):\n info = {\"id\": self.operation_id,\n \"name\": self.name,\n \"status\": status,\n \"parent_net_key\": None,\n \"parent_operation_id\": self.parent_operation_id}\n\n # XXX the name 'parallel_index' is suspicious\n optional_attrs = ['parent_net_key',\n 'peer_operation_id', 'parallel_index']\n for attr in optional_attrs:\n value = getattr(self, attr, None)\n if value is not None:\n info[attr] = value\n\n args = {\"children_info\": [info]}\n args.update(kwargs)\n\n return future.FutureAction(cls=WorkflowHistorianUpdateAction, args=args)\n","new_contents":"from flow.petri_net import future\nfrom flow.petri_net.success_failure_net import SuccessFailureNet\n\n\n# XXX Maybe this turns into a historian mixin?\nclass GenomeNetBase(SuccessFailureNet):\n \"\"\"\n Basically a success-failure net with operation_id and parent_operation_id and\n the ability to construct historian_actions\n \"\"\"\n def __init__(self, name, operation_id, parent_operation_id=None):\n SuccessFailureNet.__init__(self, name=name)\n self.operation_id = operation_id\n self.parent_operation_id = parent_operation_id\n\n\n def historian_action(self, status, **kwargs):\n info = {\"id\": self.operation_id,\n \"name\": self.name,\n \"status\": status,\n \"parent_net_key\": None,\n \"parent_operation_id\": self.parent_operation_id}\n\n # XXX the name 'parallel_index' is suspicious\n optional_attrs = ['parent_net_key',\n 'peer_operation_id', 'parallel_index']\n for attr in optional_attrs:\n value = getattr(self, attr, None)\n if value is not None:\n info[attr] = value\n\n args = {\"children_info\": [info]}\n args.update(kwargs)\n\n return future.FutureAction(cls=WorkflowHistorianUpdateAction, args=args)\n","subject":"Add comments and clean-up import of GenomeNetBase","message":"Add comments and clean-up import of GenomeNetBase\n","lang":"Python","license":"agpl-3.0","repos":"genome\/flow-workflow,genome\/flow-workflow,genome\/flow-workflow"} {"commit":"de348d8816151f2674410566f3eaff9d43d9dcde","old_file":"src\/markdoc\/cli\/main.py","new_file":"src\/markdoc\/cli\/main.py","old_contents":"# -*- coding: utf-8 -*-\n\nimport os\nimport argparse\n\nfrom markdoc.cli import commands\nfrom markdoc.cli.parser import parser\nfrom markdoc.config import Config, ConfigNotFound\n\n\ndef main(cmd_args=None):\n \"\"\"The main entry point for running the Markdoc CLI.\"\"\"\n \n if cmd_args is not None:\n args = parser.parse_args(cmd_args)\n else:\n args = parser.parse_args()\n \n if args.command != 'init':\n try:\n args.config = os.path.abspath(args.config)\n \n if os.path.isdir(args.config):\n config = Config.for_directory(args.config)\n elif os.path.isfile(args.config):\n config = Config.for_file(args.config)\n else:\n raise ConfigNotFound(\"Couldn't locate Markdoc config.\")\n except ConfigNotFound, exc:\n parser.error(str(exc))\n else:\n config = None\n \n command = getattr(commands, args.command.replace('-', '_'))\n return command(config, args)\n\n\nif __name__ == '__main__':\n main()\n","new_contents":"# -*- coding: utf-8 -*-\n\nimport logging\nimport os\nimport argparse\n\nfrom markdoc.cli import commands\nfrom markdoc.cli.parser import parser\nfrom markdoc.config import Config, ConfigNotFound\n\n\ndef main(cmd_args=None):\n \"\"\"The main entry point for running the Markdoc CLI.\"\"\"\n \n if cmd_args is not None:\n args = parser.parse_args(cmd_args)\n else:\n args = parser.parse_args()\n \n if args.command != 'init':\n try:\n args.config = os.path.abspath(args.config)\n \n if os.path.isdir(args.config):\n config = Config.for_directory(args.config)\n elif os.path.isfile(args.config):\n config = Config.for_file(args.config)\n else:\n raise ConfigNotFound(\"Couldn't locate Markdoc config.\")\n except ConfigNotFound, exc:\n parser.error(str(exc))\n else:\n config = None\n \n if args.quiet:\n logging.getLogger('markdoc').setLevel(logging.ERROR)\n \n command = getattr(commands, args.command.replace('-', '_'))\n return command(config, args)\n\n\nif __name__ == '__main__':\n main()\n","subject":"Use logging levels to suppress non-error output with --quiet on the CLI.","message":"Use logging levels to suppress non-error output with --quiet on the CLI.\n","lang":"Python","license":"unlicense","repos":"wlonk\/markdoc,lrem\/phdoc,lrem\/phdoc,zacharyvoase\/markdoc,snoozbuster\/markdoc,wlonk\/markdoc,snoozbuster\/markdoc"} {"commit":"a9666ecaa7ed904cb9ded38e41ea381eb08d7d65","old_file":"citrination_client\/models\/design\/target.py","new_file":"citrination_client\/models\/design\/target.py","old_contents":"from citrination_client.base.errors import CitrinationClientError\n\n\nclass Target(object):\n \"\"\"\n The optimization target for a design run. Consists of\n the name of the output column to optimize and the objective\n (either \"Max\" or \"Min\")\n \"\"\"\n\n def __init__(self, name, objective):\n \"\"\"\n Constructor.\n\n :param name: The name of the target output column\n :type name: str\n :param objective: The optimization objective; \"Min\", \"Max\", or a scalar value (such as \"5.0\")\n :type objective: str\n \"\"\"\n\n try:\n self._objective = float(objective)\n except ValueError:\n if objective.lower() not in [\"max\", \"min\"]:\n raise CitrinationClientError(\n \"Target objective must either be \\\"min\\\" or \\\"max\\\"\"\n )\n self._objective = objective\n\n self._name = name\n\n def to_dict(self):\n return {\n \"descriptor\": self._name,\n \"objective\": self._objective\n }\n","new_contents":"from citrination_client.base.errors import CitrinationClientError\n\n\nclass Target(object):\n \"\"\"\n The optimization target for a design run. Consists of\n the name of the output column to optimize and the objective\n (either \"Max\" or \"Min\", or a scalar value (such as \"5.0\"))\n \"\"\"\n\n def __init__(self, name, objective):\n \"\"\"\n Constructor.\n\n :param name: The name of the target output column\n :type name: str\n :param objective: The optimization objective; \"Min\", \"Max\", or a scalar value (such as \"5.0\")\n :type objective: str\n \"\"\"\n\n try:\n self._objective = float(objective)\n except ValueError:\n if objective.lower() not in [\"max\", \"min\"]:\n raise CitrinationClientError(\n \"Target objective must either be \\\"min\\\" or \\\"max\\\"\"\n )\n self._objective = objective\n\n self._name = name\n\n def to_dict(self):\n return {\n \"descriptor\": self._name,\n \"objective\": self._objective\n }\n","subject":"Update outdated design Target docstring","message":"Update outdated design Target docstring\n","lang":"Python","license":"apache-2.0","repos":"CitrineInformatics\/python-citrination-client"} {"commit":"dc57eb8fa84f10ffa9ba3f8133563b7de3945034","old_file":"whalelinter\/commands\/common.py","new_file":"whalelinter\/commands\/common.py","old_contents":"#!\/usr\/bin\/env python3\nfrom whalelinter.app import App\nfrom whalelinter.dispatcher import Dispatcher\nfrom whalelinter.commands.command import Command\nfrom whalelinter.commands.apt import Apt\n\n\n@Dispatcher.register(token='run', command='cd')\nclass Cd(Command):\n def __init__(self, **kwargs):\n App._collecter.throw(2002, self.line)\n return False\n\n\n@Dispatcher.register(token='run', command='rm')\nclass Rm(Command):\n def __init__(self, **kwargs):\n if (\n '-rf' in kwargs.get('args') or\n '-fr' in kwargs.get('args') or\n ('-r' in kwargs.get('args') and '-f' in kwargs.get('args'))\n ) and ('\/var\/lib\/apt\/lists' in kwargs.get('args')):\n\n if (int(Apt._has_been_used) < int(kwargs.get('lineno'))):\n Apt._has_been_used = 0\n","new_contents":"#!\/usr\/bin\/env python3\nimport re\n\nfrom whalelinter.app import App\nfrom whalelinter.dispatcher import Dispatcher\nfrom whalelinter.commands.command import Command\nfrom whalelinter.commands.apt import Apt\n\n\n@Dispatcher.register(token='run', command='cd')\nclass Cd(Command):\n def __init__(self, **kwargs):\n App._collecter.throw(2002, self.line)\n return False\n\n\n@Dispatcher.register(token='run', command='rm')\nclass Rm(Command):\n def __init__(self, **kwargs):\n rf_flags_regex = re.compile(\"(-.*[rRf].+-?[rRf]|-[rR]f|-f[rR])\")\n rf_flags = True if [i for i in kwargs.get('args') if rf_flags_regex.search(i)] else False\n cache_path_regex = re.compile(\"\/var\/lib\/apt\/lists(\\\/\\*?)?\")\n cache_path = True if [i for i in kwargs.get('args') if cache_path_regex.search(i)] else False\n\n if rf_flags and cache_path:\n if (int(Apt._has_been_used) < int(kwargs.get('lineno'))):\n Apt._has_been_used = 0\n","subject":"Enhance flags detection with regex when trying to remove apt cache","message":"Enhance flags detection with regex when trying to remove apt cache\n","lang":"Python","license":"mit","repos":"jeromepin\/whale-linter"} {"commit":"bb3ba296038f45c2de6517c1f980843ce2042aa9","old_file":"etcd3\/__init__.py","new_file":"etcd3\/__init__.py","old_contents":"from __future__ import absolute_import\n\nimport etcd3.etcdrpc as etcdrpc\nfrom etcd3.client import Etcd3Client\nfrom etcd3.client import Transactions\nfrom etcd3.client import client\nfrom etcd3.leases import Lease\nfrom etcd3.locks import Lock\nfrom etcd3.members import Member\n\n__author__ = 'Louis Taylor'\n__email__ = 'louis@kragniz.eu'\n__version__ = '0.6.0'\n\n__all__ = (\n 'Etcd3Client',\n 'Lease',\n 'Lock',\n 'Member',\n 'Transactions',\n 'client',\n 'etcdrpc',\n 'utils',\n)\n","new_contents":"from __future__ import absolute_import\n\nimport etcd3.etcdrpc as etcdrpc\nfrom etcd3.client import Etcd3Client\nfrom etcd3.client import Transactions\nfrom etcd3.client import client\nfrom etcd3.leases import Lease\nfrom etcd3.locks import Lock\nfrom etcd3.members import Member\n\n__author__ = 'Louis Taylor'\n__email__ = 'louis@kragniz.eu'\n__version__ = '0.6.0'\n\n__all__ = (\n 'Etcd3Client',\n 'Lease',\n 'Lock',\n 'Member',\n 'Transactions',\n 'client',\n 'etcdrpc',\n)\n","subject":"Remove obsolete 'utils' entry from '__all__","message":"Remove obsolete 'utils' entry from '__all__\n","lang":"Python","license":"apache-2.0","repos":"kragniz\/python-etcd3"} {"commit":"f698dbc8b10aacf6ac8ee2a5d0d63ad01bd73674","old_file":"octopus\/image\/data.py","new_file":"octopus\/image\/data.py","old_contents":"# System Imports\nimport StringIO\nimport urllib\n\n# Twisted Imports\nfrom twisted.python.util import unsignedID\n\n# Package Imports\nfrom ..data.errors import Immutable\n\nclass Image (object):\n\n\t@property\n\tdef value (self):\n\t\toutput = StringIO.StringIO()\n\t\timg = self._image_fn()\n\t\timg.scale(0.25).getPIL().save(output, format = \"PNG\")\n\t\tencoded = \"data:image\/png;base64,\" + urllib.quote(output.getvalue().encode('base64'))\n\n\t\treturn encoded\n\n\t@property\n\tdef type (self):\n\t\treturn \"Image\"\n\n\tdef serialize (self):\n\t\tif self.alias is None:\n\t\t\treturn \"[Image]\"\n\t\telse:\n\t\t\treturn str(self.alias)\n\n\tdef __init__ (self, title, fn):\n\t\tself.alias = None\n\t\tself.title = title\n\t\tself._image_fn = fn\n\n\tdef set (self, value):\n\t\traise Immutable\n\n\tdef setLogFile (self, logFile):\n\t\tpass\n\n\tdef stopLogging (self):\n\t\tpass\n\n\tdef __str__ (self):\n\t\treturn \"Image\"\n\n\tdef __repr__ (self):\n\t\treturn \"<%s at %s>\" % (\n\t\t\tself.__class__.__name__, \n\t\t\thex(unsignedID(self))\n\t\t)\n","new_contents":"# System Imports\nimport StringIO\nimport urllib\n\n# Package Imports\nfrom ..data.errors import Immutable\n\nclass Image (object):\n\n\t@property\n\tdef value (self):\n\t\toutput = StringIO.StringIO()\n\t\timg = self._image_fn()\n\t\timg.scale(0.25).getPIL().save(output, format = \"PNG\")\n\t\tencoded = \"data:image\/png;base64,\" + urllib.quote(output.getvalue().encode('base64'))\n\n\t\treturn encoded\n\n\t@property\n\tdef type (self):\n\t\treturn \"Image\"\n\n\tdef serialize (self):\n\t\tif self.alias is None:\n\t\t\treturn \"[Image]\"\n\t\telse:\n\t\t\treturn str(self.alias)\n\n\tdef __init__ (self, title, fn):\n\t\tself.alias = None\n\t\tself.title = title\n\t\tself._image_fn = fn\n\n\tdef set (self, value):\n\t\traise Immutable\n\n\tdef setLogFile (self, logFile):\n\t\tpass\n\n\tdef stopLogging (self):\n\t\tpass\n\n\tdef __str__ (self):\n\t\treturn \"Image\"\n\n\tdef __repr__ (self):\n\t\treturn \"<%s at %s>\" % (\n\t\t\tself.__class__.__name__, \n\t\t\thex(id(self))\n\t\t)\n","subject":"Replace another call to unsignedID.","message":"Replace another call to unsignedID.\n","lang":"Python","license":"mit","repos":"richardingham\/octopus,rasata\/octopus,rasata\/octopus,richardingham\/octopus,richardingham\/octopus,richardingham\/octopus,rasata\/octopus"} {"commit":"7881e6d06a34eddef5523df88ee601fb5e5d3ba6","old_file":"encryptit\/dump_json.py","new_file":"encryptit\/dump_json.py","old_contents":"import json\n\nfrom .compat import OrderedDict\nfrom .openpgp_message import OpenPGPMessage\n\n\ndef dump_stream(f, output_stream, indent=4):\n message = OpenPGPMessage.from_stream(f)\n return json.dump(message, output_stream, indent=indent,\n cls=OpenPGPJsonEncoder)\n\n\nclass OpenPGPJsonEncoder(json.JSONEncoder):\n def default(self, obj):\n if isinstance(obj, bytearray):\n return self.serialize_bytes(obj)\n\n if getattr(obj, 'serialize', None):\n return obj.serialize()\n\n return super(OpenPGPJsonEncoder, self).default(obj)\n\n def encode(self, obj):\n # If a builtin type provides a `serialize` method, use that instead of\n # the default serialisation, eg. namedtuple\n\n if getattr(obj, 'serialize', None):\n obj = obj.serialize()\n\n return super(OpenPGPJsonEncoder, self).encode(obj)\n\n @staticmethod\n def serialize_bytes(some_bytes):\n return OrderedDict([\n ('octets', ':'.join(['{0:02x}'.format(byte)\n for byte in some_bytes])),\n ('length', len(some_bytes)),\n ])\n","new_contents":"import json\n\nfrom .compat import OrderedDict\nfrom .openpgp_message import OpenPGPMessage\n\n\ndef dump_stream(f, output_stream, indent=4):\n message = OpenPGPMessage.from_stream(f)\n return json.dump(message, output_stream, indent=indent,\n cls=OpenPGPJsonEncoder)\n\n\nclass OpenPGPJsonEncoder(json.JSONEncoder):\n def default(self, obj):\n if isinstance(obj, bytearray):\n return self.serialize_bytes(obj)\n\n if getattr(obj, 'serialize', None):\n return obj.serialize()\n\n return super(OpenPGPJsonEncoder, self).default(obj)\n\n @staticmethod\n def serialize_bytes(some_bytes):\n return OrderedDict([\n ('octets', ':'.join(['{0:02x}'.format(byte)\n for byte in some_bytes])),\n ('length', len(some_bytes)),\n ])\n","subject":"Revert \"Fix JSON encoding of `PacketLocation`\"","message":"Revert \"Fix JSON encoding of `PacketLocation`\"\n\nThis reverts commit 9e91912c6c1764c88890ec47df9372e6ac41612c.\n","lang":"Python","license":"agpl-3.0","repos":"paulfurley\/encryptit,paulfurley\/encryptit"} {"commit":"84a0aef34f8ab187de7e0c2b17c2e79d0e8f2330","old_file":"feedback\/forms.py","new_file":"feedback\/forms.py","old_contents":"from django import forms\n\nfrom feedback.models import Feedback\n\nclass FeedbackForm(forms.ModelForm):\n class Meta:\n model = Feedback\n exclude = (\"url\", \"resolved\", \"publish\",)","new_contents":"from django import forms\n\nfrom feedback.models import Feedback\n\nclass FeedbackForm(forms.ModelForm):\n class Meta:\n model = Feedback\n exclude = (\"user\", \"url\", \"resolved\", \"publish\",)","subject":"Remove the user field from the form to counter-balance making the field editable.","message":"Remove the user field from the form to counter-balance making the field editable.","lang":"Python","license":"bsd-3-clause","repos":"gabrielhurley\/django-user-feedback"} {"commit":"1239128a082757c3a7d53e7b14c189dda06f4171","old_file":"flaws\/__init__.py","new_file":"flaws\/__init__.py","old_contents":"#!\/usr\/bin\/env python\nimport sys\n\nfrom funcy import split, map\n\nfrom .analysis import global_usage, local_usage, FileSet\n\n\nimport sys, ipdb, traceback\n\ndef info(type, value, tb):\n traceback.print_exception(type, value, tb)\n print\n ipdb.pm()\n\nsys.excepthook = info\n\n\ndef main():\n command = sys.argv[1]\n kwargs, args = split(r'^--', sys.argv[2:])\n kwargs = dict(map(r'^--(\\w+)=(.+)', kwargs))\n\n from .ext import django\n django.register(args, kwargs)\n\n files = FileSet(args, base=kwargs.get('base'), ignore=kwargs.get('ignore'))\n if command == 'global':\n global_usage(files)\n elif command == 'local':\n local_usage(files)\n else:\n print 'Unknown command', command\n\n\nif __name__ == '__main__':\n main()\n","new_contents":"#!\/usr\/bin\/env python\nimport sys\n\nfrom funcy import split, map\n\nfrom .analysis import global_usage, local_usage, FileSet\n\n\ndef main():\n command = sys.argv[1]\n kwargs, args = split(r'^--', sys.argv[2:])\n kwargs = dict(map(r'^--(\\w+)(?:=(.+))?', kwargs))\n\n # Run ipdb on exception\n if 'ipdb' in kwargs:\n import ipdb, traceback\n\n def info(type, value, tb):\n traceback.print_exception(type, value, tb)\n print\n ipdb.pm()\n\n sys.excepthook = info\n\n # Register plugins\n from .ext import django\n django.register(args, kwargs)\n\n # Do the job\n files = FileSet(args, base=kwargs.get('base'), ignore=kwargs.get('ignore'))\n if command == 'global':\n global_usage(files)\n elif command == 'local':\n local_usage(files)\n else:\n print 'Unknown command', command\n\n\nif __name__ == '__main__':\n main()\n","subject":"Make ipdb hook turn on only when --ipdb","message":"Make ipdb hook turn on only when --ipdb\n","lang":"Python","license":"bsd-2-clause","repos":"Suor\/flaws"} {"commit":"c64682fe6204b56bd5282c46a7c7168a55b46a86","old_file":"spicedham\/__init__.py","new_file":"spicedham\/__init__.py","old_contents":"from pkg_resources import iter_entry_points\n\nfrom config import config\n\n\nplugins = []\nfor plugin in iter_entry_points(group='spicedham.classifiers', name=None):\n pluginClass = plugin.load()\n plugins.append(pluginClass())\n\n\ndef train(training_data, is_spam):\n for plugin in plugins:\n plugin.train(training_data, is_spam)\n\n\ndef classify(classification_data):\n average_score = 0\n total = 0\n for plugin in plugins:\n value = plugin.classify(classification_data)\n if value != None:\n total += 1\n average_score += value\n return average_score \/ total\n","new_contents":"from pkg_resources import iter_entry_points\n\nfrom config import config\n\n\nplugins = []\nfor plugin in iter_entry_points(group='spicedham.classifiers', name=None):\n pluginClass = plugin.load()\n plugins.append(pluginClass())\n\n\ndef train(training_data, is_spam):\n for plugin in plugins:\n plugin.train(training_data, is_spam)\n\n\ndef classify(classification_data):\n average_score = 0\n total = 0\n for plugin in plugins:\n value = plugin.classify(classification_data)\n if value != None:\n total += 1\n average_score += value\n if total > 0:\n return average_score \/ total\n else:\n return 0\n","subject":"Allow for the case where no plugin returns a score","message":"Allow for the case where no plugin returns a score\n","lang":"Python","license":"mpl-2.0","repos":"mozilla\/spicedham,mozilla\/spicedham"} {"commit":"0a5e2134fda46269626b6fac367a28218734b256","old_file":"conf_site\/accounts\/tests\/__init__.py","new_file":"conf_site\/accounts\/tests\/__init__.py","old_contents":"from factory import fuzzy\n\nfrom django.contrib.auth import get_user_model\nfrom django.test import TestCase\n\n\nclass AccountsTestCase(TestCase):\n def setUp(self):\n super(AccountsTestCase, self).setUp()\n\n self.password = fuzzy.FuzzyText(length=16)\n self.new_password = fuzzy.FuzzyText(length=16)\n\n user_model = get_user_model()\n self.user = user_model.objects.get_or_create(\n username=\"test\",\n email=\"example@example.com\",\n first_name=\"Test\",\n last_name=\"User\",\n )[0]\n self.user.set_password(self.password)\n self.user.save()\n\n def _become_superuser(self):\n \"\"\"Make this testcase's user a superuser.\"\"\"\n self.user.is_superuser = True\n self.user.save()\n","new_contents":"from factory import fuzzy\n\nfrom django.contrib.auth import get_user_model\nfrom django.test import TestCase\n\n\nclass AccountsTestCase(TestCase):\n def setUp(self):\n super(AccountsTestCase, self).setUp()\n\n self.password = fuzzy.FuzzyText(length=16)\n self.new_password = fuzzy.FuzzyText(length=16)\n\n user_model = get_user_model()\n self.user = user_model.objects.get_or_create(\n username=\"test\",\n email=\"example@example.com\",\n first_name=\"Test\",\n last_name=\"User\",\n )[0]\n self.user.set_password(self.password)\n self.user.save()\n\n def _become_staff(self):\n \"\"\"Make this testcase's user a staff user.\"\"\"\n self.user.is_staff = True\n self.user.is_superuser = False\n self.user.save()\n\n def _become_superuser(self):\n \"\"\"Make this testcase's user a superuser.\"\"\"\n self.user.is_superuser = True\n self.user.save()\n","subject":"Add `_become_staff` method to AccountsTestCase.","message":"Add `_become_staff` method to AccountsTestCase.\n","lang":"Python","license":"mit","repos":"pydata\/conf_site,pydata\/conf_site,pydata\/conf_site"} {"commit":"76e436daef154bdf6acd1b0569f6fa2baa61addd","old_file":"pyxform\/tests_v1\/test_audit.py","new_file":"pyxform\/tests_v1\/test_audit.py","old_contents":"from pyxform.tests_v1.pyxform_test_case import PyxformTestCase\n\n\nclass AuditTest(PyxformTestCase):\n def test_audit(self):\n self.assertPyxformXform(\n name=\"meta_audit\",\n md=\"\"\"\n | survey | | | |\n | | type | name | label |\n | | audit | audit | |\n \"\"\",\n xml__contains=[\n '<meta>',\n '<audit\/>',\n '<\/meta>',\n '<bind nodeset=\"\/meta_audit\/meta\/audit\" type=\"binary\"\/>'],\n )\n\n def test_audit_random_name(self):\n self.assertPyxformXform(\n name=\"meta_audit\",\n md=\"\"\"\n | survey | | | |\n | | type | name | label |\n | | audit | bobby | |\n \"\"\",\n xml__contains=[\n '<meta>',\n '<audit\/>',\n '<\/meta>',\n '<bind nodeset=\"\/meta_audit\/meta\/audit\" type=\"binary\"\/>'],\n )","new_contents":"from pyxform.tests_v1.pyxform_test_case import PyxformTestCase\n\n\nclass AuditTest(PyxformTestCase):\n def test_audit(self):\n self.assertPyxformXform(\n name=\"meta_audit\",\n md=\"\"\"\n | survey | | | |\n | | type | name | label |\n | | audit | audit | |\n \"\"\",\n xml__contains=[\n '<meta>',\n '<audit\/>',\n '<\/meta>',\n '<bind nodeset=\"\/meta_audit\/meta\/audit\" type=\"binary\"\/>'],\n )\n\n def test_audit_random_name(self):\n self.assertPyxformXform(\n name=\"meta_audit\",\n md=\"\"\"\n | survey | | | |\n | | type | name | label |\n | | audit | bobby | |\n \"\"\",\n xml__contains=[\n '<meta>',\n '<audit\/>',\n '<\/meta>',\n '<bind nodeset=\"\/meta_audit\/meta\/audit\" type=\"binary\"\/>'],\n )\n\n def test_audit_blank_name(self):\n self.assertPyxformXform(\n name=\"meta_audit\",\n md=\"\"\"\n | survey | | | |\n | | type | name | label |\n | | audit | | |\n \"\"\",\n xml__contains=[\n '<meta>',\n '<audit\/>',\n '<\/meta>',\n '<bind nodeset=\"\/meta_audit\/meta\/audit\" type=\"binary\"\/>'],\n )","subject":"Add test for blank audit name.","message":"Add test for blank audit name.\n","lang":"Python","license":"bsd-2-clause","repos":"XLSForm\/pyxform,XLSForm\/pyxform"} {"commit":"fb25fa04cf553b1084425a1f2af6a9315266ffaf","old_file":"salt\/renderers\/yaml_jinja.py","new_file":"salt\/renderers\/yaml_jinja.py","old_contents":"'''\nThe default rendering engine, process yaml with the jinja2 templating engine\n\nThis renderer will take a yaml file with the jinja2 template and render it to a\nhigh data format for salt states.\n'''\n\n# Import Python Modules\nimport os\n\n# Import thirt party modules\nimport yaml\ntry:\n yaml.Loader = yaml.CLoader\n yaml.Dumper = yaml.CDumper\nexcept:\n pass\n\n# Import Salt libs\nfrom salt.utils.jinja import get_template\n\ndef render(template_file, env='', sls=''):\n '''\n Render the data passing the functions and grains into the rendering system\n '''\n if not os.path.isfile(template_file):\n return {}\n\n passthrough = {}\n passthrough['salt'] = __salt__\n passthrough['grains'] = __grains__\n passthrough['env'] = env\n passthrough['sls'] = sls\n\n template = get_template(template_file, __opts__, env)\n\n yaml_data = template.render(**passthrough)\n\n return yaml.safe_load(yaml_data)\n","new_contents":"'''\nThe default rendering engine, process yaml with the jinja2 templating engine\n\nThis renderer will take a yaml file with the jinja2 template and render it to a\nhigh data format for salt states.\n'''\n\n# Import Python Modules\nimport os\n\n# Import thirt party modules\nimport yaml\ntry:\n yaml.Loader = yaml.CLoader\n yaml.Dumper = yaml.CDumper\nexcept:\n pass\n\n# Import Salt libs\nfrom salt.utils.jinja import get_template\n\ndef render(template_file, env='', sls=''):\n '''\n Render the data passing the functions and grains into the rendering system\n '''\n if not os.path.isfile(template_file):\n return {}\n\n passthrough = {}\n passthrough['salt'] = __salt__\n passthrough['grains'] = __grains__\n passthrough['pillar'] = __pillar__\n passthrough['env'] = env\n passthrough['sls'] = sls\n\n template = get_template(template_file, __opts__, env)\n\n yaml_data = template.render(**passthrough)\n\n return yaml.safe_load(yaml_data)\n","subject":"Add pillar data to default renderer","message":"Add pillar data to default renderer\n","lang":"Python","license":"apache-2.0","repos":"saltstack\/salt,saltstack\/salt,saltstack\/salt,saltstack\/salt,saltstack\/salt"} {"commit":"d63e792815b9bfe485e4127bdcb088374d8e983e","old_file":"salvus\/scripts\/first_boot.py","new_file":"salvus\/scripts\/first_boot.py","old_contents":"#!\/usr\/bin\/env python\n\n# This script is run by \/etc\/rc.local when booting up. It does special configuration\n# depending on what images are mounted, etc.\n\nimport os\n\nif os.path.exists('\/mnt\/home\/'):\n # Compute machine\n if not os.path.exists('\/mnt\/home\/aquota.group'):\n os.system(\"quotacheck -cug \/mnt\/home\")\n os.system(\"quotaon -a\")\n\n # disable quotas for now, so that students in my class can do Sage development.\n os.system('quotaoff -a')\n\n # Restore user accounts\n if os.path.exists('\/mnt\/home\/etc\/'):\n os.system(\"cp \/mnt\/home\/etc\/* \/etc\/\")\n else:\n os.system(\"mkdir -p \/mnt\/home\/etc\/\")\n\n # Setup \/tmp so it is on the external disk image (has that quota) and is clean, since this is a fresh boot.\n os.system(\"rm -rf \/mnt\/home\/tmp; mkdir -p \/mnt\/home\/tmp\/; chmod +t \/mnt\/home\/tmp; mount -o bind \/mnt\/home\/tmp \/tmp; chmod a+rwx \/mnt\/home\/tmp\/\")\n\n os.system(\"mkdir -p \/mnt\/home\/scratch; mkdir -p \/scratch; chmod +t \/mnt\/home\/tmp; mount -o bind \/mnt\/home\/scratch \/scratch; chmod a+rwx \/mnt\/home\/scratch\/\")\n","new_contents":"#!\/usr\/bin\/env python\n\n# This script is run by \/etc\/rc.local when booting up. It does special configuration\n# depending on what images are mounted, etc.\n\nimport os\n\nif os.path.exists('\/mnt\/home\/'):\n # Compute machine\n if not os.path.exists('\/mnt\/home\/aquota.group'):\n os.system(\"quotacheck -cug \/mnt\/home\")\n os.system(\"quotaon -a\")\n\n # disable quotas for now, so that students in my class can do Sage development.\n os.system('quotaoff -a')\n\n # Restore user accounts\n if os.path.exists('\/mnt\/home\/etc\/'):\n os.system(\"cp \/mnt\/home\/etc\/* \/etc\/\")\n else:\n os.system(\"mkdir -p \/mnt\/home\/etc\/\")\n\n # Setup \/tmp so it is on the external disk image (has that quota) and is clean, since this is a fresh boot.\n os.system(\"rm -rf \/mnt\/home\/tmp; mkdir -p \/mnt\/home\/tmp\/; chmod +t \/mnt\/home\/tmp; mount -o bind \/mnt\/home\/tmp \/tmp; chmod a+rwx \/mnt\/home\/tmp\/\")\n\n os.system(\"mkdir -p \/mnt\/home\/scratch; mkdir -p \/scratch; chmod +t \/mnt\/home\/tmp; mount -o bind \/mnt\/home\/scratch \/scratch; chmod a+rwx \/mnt\/home\/scratch\/\")\n\n # Remove .ssh keys on compute node from \/mnt\/home\/salvus account, since this is a security risk (in case compute node is r00ted).\n os.system(\"rm -rf \/mnt\/home\/salvus\/.ssh\/id_rsa*\")\n","subject":"Remove .ssh keys on compute node from \/mnt\/home\/salvus account, since this is a security risk (in case compute node is r00ted).","message":"Remove .ssh keys on compute node from \/mnt\/home\/salvus account, since this is a security risk (in case compute node is r00ted).\n","lang":"Python","license":"agpl-3.0","repos":"tscholl2\/smc,sagemathinc\/smc,sagemathinc\/smc,sagemathinc\/smc,DrXyzzy\/smc,tscholl2\/smc,tscholl2\/smc,DrXyzzy\/smc,DrXyzzy\/smc,DrXyzzy\/smc,tscholl2\/smc,tscholl2\/smc,sagemathinc\/smc"} {"commit":"e645104656fda22f4c0c2b3d9841ed792b1e7103","old_file":"conftest.py","new_file":"conftest.py","old_contents":"import sys\n\n\npytest_plugins = 'setuptools.tests.fixtures'\n\n\ndef pytest_addoption(parser):\n parser.addoption(\n \"--package_name\", action=\"append\", default=[],\n help=\"list of package_name to pass to test functions\",\n )\n\n\ncollect_ignore = [\n 'tests\/manual_test.py',\n 'setuptools\/tests\/mod_with_constant.py',\n 'setuptools\/_distutils',\n '_distutils_hack',\n 'setuptools\/extern',\n 'pkg_resources\/extern',\n 'pkg_resources\/tests\/data',\n 'setuptools\/_vendor',\n 'pkg_resources\/_vendor',\n]\n\n\nif sys.version_info < (3, 6):\n collect_ignore.append('docs\/conf.py') # uses f-strings\n collect_ignore.append('pavement.py')\n","new_contents":"import sys\n\nimport pytest\n\n\npytest_plugins = 'setuptools.tests.fixtures'\n\n\ndef pytest_addoption(parser):\n parser.addoption(\n \"--package_name\", action=\"append\", default=[],\n help=\"list of package_name to pass to test functions\",\n )\n parser.addoption(\n \"--integration\", action=\"store_true\", default=False,\n help=\"run integration tests (only)\"\n )\n\n\ndef pytest_configure(config):\n config.addinivalue_line(\"markers\", \"integration: indicate integration tests\")\n\n if config.option.integration:\n # Assume unit tests and flake already run\n config.option.flake8 = False\n\n\ncollect_ignore = [\n 'tests\/manual_test.py',\n 'setuptools\/tests\/mod_with_constant.py',\n 'setuptools\/_distutils',\n '_distutils_hack',\n 'setuptools\/extern',\n 'pkg_resources\/extern',\n 'pkg_resources\/tests\/data',\n 'setuptools\/_vendor',\n 'pkg_resources\/_vendor',\n]\n\n\nif sys.version_info < (3, 6):\n collect_ignore.append('docs\/conf.py') # uses f-strings\n collect_ignore.append('pavement.py')\n\n\n@pytest.fixture(autouse=True)\ndef _skip_integration(request):\n running_integration_tests = request.config.getoption(\"--integration\")\n is_integration_test = request.node.get_closest_marker(\"integration\")\n if running_integration_tests and not is_integration_test:\n pytest.skip(\"running integration tests only\")\n if not running_integration_tests and is_integration_test:\n pytest.skip(\"skipping integration tests\")\n","subject":"Configure pytest to enable\/disable integration tests","message":"Configure pytest to enable\/disable integration tests\n","lang":"Python","license":"mit","repos":"pypa\/setuptools,pypa\/setuptools,pypa\/setuptools"} {"commit":"84d9a421b33660f4ad17432fef8604a55b0e2302","old_file":"calvin\/runtime\/south\/plugins\/io\/sensors\/environmental\/platform\/raspberry_pi\/sensehat_impl\/environmental.py","new_file":"calvin\/runtime\/south\/plugins\/io\/sensors\/environmental\/platform\/raspberry_pi\/sensehat_impl\/environmental.py","old_contents":"# -*- coding: utf-8 -*-\n\n# Copyright (c) 2015 Ericsson AB\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom sense_hat import SenseHat\nfrom calvin.runtime.south.plugins.io.sensors.environmental import base_environmental\n\n\nclass Environmental(base_environmental.EnvironmentalBase):\n\n \"\"\"\n Raspberry Pi Sense HAT environmental sensors\n \"\"\"\n\n def __init__(self):\n self.sense = SenseHat()\n\n def get_temperature(self):\n return self.sense.get_temperature()\n\n def get_humidity(self):\n return self.sense.get_humidity()\n\n def get_pressure(self):\n return self.sense.get_pressure()\n","new_contents":"# -*- coding: utf-8 -*-\n\n# Copyright (c) 2015 Ericsson AB\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom sense_hat import SenseHat\nfrom calvin.runtime.south.plugins.io.sensors.environmental import base_environmental\n\n\nclass SenseHatResourceHandler(object):\n \n sense_hat = SenseHat()\n\n def __init__(self):\n pass\n \n def get_temperature(self):\n return self.sense_hat.get_temperature()\n \n def get_humidity(self):\n return self.sense_hat.get_humidity()\n \n def get_pressure(self):\n return self.sense_hat.get_pressure()\n\n\nclass Environmental(base_environmental.EnvironmentalBase):\n\n \"\"\"\n Raspberry Pi Sense HAT environmental sensors\n \"\"\"\n\n def __init__(self):\n self.sense = SenseHatResourceHandler()\n\n def get_temperature(self):\n return self.sense.get_temperature()\n\n def get_humidity(self):\n return self.sense.get_humidity()\n\n def get_pressure(self):\n return self.sense.get_pressure()\n","subject":"Allow use of sensors from more than one actor concurrenctly","message":"Sensehat: Allow use of sensors from more than one actor concurrenctly\n","lang":"Python","license":"apache-2.0","repos":"EricssonResearch\/calvin-base,EricssonResearch\/calvin-base,EricssonResearch\/calvin-base,EricssonResearch\/calvin-base"} {"commit":"fa30c15c6bdaa49d3af2e717f559d279da770b46","old_file":"src\/streamlink\/plugins\/arconai.py","new_file":"src\/streamlink\/plugins\/arconai.py","old_contents":"import re\n\nfrom streamlink.plugin import Plugin\nfrom streamlink.plugin.api import http\nfrom streamlink.plugin.api import useragents\nfrom streamlink.stream import HLSStream\n\n_url_re = re.compile(r'''https?:\/\/(www\\.)?arconaitv\\.me\/stream\\.php\\?id=\\d+''')\n_playlist_re = re.compile(r'''source\\ssrc=[\"'](?P<url>[^\"']+)[\"']''')\n\n\nclass ArconaiTv(Plugin):\n @classmethod\n def can_handle_url(cls, url):\n return _url_re.match(url)\n\n def _get_streams(self):\n headers = {\n 'User-Agent': useragents.CHROME,\n 'Referer': self.url\n }\n\n res = http.get(self.url, headers=headers)\n\n match = _playlist_re.search(res.text)\n if match is None:\n return\n\n url = match.group('url')\n\n if url:\n self.logger.debug('HLS URL: {0}'.format(url))\n yield 'live', HLSStream(self.session, url, headers=headers)\n\n__plugin__ = ArconaiTv\n","new_contents":"import re\n\nfrom streamlink.plugin import Plugin\nfrom streamlink.plugin.api import http\nfrom streamlink.plugin.api import useragents\nfrom streamlink.stream import HLSStream\n\n_url_re = re.compile(r'''https?:\/\/(www\\.)?arconaitv\\.co\/stream\\.php\\?id=\\d+''')\n_playlist_re = re.compile(r'''source\\ssrc=[\"'](?P<url>[^\"']+)[\"']''')\n\n\nclass ArconaiTv(Plugin):\n @classmethod\n def can_handle_url(cls, url):\n return _url_re.match(url)\n\n def _get_streams(self):\n headers = {\n 'User-Agent': useragents.CHROME,\n 'Referer': self.url\n }\n\n res = http.get(self.url, headers=headers)\n\n match = _playlist_re.search(res.text)\n if match is None:\n return\n\n url = match.group('url')\n\n if url:\n self.logger.debug('HLS URL: {0}'.format(url))\n yield 'live', HLSStream(self.session, url, headers=headers)\n\n__plugin__ = ArconaiTv\n","subject":"Update Arconaitv to new url","message":"Update Arconaitv to new url","lang":"Python","license":"bsd-2-clause","repos":"bastimeyer\/streamlink,streamlink\/streamlink,back-to\/streamlink,melmorabity\/streamlink,bastimeyer\/streamlink,gravyboat\/streamlink,melmorabity\/streamlink,javiercantero\/streamlink,gravyboat\/streamlink,javiercantero\/streamlink,wlerin\/streamlink,wlerin\/streamlink,beardypig\/streamlink,chhe\/streamlink,streamlink\/streamlink,beardypig\/streamlink,back-to\/streamlink,chhe\/streamlink"} {"commit":"98aa2b25c63ec5bd6384a9d398a70996799b135e","old_file":"mygpoauth\/urls.py","new_file":"mygpoauth\/urls.py","old_contents":"from django.conf.urls import include, url\nfrom django.contrib import admin\nfrom django.views.generic.base import RedirectView\n\nfrom mygpoauth import oauth2\n\n\nurlpatterns = [\n # Examples:\n # url(r'^$', 'mygpoauth.views.home', name='home'),\n # url(r'^blog\/', include('blog.urls')),\n\n url(r'^$', RedirectView.as_view(url='http:\/\/mygpo-auth.rtfd.org\/'),\n name='index'),\n url(r'^admin\/', include(admin.site.urls)),\n url(r'^oauth2\/', include('mygpoauth.oauth2.urls', namespace='oauth2')),\n]\n","new_contents":"from django.conf.urls import include, url\nfrom django.contrib import admin\nfrom django.views.generic.base import RedirectView\n\nfrom mygpoauth import oauth2\n\n\nurlpatterns = [\n # Examples:\n # url(r'^$', 'mygpoauth.views.home', name='home'),\n # url(r'^blog\/', include('blog.urls')),\n\n url(r'^$', RedirectView.as_view(url='http:\/\/mygpo-auth.rtfd.org\/',\n permanent=False),\n name='index'),\n url(r'^admin\/', include(admin.site.urls)),\n url(r'^oauth2\/', include('mygpoauth.oauth2.urls', namespace='oauth2')),\n]\n","subject":"Make \"\/\" a non-permanent redirect","message":"Make \"\/\" a non-permanent redirect\n","lang":"Python","license":"agpl-3.0","repos":"gpodder\/mygpo-auth,gpodder\/mygpo-auth"} {"commit":"b706c1a949b10a7dd4b3206c02de8d4abda088a9","old_file":"pytac\/mini_project.py","new_file":"pytac\/mini_project.py","old_contents":"import pytac.load_csv\nimport pytac.epics\n\ndef main():\n lattice = pytac.load_csv.load('VMX', pytac.epics.EpicsControlSystem())\n bpms = lattice.get_elements('BPM')\n bpms_n = 0\n try:\n for bpm in bpms:\n bpm.get_pv_name('y')\n bpms_n += 1\n print 'There exist {0} BPMy elements in the ring.'.format(bpms_n)\n except:\n print 'Warning! There exists a bpm with no y field.'\n\n\nif __name__=='__main__':\n main()\n","new_contents":"import pytac.load_csv\nimport pytac.epics\n\ndef main():\n # First task: print the number of bpm y elements in the ring.\n lattice = pytac.load_csv.load('VMX', pytac.epics.EpicsControlSystem())\n bpms = lattice.get_elements('BPM')\n bpms_n = 0\n try:\n for bpm in bpms:\n bpm.get_pv_name('y')\n bpms_n += 1\n print 'There exist {0} BPMy elements in the ring.'.format(bpms_n)\n except:\n print 'Warning! There exists a bpm with no y field.'\n\n # Second task: Print each bpmx pv along with the associated value.\n for bpm in bpms:\n print bpm.get_pv_name('x', 'readback')\n print bpm.get_pv_value('x', 'readback')\n\n\nif __name__=='__main__':\n main()\n","subject":"Print each bpmx pv name along with the associated value to stdout","message":"Print each bpmx pv name along with the associated value to stdout\n","lang":"Python","license":"apache-2.0","repos":"razvanvasile\/Work-Mini-Projects,razvanvasile\/Work-Mini-Projects,razvanvasile\/Work-Mini-Projects"} {"commit":"0b8b32a044e92f4e996af734d44a2d93d1492684","old_file":"project_code\/bulk_fitting.py","new_file":"project_code\/bulk_fitting.py","old_contents":"\n'''\nBulk spectral line fitting with SDSS galaxy spectra\n'''\n\nimport os\nfrom astropy.io import fits\nfrom pandas import concat\n\n# Bring in the package funcs\nfrom specfit import do_specfit\nfrom download_spectra import download_spectra\n\n\ndef bulk_fit(obs_file, output_file, keep_spectra=False):\n '''\n Downloads files based off of the entries in the given file, performs\n spectral line fitting and saves the results to a FITS table.\n '''\n\n # Open the file\n data_file = fits.open(obs_file)\n\n spectra_data = data_file[1].data\n del data_file\n\n num_spectra = spectra_data['Z'].shape[0]\n\n for i in range(num_spectra):\n spec_info = spectra_data[i]\n\n # Download the spectrum\n spec_name = \\\n download_spectra(spec_info['PLATEID'], spec_info['FIBREID'],\n spec_info['MJD'], spec_info['SURVEY'])\n\n spec_df = do_specfit(spec_name, verbose=False)\n\n if i == 0:\n df = spec_df\n else:\n df = concat([df, spec_df])\n\n if not keep_spectra:\n os.system('rm ' + spec_name)\n\n df.write(output_file)\n\n return\n","new_contents":"\n'''\nBulk spectral line fitting with SDSS galaxy spectra\n'''\n\nimport os\nfrom astropy.io import fits\nfrom pandas import DataFrame\n\n# Bring in the package funcs\nfrom specfit import do_specfit\nfrom download_spectra import download_spectra\n\n\ndef bulk_fit(obs_file, output_file, keep_spectra=False):\n '''\n Downloads files based off of the entries in the given file, performs\n spectral line fitting and saves the results to a FITS table.\n '''\n\n # Open the file\n data_file = fits.open(obs_file)\n\n spectra_data = data_file[1].data\n del data_file\n\n num_spectra = spectra_data['Z'].shape[0]\n\n for i in range(num_spectra):\n spec_info = spectra_data[i]\n\n # Download the spectrum\n spec_name = \\\n download_spectra(spec_info['PLATE'], spec_info['FIBERID'],\n spec_info['MJD'], spec_info['SURVEY'])\n\n spec_df = do_specfit(spec_name, verbose=False)\n\n if i == 0:\n df = DataFrame(spec_df, columns=[spec_name[:-5]])\n else:\n df[spec_name[:-5]] = spec_df\n\n if not keep_spectra:\n os.system('rm ' + spec_name)\n\n df.to_csv(output_file)\n\n return\n","subject":"Correct names, concat dataframes properly","message":"Correct names, concat dataframes properly\n","lang":"Python","license":"mit","repos":"e-koch\/Phys-595"} {"commit":"3e41a447076c4aa183923700c1c8203afdf07377","old_file":"bitbots_body_behavior\/src\/bitbots_body_behavior\/decisions\/ball_close.py","new_file":"bitbots_body_behavior\/src\/bitbots_body_behavior\/decisions\/ball_close.py","old_contents":"from dynamic_stack_decider.abstract_decision_element import AbstractDecisionElement\n\n\nclass BallClose(AbstractDecisionElement):\n def __init__(self, blackboard, dsd, parameters=None):\n super(BallClose, self).__init__(blackboard, dsd, parameters)\n self.ball_close_distance = self.blackboard.config['ball_close_distance']\n\n def perform(self, reevaluate=False):\n \"\"\"\n Determines whether the ball is in close range to the robot. The distance threshold is set in the config file.\n :param reevaluate:\n :return:\n \"\"\"\n self.publish_debug_data(\"ball_distance\", self.blackboard.world_model.get_ball_distance())\n\n if self.blackboard.world_model.get_ball_distance() < self.ball_close_distance:\n return 'YES'\n return 'NO'\n\n def get_reevaluate(self):\n return True\n","new_contents":"from dynamic_stack_decider.abstract_decision_element import AbstractDecisionElement\n\n\nclass BallClose(AbstractDecisionElement):\n def __init__(self, blackboard, dsd, parameters=None):\n super(BallClose, self).__init__(blackboard, dsd, parameters)\n self.ball_close_distance = parameters.get(\"distance\", self.blackboard.config['ball_close_distance'])\n\n def perform(self, reevaluate=False):\n \"\"\"\n Determines whether the ball is in close range to the robot. The distance threshold is set in the config file.\n :param reevaluate:\n :return:\n \"\"\"\n self.publish_debug_data(\"ball_distance\", self.blackboard.world_model.get_ball_distance())\n\n if self.blackboard.world_model.get_ball_distance() < self.ball_close_distance:\n return 'YES'\n return 'NO'\n\n def get_reevaluate(self):\n return True\n","subject":"Add param to ball close","message":"Add param to ball close\n","lang":"Python","license":"bsd-3-clause","repos":"bit-bots\/bitbots_behaviour"} {"commit":"ba4953423450c3bf2924aa76f37694b405c8ee85","old_file":"parse-zmmailbox-ids.py","new_file":"parse-zmmailbox-ids.py","old_contents":"import re\nimport sys\n\n# $ zmmailbox -z -m username@domain.tld search -l 200 \"in:\/inbox (before:today)\"\n# num: 200, more: true\n#\n# Id Type From Subject Date\n# ------- ---- -------------------- -------------------------------------------------- --------------\n# 1. -946182 conv admin Daily mail report 09\/24\/15 23:57\n# 2. 421345 conv John Some great news for you 09\/24\/15 23:57\n\nREGEX_HEAD = re.compile(r'^Id')\nREGEX_HEAD_SEP = re.compile(r'^---')\n\nREGEX_DATA = re.compile(r'^(\\d+)\\.\\s+\\-?(\\d+)\\s+(\\S+)')\n\n\ndef main():\n lines = [line.strip() for line in sys.stdin.readlines() if line.strip()]\n\n while True:\n line = lines.pop(0)\n if REGEX_HEAD.search(line):\n break\n\n line = lines.pop(0)\n assert REGEX_HEAD_SEP.search(line)\n\n ids = []\n\n for line in lines:\n matched = REGEX_DATA.match(line)\n if matched:\n ids.append(matched.group(2))\n else:\n sys.stderr.write(\"Couldn't parse line: {0}\\n\".format(line))\n sys.exit(1)\n\n for an_id in ids:\n print an_id\n\nif __name__ == '__main__':\n main()\n","new_contents":"import re\nimport sys\n\n# $ zmmailbox -z -m username@domain.tld search -l 200 \"in:\/inbox (before:today)\"\n# num: 200, more: true\n#\n# Id Type From Subject Date\n# ------- ---- -------------------- -------------------------------------------------- --------------\n# 1. -946182 conv admin Daily mail report 09\/24\/15 23:57\n# 2. 421345 conv John Some great news for you 09\/24\/15 23:57\n\nREGEX_HEAD = re.compile(r'^Id')\nREGEX_HEAD_SEP = re.compile(r'^---')\n\nREGEX_DATA = re.compile(r'^(\\d+)\\.\\s+(\\-?\\d+)\\s+(\\S+)')\n\n\ndef main():\n lines = [line.strip() for line in sys.stdin.readlines() if line.strip()]\n\n while True:\n line = lines.pop(0)\n if REGEX_HEAD.search(line):\n break\n\n line = lines.pop(0)\n assert REGEX_HEAD_SEP.search(line)\n\n ids = []\n\n for line in lines:\n matched = REGEX_DATA.match(line)\n if matched:\n ids.append(matched.group(2))\n else:\n sys.stderr.write(\"Couldn't parse line: {0}\\n\".format(line))\n sys.exit(1)\n\n print ','.join(ids)\n\n\nif __name__ == '__main__':\n main()\n","subject":"Include '-' in ID, print IDs separated by ','","message":"Include '-' in ID, print IDs separated by ','\n","lang":"Python","license":"apache-2.0","repos":"hgdeoro\/zimbra7-to-zimbra8-password-migrator"} {"commit":"62297b3c937d386b759ec14a078cee36f2550d44","old_file":"src\/aiy\/_drivers\/_alsa.py","new_file":"src\/aiy\/_drivers\/_alsa.py","old_contents":"# Copyright 2017 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Helpers for ALSA tools, including aplay and arecord.\"\"\"\n\n\ndef sample_width_to_string(sample_width):\n \"\"\"Convert sample width (bytes) to ALSA format string.\"\"\"\n return {1: 's8', 2: 's16', 4: 's32'}[sample_width]\n","new_contents":"# Copyright 2017 Google Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\"\"\"Helpers for ALSA tools, including aplay and arecord.\"\"\"\n\n\ndef sample_width_to_string(sample_width):\n \"\"\"Convert sample width (bytes) to ALSA format string.\"\"\"\n return {1: 's8', 2: 's16', 4: 's32'}.get(sample_width, None)\n","subject":"Return None with invalid sample_width from sample_width_to_string.","message":"Return None with invalid sample_width from sample_width_to_string.\n","lang":"Python","license":"apache-2.0","repos":"google\/aiyprojects-raspbian,t1m0thyj\/aiyprojects-raspbian,google\/aiyprojects-raspbian,t1m0thyj\/aiyprojects-raspbian,google\/aiyprojects-raspbian,google\/aiyprojects-raspbian,google\/aiyprojects-raspbian"} {"commit":"ec07e139b5585a8ed9bed14426dac987267ebf05","old_file":"sbtsettings.py","new_file":"sbtsettings.py","old_contents":"import sublime\n\n\nclass SBTSettings(object):\n\n def __init__(self, window):\n self.window = window\n self._plugin_settings = sublime.load_settings('SublimeSBT.sublime-settings')\n\n def sbt_command(self):\n return self._view_settings().get('sbt_command', self._plugin_settings.get('sbt_command'))\n\n def play_command(self):\n return self._view_settings().get('sbt_command', self._plugin_settings.get('play_command'))\n\n def color_scheme(self):\n self.get('color_scheme')\n\n def get(self, name):\n return self._view_settings().get(name, self._plugin_settings.get(name))\n\n def add_on_change(self, on_change):\n self._plugin_settings.add_on_change('SublimeSBT', on_change)\n\n def _view_settings(self):\n return self.window.active_view().settings().get('SublimeSBT', {})\n","new_contents":"import sublime\n\nfrom util import maybe\n\n\nclass SBTSettings(object):\n\n def __init__(self, window):\n self.window = window\n self._plugin_settings = sublime.load_settings('SublimeSBT.sublime-settings')\n\n def sbt_command(self):\n return self._view_settings().get('sbt_command', self._plugin_settings.get('sbt_command'))\n\n def play_command(self):\n return self._view_settings().get('sbt_command', self._plugin_settings.get('play_command'))\n\n def color_scheme(self):\n self.get('color_scheme')\n\n def get(self, name):\n return self._view_settings().get(name, self._plugin_settings.get(name))\n\n def add_on_change(self, on_change):\n self._plugin_settings.add_on_change('SublimeSBT', on_change)\n\n def _view_settings(self):\n for view in maybe(self.window.active_view()):\n return view.settings().get('SublimeSBT', {})\n return {}\n","subject":"Fix AttributeError getting project settings when no active view","message":"Fix AttributeError getting project settings when no active view\n","lang":"Python","license":"mit","repos":"jarhart\/SublimeSBT"} {"commit":"e1b23cdc089b3a05ae4959c9859e16e5e21b5c91","old_file":"apps\/careeropportunity\/views.py","new_file":"apps\/careeropportunity\/views.py","old_contents":"#-*- coding: utf-8 -*-\nfrom django.shortcuts import render_to_response\nfrom django.shortcuts import get_object_or_404\nfrom django.template import RequestContext\nfrom apps.careeropportunity.models import CareerOpportunity\n\nimport datetime\n\n\ndef index(request):\n opportunities = CareerOpportunity.objects.all()\n \n return render_to_response('careeropportunity\/index.html', \\\n {'opportunities': opportunities}, \\\n context_instance=RequestContext(request))\n\ndef details(request, opportunity_id):\n opportunity = get_object_or_404(CareerOpportunity, pk=opportunity_id)\n\n return render_to_response('careeropportunity\/details.html', \\\n {'opportunity': opportunity}, \\\n context_instance=RequestContext(request))\n","new_contents":"#-*- coding: utf-8 -*-\nfrom datetime import datetime\nfrom django.shortcuts import render_to_response\nfrom django.shortcuts import get_object_or_404\nfrom django.template import RequestContext\nfrom apps.careeropportunity.models import CareerOpportunity\n\n\ndef index(request):\n opportunities = CareerOpportunity.objects.filter(start__lte=datetime.now(), end__gte=datetime.now())\n \n return render_to_response('careeropportunity\/index.html', \\\n {'opportunities': opportunities}, \\\n context_instance=RequestContext(request))\n\ndef details(request, opportunity_id):\n opportunity = get_object_or_404(CareerOpportunity, pk=opportunity_id)\n\n return render_to_response('careeropportunity\/details.html', \\\n {'opportunity': opportunity}, \\\n context_instance=RequestContext(request))\n","subject":"Make careerop only display active ops","message":"Make careerop only display active ops\n","lang":"Python","license":"mit","repos":"dotKom\/onlineweb4,dotKom\/onlineweb4,dotKom\/onlineweb4,dotKom\/onlineweb4"} {"commit":"fe167bfd25c0c86b3c6fb5ef76eb24036ad2b6da","old_file":"tests\/ne_np\/__init__.py","new_file":"tests\/ne_np\/__init__.py","old_contents":"from __future__ import unicode_literals\n\nimport unittest\nimport re\n\nfrom faker import Factory\nfrom faker.utils import text\nfrom .. import string_types\n\n\nclass ne_NP_FactoryTestCase(unittest.TestCase):\n \n def setUp(self):\n self.factory = Factory.create('ne_NP')\n \n def test_address(self):\n from faker.providers.address.ne_NP import Provider\n countries = Provider.countries\n country = self.factory.country()\n assert country\n assert isinstance(country, string_types)\n assert country in countries\n \n districts = Provider.districts\n district = self.factory.district()\n assert district\n assert isinstance(district, string_types)\n assert district in districts\n\n cities = Provider.cities\n city = self.factory.city()\n assert city\n assert isinstance(city, string_types)\n assert city in cities\n \n def test_names(self):\n from faker.providers.person.ne_NP import Provider\n first_names = Provider.first_names\n name = self.factory.name()\n first_name, last_name = name.split()\n assert first_name\n assert isinstance(first_name, string_types)\n assert first_name in first_names\n \n last_names = Provider.last_names\n assert last_names\n assert isinstance(last_name, string_types)\n assert last_name in last_names\n \n \n ","new_contents":"from __future__ import unicode_literals\n\nimport unittest\n\nfrom faker import Factory\nfrom .. import string_types\n\n\nclass NeNPFactoryTestCase(unittest.TestCase):\n\n def setUp(self):\n self.factory = Factory.create('ne_NP')\n \n def test_address(self):\n from faker.providers.address.ne_NP import Provider\n country = self.factory.country()\n assert isinstance(country, string_types)\n assert country in Provider.countries\n\n district = self.factory.district()\n assert isinstance(district, string_types)\n assert district in Provider.districts\n\n city = self.factory.city()\n assert isinstance(city, string_types)\n assert city in Provider.cities\n\n def test_names(self):\n from faker.providers.person.ne_NP import Provider\n for _ in range(10000):\n name = self.factory.name().split()\n assert all(isinstance(n, string_types) for n in name)\n # name should always be 2-3 words. If 3, first word\n # should be a prefix.\n assert name[-2] in Provider.first_names\n assert name[-1] in Provider.last_names\n prefixes = Provider.prefixes_male + Provider.prefixes_female\n if len(name) == 3:\n assert name[0] in prefixes\n","subject":"Fix incorrect ne_NP locale tests","message":"Fix incorrect ne_NP locale tests\n\nThis test incorrectly assumes a call to name() will\nyield only a first\/last name, which isn't always true for this\nlocale. I suspect it hasn't been uncovered yet because the\ntests are seeded the same at the beginning of every run. It only\nbecomes a problem when you start moving tests around. This change\naddresses the incorrect assertions as well as makes the file PEP8\ncompliant.\n","lang":"Python","license":"mit","repos":"trtd\/faker,joke2k\/faker,joke2k\/faker,danhuss\/faker"} {"commit":"6b9d9c33b4d68a008bb992b9a11ab2f02a4d5cbd","old_file":"shelltest\/tests\/test_runner.py","new_file":"shelltest\/tests\/test_runner.py","old_contents":"import tempfile\nimport StringIO\n\nimport pytest\n\nfrom shelltest.shelltest import ShellTest, ShellTestSource, ShellTestRunner\n\n\n@pytest.fixture\ndef tests():\n return [ShellTest('echo hello', 'hello\\n', ShellTestSource('', 0)),\n ShellTest('echo $?', '0\\n', ShellTestSource('', 2))]\n\n\ndef test_run(tests):\n r = ShellTestRunner(tests)\n results = r.run()\n assert len(results) == 2\n assert results[0].success\n assert results[0].ret_code == 0\n assert results[0].test == tests[0]\n assert results[0].actual_output == tests[0].expected_output\n assert results[1].success\n assert results[1].ret_code == 0\n assert results[1].test == tests[1]\n assert results[1].actual_output == tests[1].expected_output\n","new_contents":"import tempfile\nimport StringIO\n\nimport pytest\n\nfrom shelltest.shelltest import ShellTest, ShellTestSource, ShellTestRunner\n\n\ndef runner(tests):\n tests = [ShellTest(cmd, output, ShellTestSource('', 0)) for cmd, output in tests]\n return ShellTestRunner(tests)\n\n\n@pytest.mark.parametrize(\"cmd,output,ret_code,success\", (\n ('echo hello', 'hello\\n', 0, True),\n ('echo $?', '0\\n', 0, True),\n ('exit 42', '', 42, True)))\ndef test_echo(cmd, output, ret_code, success):\n r = runner([(cmd, output)])\n res = r.run()[0]\n assert res.success == success\n assert res.ret_code == ret_code\n assert res.test == r.tests[0]\n assert res.actual_output == output\n","subject":"Update runner tests to use parameters","message":"Update runner tests to use parameters\n","lang":"Python","license":"mit","repos":"jthacker\/shelltest,jthacker\/shelltest"} {"commit":"82f8861df01d67335499682743f69b1763cc3c35","old_file":"uberlogs\/handlers\/kill_process.py","new_file":"uberlogs\/handlers\/kill_process.py","old_contents":"import sys\nimport os\nfrom logging import Handler as LoggingHandler\n\n\nclass KillProcessHandler(LoggingHandler):\n\n def emit(self, record):\n if record.levelno != self.level:\n return\n\n try:\n # flush text before exiting\n for fd in [sys.stdout, sys.stderr]:\n fd.flush()\n finally:\n # Twisted writes unhandled errors in different calls\n # If we exit on the first call, we'd lose the actual error\n for log_to_ignore in [\"Unhandled error in Deferred\"]:\n if log_to_ignore.lower() in record.getMessage().lower():\n return\n os._exit(1)\n","new_contents":"import sys\nimport os\nfrom logging import Handler as LoggingHandler\n\n\nclass KillProcessHandler(LoggingHandler):\n\n def emit(self, record):\n if record.levelno != self.level:\n return\n\n # flush text before exiting\n for fd in [sys.stdout, sys.stderr]:\n fd.flush()\n\n # Twisted writes unhandled errors in different calls\n # If we exit on the first call, we'd lose the actual error\n for log_to_ignore in [\"Unhandled error in Deferred\"]:\n if log_to_ignore.lower() in record.getMessage().lower():\n return\n os._exit(1)\n","subject":"Remove redundant try\/catch block in kill process handler","message":"Remove redundant try\/catch block in kill process handler\n","lang":"Python","license":"mit","repos":"odedlaz\/uberlogs,odedlaz\/uberlogs"} {"commit":"6a531ebe5e097d277a7b07e142e98009d622253f","old_file":"tests\/registryd\/test_root_accessible.py","new_file":"tests\/registryd\/test_root_accessible.py","old_contents":"# Pytest will pick up this module automatically when running just \"pytest\".\n#\n# Each test_*() function gets passed test fixtures, which are defined\n# in conftest.py. So, a function \"def test_foo(bar)\" will get a bar()\n# fixture created for it.\n\nPROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'\nACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'\n\ndef get_property(proxy, iface_name, prop_name):\n return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE)\n\ndef test_accessible_iface_properties(registry_root, session_manager):\n values = [\n ('Name', 'main'),\n ('Description', ''),\n ]\n\n for prop_name, expected in values:\n assert get_property(registry_root, ACCESSIBLE_IFACE, prop_name) == expected\n\ndef test_registry_root_has_null_parent(registry_root, session_manager):\n assert get_property(registry_root, ACCESSIBLE_IFACE, 'Parent') == ('', '\/org\/a11y\/atspi\/null')\n\ndef test_empty_registry_has_zero_children(registry_root, session_manager):\n assert get_property(registry_root, ACCESSIBLE_IFACE, 'ChildCount') == 0\n","new_contents":"# Pytest will pick up this module automatically when running just \"pytest\".\n#\n# Each test_*() function gets passed test fixtures, which are defined\n# in conftest.py. So, a function \"def test_foo(bar)\" will get a bar()\n# fixture created for it.\n\nPROPERTIES_IFACE = 'org.freedesktop.DBus.Properties'\nACCESSIBLE_IFACE = 'org.a11y.atspi.Accessible'\n\ndef get_property(proxy, iface_name, prop_name):\n return proxy.Get(iface_name, prop_name, dbus_interface=PROPERTIES_IFACE)\n\ndef test_accessible_iface_properties(registry_root, session_manager):\n values = [\n ('Name', 'main'),\n ('Description', ''),\n ('Parent', ('', '\/org\/a11y\/atspi\/null')),\n ('ChildCount', 0),\n ]\n\n for prop_name, expected in values:\n assert get_property(registry_root, ACCESSIBLE_IFACE, prop_name) == expected\n","subject":"Put all the Accessibility property tests in a single function","message":"Put all the Accessibility property tests in a single function\n\nWe already had machinery for that, anyway.\n","lang":"Python","license":"lgpl-2.1","repos":"GNOME\/at-spi2-core,GNOME\/at-spi2-core,GNOME\/at-spi2-core"} {"commit":"b0e3ed09d401389680db14c6892e84f016423c97","old_file":"simplesqlite\/error.py","new_file":"simplesqlite\/error.py","old_contents":"\"\"\"\n.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>\n\"\"\"\n\nimport sqlite3\nfrom typing import Optional\n\nfrom tabledata import NameValidationError # noqa: W0611\n\n\nclass DatabaseError(sqlite3.DatabaseError):\n \"\"\"\n Exception raised for errors that are related to the database.\n\n .. seealso::\n - `sqlite3.DatabaseError <https:\/\/docs.python.org\/3\/library\/sqlite3.html#sqlite3.DatabaseError>`__\n \"\"\"\n\n\nclass NullDatabaseConnectionError(DatabaseError):\n \"\"\"\n Exception raised when executing an operation of\n :py:class:`~simplesqlite.SimpleSQLite` instance without connection to\n a SQLite database file.\n \"\"\"\n\n\nclass TableNotFoundError(DatabaseError):\n \"\"\"\n Exception raised when accessed the table that not exists in the database.\n \"\"\"\n\n\nclass AttributeNotFoundError(DatabaseError):\n \"\"\"\n Exception raised when accessed the attribute that not exists in the table.\n \"\"\"\n\n\nclass SqlSyntaxError(Exception):\n \"\"\"\n Exception raised when a SQLite query syntax is invalid.\n \"\"\"\n\n\nclass OperationalError(sqlite3.OperationalError):\n \"\"\"\n Exception raised when failed to execute a query.\n \"\"\"\n\n @property\n def message(self) -> Optional[str]:\n return self.__message\n\n def __init__(self, *args, **kwargs) -> None:\n self.__message = kwargs.pop(\"message\", None)\n\n super().__init__()\n","new_contents":"\"\"\"\n.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>\n\"\"\"\n\nimport sqlite3\nfrom typing import Optional\n\nfrom tabledata import NameValidationError # noqa: W0611\n\n\nclass DatabaseError(sqlite3.DatabaseError):\n \"\"\"\n Exception raised for errors that are related to the database.\n\n .. seealso::\n - `sqlite3.DatabaseError <https:\/\/docs.python.org\/3\/library\/sqlite3.html#sqlite3.DatabaseError>`__\n \"\"\"\n\n\nclass NullDatabaseConnectionError(DatabaseError):\n \"\"\"\n Exception raised when executing an operation of\n :py:class:`~simplesqlite.SimpleSQLite` instance without connection to\n a SQLite database file.\n \"\"\"\n\n\nclass TableNotFoundError(DatabaseError):\n \"\"\"\n Exception raised when accessed the table that not exists in the database.\n \"\"\"\n\n\nclass AttributeNotFoundError(DatabaseError):\n \"\"\"\n Exception raised when accessed the attribute that not exists in the table.\n \"\"\"\n\n\nclass SqlSyntaxError(Exception):\n \"\"\"\n Exception raised when a SQLite query syntax is invalid.\n \"\"\"\n\n\nclass OperationalError(sqlite3.OperationalError):\n \"\"\"\n Exception raised when failed to execute a query.\n \"\"\"\n\n @property\n def message(self) -> Optional[str]:\n return self.__message\n\n def __init__(self, *args, **kwargs) -> None:\n self.__message = kwargs.pop(\"message\", None)\n\n super().__init__(*args)\n","subject":"Modify to pass args to the base class constructor","message":"Modify to pass args to the base class constructor\n","lang":"Python","license":"mit","repos":"thombashi\/SimpleSQLite,thombashi\/SimpleSQLite"} {"commit":"0d0b470e470ee913cb8983f932323921d405607b","old_file":"refabric\/context_managers.py","new_file":"refabric\/context_managers.py","old_contents":"# coding=utf-8\nfrom contextlib import contextmanager\n\nfrom fabric.context_managers import settings, hide, prefix\nfrom fabric.state import env\n\n__all__ = ['get_sudo_context', 'sudo', 'only_messages', 'prefix']\n\n\n@contextmanager\ndef sudo(user=None):\n with settings(sudo_user=user or env.sudo_user or env.user, use_sudo=True):\n yield\n\n\nsilent = lambda: settings(hide('commands'), warn_only=True)\nhide_prefix = lambda: settings(output_prefix=False)\nabort_on_error = lambda: settings(warn_only=False)\n\n\n@contextmanager\ndef shell_env(**env_vars):\n orig_shell = env['shell']\n env_vars_str = ' '.join('{0}={1}'.format(key, value)\n for key, value in env_vars.items())\n env['shell'] = '{0} {1}'.format(env_vars_str, orig_shell)\n yield\n env['shell'] = orig_shell\n","new_contents":"# coding=utf-8\nfrom contextlib import contextmanager\n\nfrom fabric.context_managers import settings, hide, prefix\nfrom fabric.state import env\n\n__all__ = ['get_sudo_context', 'sudo', 'only_messages', 'prefix']\n\n\n@contextmanager\ndef sudo(user=None):\n with settings(sudo_user=user or env.sudo_user or env.user, use_sudo=True):\n yield\n\n\nsilent = lambda *h: settings(hide('commands', *h), warn_only=True)\nhide_prefix = lambda: settings(output_prefix=False)\nabort_on_error = lambda: settings(warn_only=False)\n\n\n@contextmanager\ndef shell_env(**env_vars):\n orig_shell = env['shell']\n env_vars_str = ' '.join('{0}={1}'.format(key, value)\n for key, value in env_vars.items())\n env['shell'] = '{0} {1}'.format(env_vars_str, orig_shell)\n yield\n env['shell'] = orig_shell\n","subject":"Add fine tuning to silent helper","message":"Add fine tuning to silent helper","lang":"Python","license":"mit","repos":"5monkeys\/refabric"} {"commit":"4aeb85126cf5f75d89cc466c3f7fea2f53702a13","old_file":"bluebottle\/votes\/serializers.py","new_file":"bluebottle\/votes\/serializers.py","old_contents":"from bluebottle.votes.models import Vote\nfrom bluebottle.bb_accounts.serializers import UserPreviewSerializer\nfrom rest_framework import serializers\n\n\nclass VoteSerializer(serializers.ModelSerializer):\n voter = UserPreviewSerializer(read_only=True)\n project = serializers.SlugRelatedField(source='project', slug_field='slug')\n\n class Meta:\n model = Vote\n fields = ('id', 'voter', 'project')\n","new_contents":"from bluebottle.votes.models import Vote\nfrom bluebottle.bb_accounts.serializers import UserPreviewSerializer\nfrom rest_framework import serializers\n\n\nclass VoteSerializer(serializers.ModelSerializer):\n voter = UserPreviewSerializer(read_only=True)\n project = serializers.SlugRelatedField(source='project', slug_field='slug')\n\n class Meta:\n model = Vote\n fields = ('id', 'voter', 'project', 'created')\n","subject":"Add created to votes api serializer","message":"Add created to votes api serializer\n","lang":"Python","license":"bsd-3-clause","repos":"onepercentclub\/bluebottle,onepercentclub\/bluebottle,jfterpstra\/bluebottle,onepercentclub\/bluebottle,jfterpstra\/bluebottle,onepercentclub\/bluebottle,jfterpstra\/bluebottle,jfterpstra\/bluebottle,onepercentclub\/bluebottle"} {"commit":"8425efaf60b642418786c523d142a370dae3c3a0","old_file":"quilt_server\/config.py","new_file":"quilt_server\/config.py","old_contents":"# Copyright (c) 2017 Quilt Data, Inc. All rights reserved.\n\n\"\"\"\nDefault config values the Flask app.\nShared between dev, stage, and production.\n\nSee `app.config.from_object('...')` in __init__.py.\n\"\"\"\n\nSQLALCHEMY_TRACK_MODIFICATIONS = False\nSQLALCHEMY_ECHO = False # Turn it on for debugging.\n\nPACKAGE_URL_EXPIRATION = 60*60*12 # 12 Hours\n\nJSON_USE_ENCODE_METHODS = True # Support the __json__ method in Node\n\n# 100MB max for request body.\nMAX_CONTENT_LENGTH = 100 * 1024 * 1024\n","new_contents":"# Copyright (c) 2017 Quilt Data, Inc. All rights reserved.\n\n\"\"\"\nDefault config values the Flask app.\nShared between dev, stage, and production.\n\nSee `app.config.from_object('...')` in __init__.py.\n\"\"\"\n\nSQLALCHEMY_TRACK_MODIFICATIONS = False\nSQLALCHEMY_ECHO = False # Turn it on for debugging.\n\nPACKAGE_URL_EXPIRATION = 60*60*24 # 24 hours\n\nJSON_USE_ENCODE_METHODS = True # Support the __json__ method in Node\n\n# 100MB max for request body.\nMAX_CONTENT_LENGTH = 100 * 1024 * 1024\n","subject":"Change the S3 URL expiration time from 12 to 24 hours","message":"Change the S3 URL expiration time from 12 to 24 hours\n","lang":"Python","license":"apache-2.0","repos":"quiltdata\/quilt-compiler,quiltdata\/quilt-compiler,quiltdata\/quilt,quiltdata\/quilt-compiler,quiltdata\/quilt-compiler,quiltdata\/quilt,quiltdata\/quilt,quiltdata\/quilt,quiltdata\/quilt"} {"commit":"83036bf711dd5047ef87a56ea9d8def604923882","old_file":"ts3observer\/features.py","new_file":"ts3observer\/features.py","old_contents":"'''\nCreated on Nov 10, 2014\n\n@author: fechnert\n'''\n\nimport logging\n\n\nclass Feature(object):\n ''' Represents a abstract Feature '''\n\n def __init__(self, config, clients, channels):\n ''' Initialize the Object '''\n self.config = config\n self.clients = clients\n self.channels = channels\n\n def run(self):\n raise NotImplementedError\n\n\nclass UsernameBlacklist(Feature):\n pass\n\n\nclass AutoMove(Feature):\n pass\n\n\nclass MusicbotDetect(Feature):\n pass\n","new_contents":"'''\nCreated on Nov 10, 2014\n\n@author: fechnert\n'''\n\nimport logging\n\n\nclass Feature(object):\n ''' Represents a abstract Feature '''\n\n def __init__(self, config, clients, channels):\n ''' Initialize the Object '''\n self.config = config\n self.clients = clients\n self.channels = channels\n\n def run(self):\n raise NotImplementedError\n\n\nclass UsernameBlacklist(Feature):\n pass\n\nclass OnAway(Feature):\n pass\n\nclass OnIdle(Feature):\n pass\n\nclass OnMute(Feature):\n pass\n\nclass OnDeaf(Feature):\n pass\n","subject":"Change Feature classes to match the new config","message":"Change Feature classes to match the new config\n","lang":"Python","license":"mit","repos":"HWDexperte\/ts3observer"} {"commit":"a991d3d66d4a021eb88ad1ff982686f3b930d468","old_file":"sqlobject\/dberrors.py","new_file":"sqlobject\/dberrors.py","old_contents":"\"\"\"dberrors: database exception classes for SQLObject.\n\n These classes are dictated by the DB API v2.0:\n\n http:\/\/www.python.org\/topics\/database\/DatabaseAPI-2.0.html\n\"\"\"\n\nimport sys\nif sys.version_info[0] >= 3:\n StandardError = Exception\n\nclass Error(StandardError):\n pass\n\n\nclass Warning(StandardError):\n pass\n\n\nclass InterfaceError(Error):\n pass\n\n\nclass DatabaseError(Error):\n pass\n\n\nclass InternalError(DatabaseError):\n pass\n\n\nclass OperationalError(DatabaseError):\n pass\n\n\nclass ProgrammingError(DatabaseError):\n pass\n\n\nclass IntegrityError(DatabaseError):\n pass\n\n\nclass DataError(DatabaseError):\n pass\n\n\nclass NotSupportedError(DatabaseError):\n pass\n\n\nclass DuplicateEntryError(IntegrityError):\n pass\n","new_contents":"\"\"\"dberrors: database exception classes for SQLObject.\n\n These classes are dictated by the DB API v2.0:\n\n http:\/\/www.python.org\/topics\/database\/DatabaseAPI-2.0.html\n\"\"\"\n\nimport sys\nif sys.version_info[0] >= 3:\n StandardError = Exception\n\n\nclass Error(StandardError):\n pass\n\n\nclass Warning(StandardError):\n pass\n\n\nclass InterfaceError(Error):\n pass\n\n\nclass DatabaseError(Error):\n pass\n\n\nclass InternalError(DatabaseError):\n pass\n\n\nclass OperationalError(DatabaseError):\n pass\n\n\nclass ProgrammingError(DatabaseError):\n pass\n\n\nclass IntegrityError(DatabaseError):\n pass\n\n\nclass DataError(DatabaseError):\n pass\n\n\nclass NotSupportedError(DatabaseError):\n pass\n\n\nclass DuplicateEntryError(IntegrityError):\n pass\n","subject":"Fix flake8 E302 expected 2 blank lines, found 1","message":"Fix flake8 E302 expected 2 blank lines, found 1\n","lang":"Python","license":"lgpl-2.1","repos":"sqlobject\/sqlobject,drnlm\/sqlobject,sqlobject\/sqlobject,drnlm\/sqlobject"} {"commit":"aa1bbbe1d4b463be8cedaaf445fa44612592513f","old_file":"minette\/test\/helper.py","new_file":"minette\/test\/helper.py","old_contents":"from time import time\n\nfrom ..core import Minette\nfrom ..models import Message\n\n\nclass MinetteForTest(Minette):\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self.default_channel = kwargs.get(\"default_channel\", \"\")\n self.case_id = str(int(time() * 10000000))\n\n def chat(self, request, **kwargs):\n self.logger.info(\"start testcase: \" + self.case_id)\n # convert to Message\n if isinstance(request, str):\n request = Message(text=request, **kwargs)\n # set channel and channel_user_id for this test case\n if request.channel == \"console\":\n request.channel = self.default_channel or request.channel\n if request.channel_user_id == \"anonymous\":\n request.channel_user_id = \"user\" + self.case_id\n # chat and return response\n response = super().chat(request)\n self.logger.info(\"end testcase: \" + self.case_id)\n return response\n","new_contents":"from time import time\n\nfrom ..core import Minette\nfrom ..models import Message\n\n\nclass MinetteForTest(Minette):\n def __init__(self, **kwargs):\n super().__init__(**kwargs)\n self.default_channel = kwargs.get(\"default_channel\", \"\")\n self.case_id = str(int(time() * 10000000))\n\n def chat(self, request, **kwargs):\n self.logger.info(\"start testcase: \" + self.case_id)\n # convert to Message\n if isinstance(request, str):\n request = Message(text=request, **kwargs)\n # set channel and channel_user_id for this test case\n if request.channel == \"console\":\n request.channel = self.default_channel or request.channel\n if request.channel_user_id == \"anonymous\":\n request.channel_user_id = \"user\" + self.case_id\n # chat and return response\n response = super().chat(request)\n if response.messages:\n response.text = response.messages[0].text\n else:\n response.text = \"\"\n self.logger.info(\"end testcase: \" + self.case_id)\n return response\n","subject":"Add `text` attribute to response from `chat`","message":"Add `text` attribute to response from `chat`\n","lang":"Python","license":"apache-2.0","repos":"uezo\/minette-python"} {"commit":"1d043a9fa2140992435bc5d6583601464d96f5b0","old_file":"wafer\/schedule\/renderers.py","new_file":"wafer\/schedule\/renderers.py","old_contents":"from django_medusa.renderers import StaticSiteRenderer\n\n\nclass ScheduleRenderer(StaticSiteRenderer):\n def get_paths(self):\n paths = [\"\/schedule\/\", ]\n return paths\n\nrenderers = [ScheduleRenderer, ]\n","new_contents":"from django_medusa.renderers import StaticSiteRenderer\nfrom wafer.schedule.models import Venue\n\n\nclass ScheduleRenderer(StaticSiteRenderer):\n def get_paths(self):\n paths = [\"\/schedule\/\", ]\n\n # Add the venues\n items = Venue.objects.all()\n for item in items:\n paths.append(item.get_absolute_url())\n return paths\n\nrenderers = [ScheduleRenderer, ]\n","subject":"Add venues to site export","message":"Add venues to site export\n","lang":"Python","license":"isc","repos":"CarlFK\/wafer,CarlFK\/wafer,CTPUG\/wafer,CTPUG\/wafer,CTPUG\/wafer,CTPUG\/wafer,CarlFK\/wafer,CarlFK\/wafer"} {"commit":"93474c192516864b2c609f2225a0f6c1fa8ca9a8","old_file":"Cauldron\/ext\/commandkeywords\/__init__.py","new_file":"Cauldron\/ext\/commandkeywords\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nAn extension for a command-based keyword.\n\"\"\"\nfrom __future__ import absolute_import\n\nfrom Cauldron.types import Boolean, DispatcherKeywordType\nfrom Cauldron.exc import NoWriteNecessary\n\nclass CommandKeyword(Boolean, DispatcherKeywordType):\n \"\"\"This keyword will receive boolean writes as 1, and will always be set to 0. \n \n Actions can then be performed in callbacks, etc., every time this keyword is triggered.\n \"\"\"\n \n KTL_REGISTERED = False\n \n KTL_TYPE = 'boolean'\n \n def __init__(self, *args, **kwargs):\n kwargs['initial'] = '0'\n super(CommandKeyword, self).__init__(*args, **kwargs)\n \n def prewrite(self, value):\n \"\"\"Before writing, trigger no-write-necssary if value is False\"\"\"\n if self.translate(value) == '0':\n raise NoWriteNecessary(\"No write needed, command not triggered.\")\n return super(CommandKeyword, self).prewrite(value)\n \n def postwrite(self, value):\n \"\"\"Special postwrite that always sets the value to '0'.\"\"\"\n self.set('0', force=True)\n \n # We don't have to do anything else here.\n ","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nAn extension for a command-based keyword.\n\"\"\"\nfrom __future__ import absolute_import\n\nfrom Cauldron.types import Boolean, DispatcherKeywordType\nfrom Cauldron.exc import NoWriteNecessary\nfrom Cauldron.utils.callbacks import Callbacks\n\nclass CommandKeyword(Boolean, DispatcherKeywordType):\n \"\"\"This keyword will receive boolean writes as 1, and will always be set to 0. \n \n Actions can then be performed in callbacks, etc., every time this keyword is triggered.\n \"\"\"\n \n KTL_REGISTERED = False\n \n KTL_TYPE = 'boolean'\n \n def __init__(self, *args, **kwargs):\n kwargs['initial'] = '0'\n super(CommandKeyword, self).__init__(*args, **kwargs)\n self._cbs = Callbacks()\n \n def command(self, func):\n \"\"\"Add command items.\"\"\"\n self._cbs.add(func)\n \n def prewrite(self, value):\n \"\"\"Before writing, trigger no-write-necssary if value is False\"\"\"\n if self.translate(value) == '0':\n raise NoWriteNecessary(\"No write needed, command not triggered.\")\n return super(CommandKeyword, self).prewrite(value)\n \n def write(self, value):\n \"\"\"Write to the commands.\"\"\"\n if str(value) == '1':\n self._cbs(self)\n \n def postwrite(self, value):\n \"\"\"Special postwrite that always sets the value to '0'.\"\"\"\n self.set('0', force=True)\n \n # We don't have to do anything else here.\n ","subject":"Make command-keyword compatible with DFW implementation","message":"Make command-keyword compatible with DFW implementation\n","lang":"Python","license":"bsd-3-clause","repos":"alexrudy\/Cauldron"} {"commit":"6cf2a3966e12af5f86781a5d20c0810953722811","old_file":"tests\/basics\/scope.py","new_file":"tests\/basics\/scope.py","old_contents":"# test scoping rules\n\n# explicit global variable\na = 1\ndef f():\n global a\n global a, a # should be able to redefine as global\n a = 2\nf()\nprint(a)\n\n# explicit nonlocal variable\ndef f():\n a = 1\n def g():\n nonlocal a\n nonlocal a, a # should be able to redefine as nonlocal\n a = 2\n g()\n return a\nprint(f())\n","new_contents":"# test scoping rules\n\n# explicit global variable\na = 1\ndef f():\n global a\n global a, a # should be able to redefine as global\n a = 2\nf()\nprint(a)\n\n# explicit nonlocal variable\ndef f():\n a = 1\n def g():\n nonlocal a\n nonlocal a, a # should be able to redefine as nonlocal\n a = 2\n g()\n return a\nprint(f())\n\n# nonlocal at inner-inner level (h)\ndef f():\n x = 1\n def g():\n def h():\n nonlocal x\n return x\n return h\n return g\nprint(f()()())\n\n# nonlocal declared at outer level (g), and referenced by inner level (h)\ndef f():\n x = 1\n def g():\n nonlocal x\n def h():\n return x\n return h\n return g\nprint(f()()())\n","subject":"Add further tests for nonlocal scoping and closures.","message":"tests\/basics: Add further tests for nonlocal scoping and closures.\n","lang":"Python","license":"mit","repos":"lowRISC\/micropython,ryannathans\/micropython,tralamazza\/micropython,micropython\/micropython-esp32,cwyark\/micropython,deshipu\/micropython,lowRISC\/micropython,alex-march\/micropython,adafruit\/micropython,Peetz0r\/micropython-esp32,SHA2017-badge\/micropython-esp32,turbinenreiter\/micropython,deshipu\/micropython,ryannathans\/micropython,HenrikSolver\/micropython,trezor\/micropython,micropython\/micropython-esp32,hiway\/micropython,kerneltask\/micropython,mhoffma\/micropython,MrSurly\/micropython-esp32,chrisdearman\/micropython,hosaka\/micropython,selste\/micropython,infinnovation\/micropython,tobbad\/micropython,adafruit\/circuitpython,tobbad\/micropython,PappaPeppar\/micropython,bvernoux\/micropython,toolmacher\/micropython,matthewelse\/micropython,dmazzella\/micropython,turbinenreiter\/micropython,torwag\/micropython,MrSurly\/micropython,hiway\/micropython,kerneltask\/micropython,hosaka\/micropython,lowRISC\/micropython,mhoffma\/micropython,cwyark\/micropython,pramasoul\/micropython,alex-robbins\/micropython,dmazzella\/micropython,infinnovation\/micropython,bvernoux\/micropython,oopy\/micropython,selste\/micropython,jmarcelino\/pycom-micropython,turbinenreiter\/micropython,mhoffma\/micropython,tuc-osg\/micropython,MrSurly\/micropython,adafruit\/circuitpython,dxxb\/micropython,pozetroninc\/micropython,SHA2017-badge\/micropython-esp32,deshipu\/micropython,Timmenem\/micropython,bvernoux\/micropython,alex-robbins\/micropython,turbinenreiter\/micropython,puuu\/micropython,dxxb\/micropython,pozetroninc\/micropython,TDAbboud\/micropython,puuu\/micropython,hosaka\/micropython,deshipu\/micropython,tuc-osg\/micropython,AriZuu\/micropython,Peetz0r\/micropython-esp32,adafruit\/circuitpython,adafruit\/circuitpython,MrSurly\/micropython,toolmacher\/micropython,hosaka\/micropython,lowRISC\/micropython,henriknelson\/micropython,pfalcon\/micropython,bvernoux\/micropython,HenrikSolver\/micropython,blazewicz\/micropython,PappaPeppar\/micropython,alex-march\/micropython,henriknelson\/micropython,dxxb\/micropython,blazewicz\/micropython,Timmenem\/micropython,blazewicz\/micropython,blazewicz\/micropython,tobbad\/micropython,swegener\/micropython,oopy\/micropython,TDAbboud\/micropython,jmarcelino\/pycom-micropython,hiway\/micropython,jmarcelino\/pycom-micropython,PappaPeppar\/micropython,HenrikSolver\/micropython,dmazzella\/micropython,trezor\/micropython,cwyark\/micropython,kerneltask\/micropython,tobbad\/micropython,TDAbboud\/micropython,matthewelse\/micropython,chrisdearman\/micropython,torwag\/micropython,ryannathans\/micropython,pramasoul\/micropython,trezor\/micropython,oopy\/micropython,toolmacher\/micropython,adafruit\/micropython,pfalcon\/micropython,chrisdearman\/micropython,Timmenem\/micropython,Timmenem\/micropython,puuu\/micropython,AriZuu\/micropython,dxxb\/micropython,swegener\/micropython,swegener\/micropython,MrSurly\/micropython,pramasoul\/micropython,hosaka\/micropython,alex-robbins\/micropython,micropython\/micropython-esp32,selste\/micropython,tralamazza\/micropython,tralamazza\/micropython,matthewelse\/micropython,mhoffma\/micropython,Peetz0r\/micropython-esp32,selste\/micropython,alex-march\/micropython,tralamazza\/micropython,adafruit\/circuitpython,pfalcon\/micropython,adafruit\/circuitpython,jmarcelino\/pycom-micropython,tuc-osg\/micropython,TDAbboud\/micropython,henriknelson\/micropython,MrSurly\/micropython-esp32,torwag\/micropython,chrisdearman\/micropython,puuu\/micropython,SHA2017-badge\/micropython-esp32,MrSurly\/micropython-esp32,Peetz0r\/micropython-esp32,micropython\/micropython-esp32,hiway\/micropython,torwag\/micropython,SHA2017-badge\/micropython-esp32,mhoffma\/micropython,AriZuu\/micropython,oopy\/micropython,pfalcon\/micropython,pramasoul\/micropython,MrSurly\/micropython,hiway\/micropython,toolmacher\/micropython,AriZuu\/micropython,kerneltask\/micropython,chrisdearman\/micropython,selste\/micropython,micropython\/micropython-esp32,matthewelse\/micropython,PappaPeppar\/micropython,matthewelse\/micropython,matthewelse\/micropython,deshipu\/micropython,bvernoux\/micropython,Timmenem\/micropython,cwyark\/micropython,alex-robbins\/micropython,TDAbboud\/micropython,henriknelson\/micropython,adafruit\/micropython,oopy\/micropython,alex-march\/micropython,dmazzella\/micropython,MrSurly\/micropython-esp32,turbinenreiter\/micropython,kerneltask\/micropython,PappaPeppar\/micropython,pozetroninc\/micropython,pozetroninc\/micropython,tobbad\/micropython,blazewicz\/micropython,swegener\/micropython,tuc-osg\/micropython,trezor\/micropython,alex-robbins\/micropython,torwag\/micropython,dxxb\/micropython,adafruit\/micropython,HenrikSolver\/micropython,cwyark\/micropython,infinnovation\/micropython,HenrikSolver\/micropython,pfalcon\/micropython,adafruit\/micropython,infinnovation\/micropython,ryannathans\/micropython,henriknelson\/micropython,toolmacher\/micropython,MrSurly\/micropython-esp32,lowRISC\/micropython,AriZuu\/micropython,Peetz0r\/micropython-esp32,alex-march\/micropython,SHA2017-badge\/micropython-esp32,pramasoul\/micropython,tuc-osg\/micropython,infinnovation\/micropython,puuu\/micropython,jmarcelino\/pycom-micropython,trezor\/micropython,pozetroninc\/micropython,swegener\/micropython,ryannathans\/micropython"} {"commit":"63ee6f971b99c2f030e0347c37bc9577ba9ee7cd","old_file":"getMenu.py","new_file":"getMenu.py","old_contents":"#!\/usr\/bin\/env python\n\nimport json, os, requests\nfrom awsauth import S3Auth\n\nkey = os.environ.get('UWOPENDATA_APIKEY')\n\nservice = 'FoodMenu'\n# output = 'json'\n# callback = 'None'\nrequest = 'http:\/\/api.uwaterloo.ca\/public\/v1\/'\n\ndef getMenu():\n\turl = request + '?' + 'key=' + key + '&' + 'service=' + service\n\tr = requests.get(url).text\n\treturn r\n\nmenu = getMenu()\nACCESS_KEY = os.environ.get('AWS_ACCESS_KEY_ID')\nSECRET_KEY = os.environ.get('AWS_SECRET_ACCESS_KEY')\nrequests.put('http:\/\/s3.amazonaws.com\/uwfoodmenu\/response.txt', data=menu, auth=S3Auth(ACCESS_KEY, SECRET_KEY))\n","new_contents":"#!\/usr\/bin\/env python\nimport json, os, requests\nfrom awsauth import S3Auth\n\nkey = os.environ.get('UWOPENDATA_APIKEY')\nservice = 'FoodMenu'\n\ndef getMenu():\n\tpayload = {'key': key, 'service': service}\n\tr = requests.get('http:\/\/api.uwaterloo.ca\/public\/v1\/', params=payload)\n\treturn r.text\n\nmenu = getMenu()\nACCESS_KEY = os.environ.get('AWS_ACCESS_KEY_ID')\nSECRET_KEY = os.environ.get('AWS_SECRET_ACCESS_KEY')\nrequests.put('http:\/\/s3.amazonaws.com\/uwfoodmenu\/response.txt', data=menu, auth=S3Auth(ACCESS_KEY, SECRET_KEY))\n","subject":"Allow requests module to correctly encode query parameters.","message":"Allow requests module to correctly encode query parameters.\n","lang":"Python","license":"mit","repos":"alykhank\/FoodMenu,alykhank\/FoodMenu,alykhank\/FoodMenu"} {"commit":"98a05257eaf4ca6555ffc179a9250a7cfb3a903c","old_file":"scripts\/lib\/check-database-compatibility.py","new_file":"scripts\/lib\/check-database-compatibility.py","old_contents":"#!\/usr\/bin\/env python3\nimport logging\nimport os\nimport sys\n\nZULIP_PATH = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))\nsys.path.insert(0, ZULIP_PATH)\nfrom scripts.lib.setup_path import setup_path\nfrom scripts.lib.zulip_tools import DEPLOYMENTS_DIR, assert_not_running_as_root, parse_version_from\nfrom version import ZULIP_VERSION as new_version\n\nassert_not_running_as_root()\nsetup_path()\nos.environ[\"DJANGO_SETTINGS_MODULE\"] = \"zproject.settings\"\n\nimport django\nfrom django.db import connection\nfrom django.db.migrations.loader import MigrationLoader\n\ndjango.setup()\nloader = MigrationLoader(connection)\nmissing = set(loader.applied_migrations)\nfor key, migration in loader.disk_migrations.items():\n missing.discard(key)\n missing.difference_update(migration.replaces)\nif not missing:\n sys.exit(0)\n\ncurrent_version = parse_version_from(os.path.join(DEPLOYMENTS_DIR, \"current\"))\nlogging.error(\n \"This is not an upgrade -- the current deployment (version %s) \"\n \"contains %s database migrations which %s (version %s) does not.\",\n current_version,\n len(missing),\n ZULIP_PATH,\n new_version,\n)\nsys.exit(1)\n","new_contents":"#!\/usr\/bin\/env python3\nimport logging\nimport os\nimport sys\n\nZULIP_PATH = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))\nsys.path.insert(0, ZULIP_PATH)\nfrom scripts.lib.setup_path import setup_path\nfrom scripts.lib.zulip_tools import DEPLOYMENTS_DIR, assert_not_running_as_root, parse_version_from\nfrom version import ZULIP_VERSION as new_version\n\nassert_not_running_as_root()\nsetup_path()\nos.environ[\"DJANGO_SETTINGS_MODULE\"] = \"zproject.settings\"\n\nimport django\nfrom django.db import connection\nfrom django.db.migrations.loader import MigrationLoader\n\ndjango.setup()\nloader = MigrationLoader(connection)\nmissing = set(loader.applied_migrations)\nfor key, migration in loader.disk_migrations.items():\n missing.discard(key)\n missing.difference_update(migration.replaces)\nif not missing:\n sys.exit(0)\n\nfor migration in missing:\n print(f\"Migration {migration} missing in new version.\")\n\ncurrent_version = parse_version_from(os.path.join(DEPLOYMENTS_DIR, \"current\"))\nlogging.error(\n \"This is not an upgrade -- the current deployment (version %s) \"\n \"contains %s database migrations which %s (version %s) does not.\",\n current_version,\n len(missing),\n ZULIP_PATH,\n new_version,\n)\nsys.exit(1)\n","subject":"Print names of missing migrations in compatibility check.","message":"scripts: Print names of missing migrations in compatibility check.\n\nThis will make it much easier to debug any situations where this\nhappens.\n","lang":"Python","license":"apache-2.0","repos":"rht\/zulip,rht\/zulip,rht\/zulip,andersk\/zulip,zulip\/zulip,zulip\/zulip,andersk\/zulip,kou\/zulip,andersk\/zulip,kou\/zulip,zulip\/zulip,rht\/zulip,kou\/zulip,andersk\/zulip,rht\/zulip,kou\/zulip,andersk\/zulip,rht\/zulip,zulip\/zulip,kou\/zulip,kou\/zulip,zulip\/zulip,kou\/zulip,andersk\/zulip,zulip\/zulip,andersk\/zulip,rht\/zulip,zulip\/zulip"} {"commit":"cbafc968343cd2b001bcee354d418c9886fe94b4","old_file":"tests\/test_network.py","new_file":"tests\/test_network.py","old_contents":"from nose.tools import eq_, ok_\nimport unittest\n\nimport openxc.measurements\nfrom openxc.sources import NetworkDataSource\nfrom openxc.sources import DataSourceError\n\nclass NetworkDataSourceTests(unittest.TestCase):\n def setUp(self):\n super(NetworkDataSourceTests, self).setUp()\n\n def test_create(self):\n def callback(message):\n pass\n\n try:\n s = NetworkDataSource(callback)\n except DataSourceError as e:\n pass\n","new_contents":"from nose.tools import eq_, ok_\nimport unittest\n\nimport openxc.measurements\nfrom openxc.sources import NetworkDataSource\nfrom openxc.sources import DataSourceError\n\nclass NetworkDataSourceTests(unittest.TestCase):\n def setUp(self):\n super(NetworkDataSourceTests, self).setUp()\n\n def test_create(self):\n def callback(message):\n pass\n\n try:\n s = NetworkDataSource(callback, host='localhost')\n except DataSourceError as e:\n pass\n","subject":"Use localhost for network source tests to avoid waiting for DNS.","message":"Use localhost for network source tests to avoid waiting for DNS.\n","lang":"Python","license":"bsd-3-clause","repos":"openxc\/openxc-python,openxc\/openxc-python,openxc\/openxc-python"} {"commit":"f34c80578fce1e461c6746a1a815e84b6517e382","old_file":"version_st2.py","new_file":"version_st2.py","old_contents":"# Copyright 2016 - StackStorm, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n__version__ = '1.6dev'\n","new_contents":"# Copyright 2016 - StackStorm, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n__version__ = '1.7dev'\n","subject":"Update version info for development - 1.7dev","message":"Update version info for development - 1.7dev\n","lang":"Python","license":"apache-2.0","repos":"StackStorm\/mistral,StackStorm\/mistral"} {"commit":"86d12c7d13bd7a11a93deccf42f93df4328e70fd","old_file":"admin_honeypot\/urls.py","new_file":"admin_honeypot\/urls.py","old_contents":"from admin_honeypot import views\nfrom django.conf.urls import url\n\napp_name = 'admin_honeypot'\n\nurlpatterns = [\n url(r'^login\/$', views.AdminHoneypot.as_view(), name='login'),\n url(r'^.*$', views.AdminHoneypot.as_view(), name='index'),\n]\n","new_contents":"from admin_honeypot import views\nfrom django.urls import path, re_path\n\napp_name = 'admin_honeypot'\n\nurlpatterns = [\n path('login\/', views.AdminHoneypot.as_view(), name='login'),\n re_path(r'^.*$', views.AdminHoneypot.as_view(), name='index'),\n]\n","subject":"Update url() to path() in the urlconf.","message":"Update url() to path() in the urlconf.\n","lang":"Python","license":"mit","repos":"dmpayton\/django-admin-honeypot,dmpayton\/django-admin-honeypot"} {"commit":"1726a73b81c8a7dfc3610690fe9272776e930f0f","old_file":"aero\/adapters\/bower.py","new_file":"aero\/adapters\/bower.py","old_contents":"# -*- coding: utf-8 -*-\n__author__ = 'oliveiraev'\n__all__ = ['Bower']\n\nfrom re import sub\nfrom re import split\nfrom aero.__version__ import enc\nfrom .base import BaseAdapter\n\n\nclass Bower(BaseAdapter):\n \"\"\"\n Twitter Bower - Browser package manager - Adapter\n \"\"\"\n def search(self, query):\n return {}\n\n response = self.command('search', query, ['--no-color'])[0].decode(*enc)\n lst = dict([(self.package_name(k), v) for k, v in [\n line.lstrip(' -').split(' ') for line in response.splitlines()\n if line.startswith(' - ')]\n ])\n if lst:\n return lst\n\n def install(self, query):\n return self.shell('install', query)\n\n def info(self, query):\n response = self.command('info', query, ['--no-color'])[0].decode(*enc)\n return response or ['Aborted: No info available']\n","new_contents":"# -*- coding: utf-8 -*-\n__author__ = 'oliveiraev'\n__all__ = ['Bower']\n\nfrom re import sub\nfrom re import split\nfrom aero.__version__ import enc\nfrom .base import BaseAdapter\n\n\nclass Bower(BaseAdapter):\n \"\"\"\n Twitter Bower - Browser package manager - Adapter\n \"\"\"\n def search(self, query):\n response = self.command('search', query, ['--no-color'])[0].decode(*enc)\n lst = dict([(self.package_name(k), v) for k, v in [\n line.lstrip(' -').split(' ') for line in response.splitlines()\n if line.startswith(' - ')]\n ])\n return lst or {}\n\n def install(self, query):\n return self.shell('install', query)\n\n def info(self, query):\n response = self.command('info', query, ['--no-color'])[0].decode(*enc)\n return response or ['Aborted: No info available']\n","subject":"Simplify return while we're at it","message":"Simplify return while we're at it\n","lang":"Python","license":"bsd-3-clause","repos":"Aeronautics\/aero"} {"commit":"19964dc65cecbbb043da3fe85bf355423cf9ce3c","old_file":"shop\/products\/admin\/forms.py","new_file":"shop\/products\/admin\/forms.py","old_contents":"\nfrom django.apps import apps\n\nfrom django import forms\n\nfrom suit.sortables import SortableTabularInline\nfrom multiupload.fields import MultiFileField\n\n\nclass ProductForm(forms.ModelForm):\n\n images = MultiFileField(max_num=100, min_num=1, required=False)\n\n class Meta:\n model = apps.get_model('products', 'Product')\n fields = '__all__'\n\n\nclass ProductImageInline(SortableTabularInline):\n fields = ('preview', )\n readonly_fields = ['preview']\n model = apps.get_model('products', 'ProductImage')\n extra = 0\n max_num = 0\n\n","new_contents":"\nfrom django.apps import apps\n\nfrom django import forms\n\nfrom suit.sortables import SortableTabularInline\nfrom multiupload.fields import MultiFileField\n\n\nclass ProductForm(forms.ModelForm):\n\n images = MultiFileField(max_num=100, min_num=1, required=False)\n\n def save(self, commit=True):\n product = super(ProductForm, self).save(commit)\n\n if 'category' in self.changed_data:\n product.attribute_values.all().delete()\n\n return product\n\n class Meta:\n model = apps.get_model('products', 'Product')\n fields = '__all__'\n\n\nclass ProductImageInline(SortableTabularInline):\n fields = ('preview', )\n readonly_fields = ['preview']\n model = apps.get_model('products', 'ProductImage')\n extra = 0\n max_num = 0\n\n","subject":"Clear attr values on category change.","message":"Clear attr values on category change.\n","lang":"Python","license":"isc","repos":"pmaigutyak\/mp-shop,pmaigutyak\/mp-shop,pmaigutyak\/mp-shop"} {"commit":"bd7c6e22146604183412657e68457db7ae7766ed","old_file":"script\/jsonify-book.py","new_file":"script\/jsonify-book.py","old_contents":"import sys\nfrom glob import glob\nfrom os.path import basename\nimport json\n\nbook_dir, out_dir = sys.argv[1:3]\n\nfiles = [basename(x).rstrip(\".xhtml\") for x in glob(f\"{book_dir}\/*.xhtml\")]\n\njson_data = {}\n\nfor path in files:\n with open(f\"{book_dir}\/{path}.xhtml\", \"r\") as book_part:\n content = book_part.read()\n json_data = { \"content\": str(content) }\n\n with open(f\"{out_dir}\/{path}.json\", 'w') as outfile:\n json.dump(json_data, outfile)\n","new_contents":"import sys\nfrom glob import glob\nfrom os.path import basename\nimport json\n\nbook_dir, out_dir = sys.argv[1:3]\n\nfiles = [basename(x).rstrip(\".xhtml\") for x in glob(f\"{book_dir}\/*.xhtml\")]\n\njson_data = {}\n\nfor path in files:\n with open(f\"{book_dir}\/{path}.json\", \"r\") as meta_part:\n json_data = json.load(meta_part)\n\n with open(f\"{book_dir}\/{path}.xhtml\", \"r\") as book_part:\n content = book_part.read()\n json_data[\"content\"] = str(content)\n\n with open(f\"{out_dir}\/{path}.json\", 'w') as outfile:\n json.dump(json_data, outfile)","subject":"Add metadata to jsonify output","message":"Add metadata to jsonify output\n","lang":"Python","license":"lgpl-2.1","repos":"Connexions\/cte,Connexions\/cnx-recipes,Connexions\/cnx-rulesets,Connexions\/cnx-recipes,Connexions\/cnx-rulesets,Connexions\/cnx-rulesets,Connexions\/cnx-rulesets,Connexions\/cnx-recipes,Connexions\/cnx-recipes,Connexions\/cte,Connexions\/cnx-recipes"} {"commit":"b9e3485030ef7acf5b3d312b8e9d9fc54367eded","old_file":"tests\/ext\/argcomplete_tests.py","new_file":"tests\/ext\/argcomplete_tests.py","old_contents":"\"\"\"Tests for cement.ext.ext_argcomplete.\"\"\"\n\nimport os\nfrom cement.ext import ext_argcomplete\nfrom cement.ext.ext_argparse import ArgparseController, expose\nfrom cement.utils import test\nfrom cement.utils.misc import rando\n\nAPP = rando()[:12]\n\nclass MyBaseController(ArgparseController):\n class Meta:\n label = 'base'\n\n @expose()\n def default(self):\n pass\n\nclass ArgcompleteExtTestCase(test.CementExtTestCase):\n\n def setUp(self):\n super(ArgcompleteExtTestCase, self).setUp()\n self.app = self.make_app(APP, \n base_controller=MyBaseController,\n extensions=[\n 'argparse',\n 'argcomplete'\n ],\n )\n\n def test_argcomplete(self):\n # not really sure how to test this for reals... but let's atleast get\n # coverage\n with self.app as app:\n app.run()\n","new_contents":"\"\"\"Tests for cement.ext.ext_argcomplete.\"\"\"\n\nimport os\nfrom cement.ext import ext_argcomplete\nfrom cement.ext.ext_argparse import ArgparseController, expose\nfrom cement.utils import test\nfrom cement.utils.misc import rando\n\nAPP = rando()[:12]\n\nclass MyBaseController(ArgparseController):\n class Meta:\n label = 'base'\n\n @expose()\n def default(self):\n pass\n\nclass ArgcompleteExtTestCase(test.CementExtTestCase):\n\n def setUp(self):\n super(ArgcompleteExtTestCase, self).setUp()\n self.app = self.make_app(APP, \n argv=['default'],\n base_controller=MyBaseController,\n extensions=[\n 'argparse',\n 'argcomplete'\n ],\n )\n\n def test_argcomplete(self):\n # not really sure how to test this for reals... but let's atleast get\n # coverage\n with self.app as app:\n app.run()\n","subject":"Fix Argcomplete Tests on Python <3.2","message":"Fix Argcomplete Tests on Python <3.2\n","lang":"Python","license":"bsd-3-clause","repos":"akhilman\/cement,fxstein\/cement,datafolklabs\/cement,akhilman\/cement,akhilman\/cement,fxstein\/cement,fxstein\/cement,datafolklabs\/cement,datafolklabs\/cement"} {"commit":"11b16c26c182636016e7d86cd0f94963eec42556","old_file":"project\/settings\/ci.py","new_file":"project\/settings\/ci.py","old_contents":"# Local\nfrom .base import *\n\n\n# JWT Settings\ndef jwt_get_username_from_payload_handler(payload):\n return payload.get('email')\n\nJWT_AUTH = {\n # 'JWT_SECRET_KEY': AUTH0_CLIENT_SECRET,\n 'JWT_AUDIENCE': AUTH0_CLIENT_ID,\n 'JWT_PAYLOAD_GET_USERNAME_HANDLER': jwt_get_username_from_payload_handler,\n 'JWT_AUTH_HEADER_PREFIX': 'Bearer',\n 'JWT_PUBLIC_KEY': jwt_public_key,\n 'JWT_ALGORITHM': 'RS256',\n}\n\nDATABASES['default']['TEST'] = dj_database_url.config(default=DATABASE_URL)\n\n# Static Server Config\nSTATIC_ROOT = os.path.join(PROJECT_ROOT, 'static')\nSTATIC_STORAGE = 'django.contrib.staticfiles.storage.StaticFilesStorage'\nSTATIC_URL = '\/static\/'\nSTATICFILES_STORAGE = STATIC_STORAGE\n\n# Media (aka File Upload) Server Config\nMEDIA_ROOT = os.path.join(PROJECT_ROOT, 'media')\nMEDIA_STORAGE = 'django.core.files.storage.FileSystemStorage'\nMEDIA_URL = '\/media\/'\nDEFAULT_FILE_STORAGE = MEDIA_STORAGE\n\n# Email\nEMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'\n","new_contents":"# Local\nfrom .base import *\n\n\n# JWT Settings\ndef jwt_get_username_from_payload_handler(payload):\n return payload.get('email')\n\nJWT_AUTH = {\n # 'JWT_SECRET_KEY': AUTH0_CLIENT_SECRET,\n 'JWT_AUDIENCE': AUTH0_CLIENT_ID,\n 'JWT_PAYLOAD_GET_USERNAME_HANDLER': jwt_get_username_from_payload_handler,\n 'JWT_AUTH_HEADER_PREFIX': 'Bearer',\n 'JWT_PUBLIC_KEY': jwt_public_key,\n 'JWT_ALGORITHM': 'RS256',\n}\n\n\n# Static Server Config\nSTATIC_ROOT = os.path.join(PROJECT_ROOT, 'static')\nSTATIC_STORAGE = 'django.contrib.staticfiles.storage.StaticFilesStorage'\nSTATIC_URL = '\/static\/'\nSTATICFILES_STORAGE = STATIC_STORAGE\n\n# Media (aka File Upload) Server Config\nMEDIA_ROOT = os.path.join(PROJECT_ROOT, 'media')\nMEDIA_STORAGE = 'django.core.files.storage.FileSystemStorage'\nMEDIA_URL = '\/media\/'\nDEFAULT_FILE_STORAGE = MEDIA_STORAGE\n\n# Email\nEMAIL_BACKEND = 'django.core.mail.backends.console.EmailBackend'\n","subject":"Revert \"Attempt to bypass test database\"","message":"Revert \"Attempt to bypass test database\"\n\nThis reverts commit 889713c8c4c7151ba06448a3993778a91d2abfd6.\n","lang":"Python","license":"bsd-2-clause","repos":"barberscore\/barberscore-api,dbinetti\/barberscore-django,dbinetti\/barberscore-django,barberscore\/barberscore-api,barberscore\/barberscore-api,dbinetti\/barberscore,barberscore\/barberscore-api,dbinetti\/barberscore"} {"commit":"77f99f4862ded1b8493b5895e4f9d88a3bbf722b","old_file":"source\/globals\/fieldtests.py","new_file":"source\/globals\/fieldtests.py","old_contents":"# -*- coding: utf-8 -*-\n\n## \\package globals.fieldtests\n\n# MIT licensing\n# See: LICENSE.txt\n\n\nimport wx\n\n\n## Tests if a wx control\/instance is enabled\n# \n# Function for compatibility between wx versions\n# \\param enabled\n# \\b \\e bool : Check if enabled or disabled\ndef FieldEnabled(field, enabled=True):\n if wx.MAJOR_VERSION > 2:\n return field.IsThisEnabled() == enabled\n \n else:\n return field.IsEnabled() == enabled\n\n\n## Tests multiple fields\n# \n# \\return\n# \\b \\e bool : True if all fields are enabled\ndef FieldsEnabled(field_list):\n if isinstance(field_list, (tuple, list)):\n return FieldEnabled(field_list)\n \n for F in field_list:\n if not FieldEnabled(F):\n return False\n \n return True\n","new_contents":"# -*- coding: utf-8 -*-\n\n## \\package globals.fieldtests\n\n# MIT licensing\n# See: LICENSE.txt\n\n\nimport wx\n\n\n## Tests if a wx control\/instance is enabled\/disabled\n# \n# Function for compatibility between wx versions\n# \\param field\n# \\b \\e wx.Window : the wx control to check\n# \\param enabled\n# \\b \\e bool : Check if enabled or disabled\n# \\return\n# \\b \\e bool : True if field's enabled status is same as 'enabled'\ndef FieldEnabled(field, enabled=True):\n if wx.MAJOR_VERSION > 2:\n return field.IsThisEnabled() == enabled\n \n else:\n return field.IsEnabled() == enabled\n\n\n## Tests if a wx control\/instance is disabled\n# \n# \\param field\n# \\b \\e wx.Window : The wx field to check\n# \\return\n# \\b \\e : True if field is disabled\ndef FieldDisabled(field):\n return FieldEnabled(field, False)\n\n\n## Tests multiple fields\n# \n# \\return\n# \\b \\e bool : True if all fields are enabled\ndef FieldsEnabled(field_list):\n if isinstance(field_list, (tuple, list)):\n return FieldEnabled(field_list)\n \n for F in field_list:\n if not FieldEnabled(F):\n return False\n \n return True\n","subject":"Add function FieldDisabled to test for disabled controls","message":"Add function FieldDisabled to test for disabled controls","lang":"Python","license":"mit","repos":"AntumDeluge\/desktop_recorder,AntumDeluge\/desktop_recorder"} {"commit":"6a2fb450eb51d46fe4ab53dd4095527ecdcc9266","old_file":"tests\/laundry_test.py","new_file":"tests\/laundry_test.py","old_contents":"import unittest\nfrom penn import Laundry\n\n\nclass TestLaundry(unittest.TestCase):\n\n def setUp(self):\n self.laundry = Laundry()\n\n def test_all(self):\n data = self.laundry.all_status()\n self.assertEquals('Class of 1925 House', data[0]['name'])\n self.assertEquals(55, len(data))\n\n def test_single_hall(self):\n for i in range(5):\n data = self.laundry.hall_status(i)\n self.assertEquals(data['machines'][0]['number'], '1')\n","new_contents":"from nose.tools import ok_, eq_\nfrom penn import Laundry\n\n\nclass TestLaundry():\n\n def setUp(self):\n self.laundry = Laundry()\n\n def test_all(self):\n data = self.laundry.all_status()\n eq_(55, len(data))\n eq_('Class of 1925 House', data[0]['name'])\n # Check all halls have appropriate data points\n for i, hall in enumerate(data):\n eq_(hall['hall_no'], i)\n ok_(hall['dryers_available'] >= 0)\n ok_(hall['dryers_in_use'] >= 0)\n ok_(hall['washers_available'] >= 0)\n ok_(hall['washers_in_use'] >= 0)\n\n def test_single_hall(self):\n for i in range(1):\n data = self.laundry.hall_status(i)\n machines = data['machines']\n # Check all machines have appropriate data points\n for i, machine in enumerate(machines):\n eq_(machine['number'], str(i + 1))\n ok_('available' in machine)\n ok_('machine_type' in machine)\n ok_('time_left' in machine)\n","subject":"Add more rigorous laundry tests","message":"Add more rigorous laundry tests\n","lang":"Python","license":"mit","repos":"pennlabs\/penn-sdk-python,pennlabs\/penn-sdk-python"} {"commit":"589dac7bf0305ec1289b2f81fe8c03cb61260238","old_file":"tools\/boilerplate_data\/init.py","new_file":"tools\/boilerplate_data\/init.py","old_contents":"<%inherit file=\"layout.py\"\/>\nfrom .backend import ${r.name}Backend\n\n\n__all__ = ['${r.name}Backend']\n","new_contents":"<%inherit file=\"layout.py\"\/>\nfrom .backend import ${r.classname}Backend\n\n\n__all__ = ['${r.classname}Backend']\n","subject":"Fix missing use of the class name","message":"boilerplate: Fix missing use of the class name\n","lang":"Python","license":"agpl-3.0","repos":"sputnick-dev\/weboob,RouxRC\/weboob,sputnick-dev\/weboob,willprice\/weboob,laurent-george\/weboob,Konubinix\/weboob,RouxRC\/weboob,Boussadia\/weboob,RouxRC\/weboob,yannrouillard\/weboob,frankrousseau\/weboob,yannrouillard\/weboob,franek\/weboob,Boussadia\/weboob,laurent-george\/weboob,Konubinix\/weboob,Boussadia\/weboob,laurent-george\/weboob,willprice\/weboob,franek\/weboob,franek\/weboob,Boussadia\/weboob,sputnick-dev\/weboob,yannrouillard\/weboob,nojhan\/weboob-devel,nojhan\/weboob-devel,Konubinix\/weboob,frankrousseau\/weboob,nojhan\/weboob-devel,frankrousseau\/weboob,willprice\/weboob"} {"commit":"9bb19e21ed7f3b10af9a218cf55ea3a19ee4393c","old_file":"tests\/test_command.py","new_file":"tests\/test_command.py","old_contents":"\"\"\"Unittest of command entry point.\"\"\"\n# Copyright 2015 Masayuki Yamamoto\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport unittest\nimport unittest.mock as mock\n\nimport yanico\nimport yanico.command\n\n\nclass TestCreateMainParser(unittest.TestCase):\n \"\"\"yanico.command.create_main_parser() test.\"\"\"\n\n def test_version(self):\n \"\"\"Parse '--version' option.\"\"\"\n parser = yanico.command.create_main_parser()\n with mock.patch.object(parser, '_print_message') as print_message:\n self.assertRaises(SystemExit, parser.parse_args, ['--version'])\n print_message.assert_called_once_with('yanico version ' +\n yanico.__version__ + '\\n',\n mock.ANY)\n","new_contents":"\"\"\"Unittest of command entry point.\"\"\"\n# Copyright 2015 Masayuki Yamamoto\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport unittest\nimport unittest.mock as mock\n\nimport yanico\nimport yanico.command\n\n\nclass TestCreateMainParser(unittest.TestCase):\n \"\"\"yanico.command.create_main_parser() test.\"\"\"\n\n def test_version(self):\n \"\"\"Parse '--version' option.\"\"\"\n parser = yanico.command.create_main_parser()\n with mock.patch.object(parser, '_print_message') as print_message:\n self.assertRaises(SystemExit, parser.parse_args, ['--version'])\n print_message.assert_called_once_with('yanico version ' +\n yanico.__version__ + '\\n',\n mock.ANY)\n\n def test_help_long(self):\n \"\"\"Parse '--help' option.\"\"\"\n parser = yanico.command.create_main_parser()\n with mock.patch.object(parser, 'print_help') as print_help:\n self.assertRaises(SystemExit, parser.parse_args, ['--help'])\n print_help.assert_called_once_with()\n","subject":"Add command test for '--help' option","message":"Add command test for '--help' option\n\nCheck calling 'print_help' method.\n","lang":"Python","license":"apache-2.0","repos":"ma8ma\/yanico"} {"commit":"01920b5dcced36e72a5623bf9c08c5cecfa38753","old_file":"src\/scrapy_redis\/dupefilter.py","new_file":"src\/scrapy_redis\/dupefilter.py","old_contents":"import time\n\nfrom scrapy.dupefilters import BaseDupeFilter\nfrom scrapy.utils.request import request_fingerprint\n\nfrom . import connection\n\n\nclass RFPDupeFilter(BaseDupeFilter):\n \"\"\"Redis-based request duplication filter\"\"\"\n\n def __init__(self, server, key):\n \"\"\"Initialize duplication filter\n\n Parameters\n ----------\n server : Redis instance\n key : str\n Where to store fingerprints\n \"\"\"\n self.server = server\n self.key = key\n\n @classmethod\n def from_settings(cls, settings):\n server = connection.from_settings(settings)\n # create one-time key. needed to support to use this\n # class as standalone dupefilter with scrapy's default scheduler\n # if scrapy passes spider on open() method this wouldn't be needed\n key = \"dupefilter:%s\" % int(time.time())\n return cls(server, key)\n\n @classmethod\n def from_crawler(cls, crawler):\n return cls.from_settings(crawler.settings)\n\n def request_seen(self, request):\n fp = request_fingerprint(request)\n added = self.server.sadd(self.key, fp)\n return not added\n\n def close(self, reason):\n \"\"\"Delete data on close. Called by scrapy's scheduler\"\"\"\n self.clear()\n\n def clear(self):\n \"\"\"Clears fingerprints data\"\"\"\n self.server.delete(self.key)\n","new_contents":"import time\n\nfrom scrapy.dupefilters import BaseDupeFilter\nfrom scrapy.utils.request import request_fingerprint\n\nfrom . import connection\n\n\nclass RFPDupeFilter(BaseDupeFilter):\n \"\"\"Redis-based request duplication filter\"\"\"\n\n def __init__(self, server, key):\n \"\"\"Initialize duplication filter\n\n Parameters\n ----------\n server : Redis instance\n key : str\n Where to store fingerprints\n \"\"\"\n self.server = server\n self.key = key\n\n @classmethod\n def from_settings(cls, settings):\n server = connection.from_settings(settings)\n # create one-time key. needed to support to use this\n # class as standalone dupefilter with scrapy's default scheduler\n # if scrapy passes spider on open() method this wouldn't be needed\n key = \"dupefilter:%s\" % int(time.time())\n return cls(server, key)\n\n @classmethod\n def from_crawler(cls, crawler):\n return cls.from_settings(crawler.settings)\n\n def request_seen(self, request):\n fp = self.request_fingerprint(request)\n added = self.server.sadd(self.key, fp)\n return not added\n\n def request_fingerprint(self, request):\n return request_fingerprint(request)\n\n def close(self, reason):\n \"\"\"Delete data on close. Called by scrapy's scheduler\"\"\"\n self.clear()\n\n def clear(self):\n \"\"\"Clears fingerprints data\"\"\"\n self.server.delete(self.key)\n","subject":"Allow to override request fingerprint call.","message":"Allow to override request fingerprint call.\n","lang":"Python","license":"mit","repos":"darkrho\/scrapy-redis,rolando\/scrapy-redis"} {"commit":"333afea8d8a548948f24745490c700c98500e22f","old_file":"mlab-ns-simulator\/mlabsim\/lookup.py","new_file":"mlab-ns-simulator\/mlabsim\/lookup.py","old_contents":"\"\"\"\nThis simulates the mlab-ns lookup request, whose code lives here:\n\nhttps:\/\/code.google.com\/p\/m-lab\/source\/browse\/server\/mlabns\/handlers\/lookup.py?repo=ns\n\nThe difference in this module is that we don't support features which\nooni-support does not use and we augment features which ooni-support\nwould rely on if mlab-ns were to add those features.\n\nAlso, this is a twisted web server rather than appengine.\n\"\"\"\n\n\nfrom twisted.web import resource\nfrom twisted.web.server import NOT_DONE_YET\n\n\nclass LookupSimulatorResource (resource.Resource):\n def __init__(self, db):\n # FIXME - db is some simple memory structure holding info;\n # the details will solidfy soon. This resource reads from\n # this structure.\n\n resource.Resource.__init__(self)\n self._db = db\n\n def render_GET(self, request):\n # FIXME: This is not implemented yet.\n request.setResponseCode(500, 'NOT IMPLEMENTED')\n request.finish()\n return NOT_DONE_YET\n","new_contents":"\"\"\"\nThis simulates the mlab-ns lookup request, whose code lives here:\n\nhttps:\/\/code.google.com\/p\/m-lab\/source\/browse\/server\/mlabns\/handlers\/lookup.py?repo=ns\n\nThe difference in this module is that we don't support features which\nooni-support does not use and we augment features which ooni-support\nwould rely on if mlab-ns were to add those features.\n\nAlso, this is a twisted web server rather than appengine.\n\"\"\"\n\nimport json\n\nfrom twisted.web import resource\nfrom twisted.web.server import NOT_DONE_YET\n\n\nclass LookupSimulatorResource (resource.Resource):\n def __init__(self, db):\n \"\"\"db is a dict mapping { fqdn -> other_stuff }; inserts come from mlabsim.update.\"\"\"\n resource.Resource.__init__(self)\n self._db = db\n\n def render_GET(self, request):\n if request.args['match'] == ['all'] and request.args.get('format', ['json']) == ['json']:\n request.setResponseCode(200, 'ok')\n request.write(json.dumps(self._db.values(), indent=2, sort_keys=True))\n request.finish()\n else:\n request.setResponseCode(400, 'invalid')\n request.finish()\n return NOT_DONE_YET\n","subject":"Implement the current ``GET \/ooni`` api.","message":"Implement the current ``GET \/ooni`` api.\n","lang":"Python","license":"apache-2.0","repos":"hellais\/ooni-support,m-lab\/ooni-support,m-lab\/ooni-support,hellais\/ooni-support"} {"commit":"a057798f3e54e8d74005df10ba1f7d9b93270787","old_file":"odbc2csv.py","new_file":"odbc2csv.py","old_contents":"import pypyodbc\nimport csv\n\nconn = pypyodbc.connect(\"DSN=HOSS_DB\")\ncur = conn.cursor()\n\ntables = []\n\ncur.execute(\"select * from sys.tables\")\n\nfor row in cur.fetchall():\n tables.append(row[0])\n\nfor table in tables:\n print(table)\n\n cur.execute(\"select * from {}\".format(table))\n\n column_names = []\n\n for d in cur.description:\n column_names.append(d[0])\n\n# file = open(\"{}.csv\".format(table), \"w\", encoding=\"ISO-8859-1\")\n file = open(\"{}.csv\".format(table), \"w\", encoding=\"utf-8\")\n writer = csv.writer(file)\n writer.writerow(column_names)\n\n for row in cur.fetchall():\n writer.writerow(row)\n\n file.close()\n","new_contents":"import pypyodbc\nimport csv\n\nconn = pypyodbc.connect(\"DSN=HOSS_DB\")\ncur = conn.cursor()\n\ntables = []\n\ncur.execute(\"select * from sys.tables\")\n\nfor row in cur.fetchall():\n tables.append(row[0])\n\nfor table in tables:\n print(table)\n\n cur.execute(\"select * from {}\".format(table))\n\n column_names = []\n\n for d in cur.description:\n column_names.append(d[0])\n\n# file = open(\"{}.csv\".format(table), \"w\", encoding=\"ISO-8859-1\")\n file = open(\"{}.csv\".format(table), \"w\", encoding=\"utf-8\")\n writer = csv.writer(file, lineterminator='\\n')\n writer.writerow(column_names)\n\n for row in cur.fetchall():\n writer.writerow(row)\n\n file.close()\n","subject":"Use just newline for file terminator.","message":"Use just newline for file terminator.","lang":"Python","license":"isc","repos":"wablair\/misc_scripts,wablair\/misc_scripts,wablair\/misc_scripts,wablair\/misc_scripts"} {"commit":"9f5418e5b755232e12ea18e85b131dbd04c74587","old_file":"benchmarks_sphere\/paper_jrn_parco_rexi_nonlinear\/postprocessing_pickle.py","new_file":"benchmarks_sphere\/paper_jrn_parco_rexi_nonlinear\/postprocessing_pickle.py","old_contents":"#! \/usr\/bin\/env python3\n\nimport sys\nimport math\nimport glob\n\nfrom sweet.postprocessing.pickle_SphereDataPhysicalDiff import *\n\np = pickle_SphereDataPhysicalDiff(\"_t00000000120.00000000.csv\")\n\n","new_contents":"#! \/usr\/bin\/env python3\n\nimport sys\nimport math\nimport glob\n\nfrom sweet.postprocessing.pickle_SphereDataPhysicalDiff import *\nfrom mule.exec_program import *\n\n# Ugly hack!\n#output, retval = exec_program('ls *benchref*\/*prog_h* | sort | tail -n 1 | sed \"s\/.*prog_h\/\/\"')\n#if retval != 0:\n#\tprint(output)\n#\traise Exception(\"Something went wrong\")\n\n#output = output.replace(\"\\n\", '')\n#output = output.replace(\"\\r\", '')\n\n#p = pickle_SphereDataPhysicalDiff(output)\np = pickle_SphereDataPhysicalDiff()\n","subject":"Make postprocess pickling generic to various reference files","message":"Make postprocess pickling generic to various reference files\n","lang":"Python","license":"mit","repos":"schreiberx\/sweet,schreiberx\/sweet,schreiberx\/sweet,schreiberx\/sweet"} {"commit":"36ea5e58ce97b69bfd0bf3701cbc5936bc59d100","old_file":"install_dotfiles.py","new_file":"install_dotfiles.py","old_contents":"#!\/usr\/bin\/python\n# install_dotfiles\n# This script will build platform-specific dotfiles and create the appropriate symlinks in ~\n\nimport platform\nimport os\n\nsysName = platform.system()\nos.remove('bashrc')\nbashrc = open('bashrc','a')\n\ndef writeSection(fileName, allowComments):\n f = open(fileName,'r')\n for line in f:\n if line.startswith('#'):\n if allowComments:\n bashrc.write(line)\n else:\n bashrc.write(line)\n\nif sysName == 'Linux':\n writeSection('bash_linux',True)\nelif sysName == 'Darwin':\n writeSection('bash_mac',True)\nelse:\n print \"System not supported!\"\n bashrc.close()\n exit(1)\nif os.path.isfile('bash_private'):\n writeSection('bash_private',False)\nwriteSection('bash_common',False)\nbashrc.close()\n\n","new_contents":"#!\/usr\/bin\/python\n# install_dotfiles\n# This script will build platform-specific dotfiles and create the appropriate symlinks in ~\n\nimport platform\nimport os\n\nsysName = platform.system()\nos.remove('bashrc')\nbashrc = open('bashrc','a')\nbashrc.write(\"#!\/bin\/bash\\n\")\nbashrc.write(\"# This file was generated by a script. Do not edit manually!\\n\")\n\ndef writeSection(fileName, allowComments):\n f = open(fileName,'r')\n for line in f:\n if line.startswith('#'):\n if allowComments:\n bashrc.write(line)\n else:\n bashrc.write(line)\n\nif sysName == 'Linux':\n bashrc.write(\"# ~\/.bashrc: executed by bash(1) for non-login shells.\\n\")\n if os.path.isfile('bash_private'):\n writeSection('bash_private',False)\n writeSection('bash_common',False)\n writeSection('bash_linux',True)\nelif sysName == 'Darwin':\n bashrc.write(\"# ~\/.bash_profile: executed by bash(1) for lon-login shells.\\n\")\n writeSection('bash_mac',True)\n if os.path.isfile('bash_private'):\n writeSection('bash_private',False)\n writeSection('bash_common',False)\nelse:\n print \"System not supported!\"\n bashrc.close()\n exit(1)\nbashrc.close()\n\n","subject":"Reorder writing of bashrc body sections.","message":"Reorder writing of bashrc body sections.\n","lang":"Python","license":"mit","repos":"rucker\/dotfiles-manager"} {"commit":"3661edd55553ff2dff27cb102a83d4751e033f2a","old_file":"painter\/management\/commands\/import_cards.py","new_file":"painter\/management\/commands\/import_cards.py","old_contents":"import tablib\nfrom django.core.management.base import BaseCommand\n\nfrom painter.models import Card\n\n\nclass Command(BaseCommand):\n def handle(self, *args, **options):\n dataset = tablib.Dataset()\n","new_contents":"import tablib\nfrom django.core.management.base import BaseCommand\n\nfrom painter.models import Card\n\n\nclass Command(BaseCommand):\n help = ('Clears the database of cards, then fills it with the contents of one or' +\n ' more specified CSV files.')\n\n def add_arguments(self, parser):\n parser.add_argument(\n 'filenames',\n nargs='+',\n type=str,\n help='One or more CSV file names. The extension is optional.',\n )\n\n def handle(self, *args, **options):\n dataset = tablib.Dataset()\n for filename in options['filenames']:\n print(filename)\n","subject":"Add help text and a 'filenames' argument.","message":"Add help text and a 'filenames' argument.\n\n* Make it print the filenames it's receiving for the sake of\n good testing output.\n","lang":"Python","license":"mit","repos":"adam-incuna\/imperial-painter,adam-thomas\/imperial-painter,adam-thomas\/imperial-painter,adam-incuna\/imperial-painter"} {"commit":"b457af108174821965ae8e3ee28eb3d34c0fec06","old_file":"plugins\/GCodeWriter\/__init__.py","new_file":"plugins\/GCodeWriter\/__init__.py","old_contents":"# Copyright (c) 2015 Ultimaker B.V.\n# Cura is released under the terms of the AGPLv3 or higher.\n\nfrom . import GCodeWriter\n\nfrom UM.i18n import i18nCatalog\ncatalog = i18nCatalog(\"cura\")\n\ndef getMetaData():\n return {\n \"plugin\": {\n \"name\": catalog.i18nc(\"@label\", \"GCode Writer\"),\n \"author\": \"Ultimaker\",\n \"version\": \"1.0\",\n \"description\": catalog.i18nc(\"@info:whatsthis\", \"Writes GCode to a file\"),\n \"api\": 2\n },\n\n \"mesh_writer\": {\n \"output\": [{\n \"extension\": \"gcode\",\n \"description\": catalog.i18nc(\"@item:inlistbox\", \"GCode File\"),\n \"mime_type\": \"text\/x-gcode\",\n \"mode\": GCodeWriter.GCodeWriter.OutputMode.TextMode\n }]\n }\n }\n\ndef register(app):\n return { \"mesh_writer\": GCodeWriter.GCodeWriter() }\n","new_contents":"# Copyright (c) 2015 Ultimaker B.V.\n# Cura is released under the terms of the AGPLv3 or higher.\n\nfrom . import GCodeWriter\n\nfrom UM.i18n import i18nCatalog\ncatalog = i18nCatalog(\"cura\")\n\ndef getMetaData():\n return {\n \"plugin\": {\n \"name\": catalog.i18nc(\"@label\", \"GCode Writer\"),\n \"author\": \"Ultimaker\",\n \"version\": \"1.0\",\n \"description\": catalog.i18nc(\"@info:whatsthis\", \"Writes GCode to a file.\"),\n \"api\": 2\n },\n\n \"mesh_writer\": {\n \"output\": [{\n \"extension\": \"gcode\",\n \"description\": catalog.i18nc(\"@item:inlistbox\", \"GCode File\"),\n \"mime_type\": \"text\/x-gcode\",\n \"mode\": GCodeWriter.GCodeWriter.OutputMode.TextMode\n }]\n }\n }\n\ndef register(app):\n return { \"mesh_writer\": GCodeWriter.GCodeWriter() }\n","subject":"Add period at end of plug-in description","message":"Add period at end of plug-in description\n\nAll other plug-in descriptions have that too. So for consistency.\n\nContributes to issue CURA-1190.\n","lang":"Python","license":"agpl-3.0","repos":"fieldOfView\/Cura,fieldOfView\/Cura,senttech\/Cura,ynotstartups\/Wanhao,senttech\/Cura,hmflash\/Cura,totalretribution\/Cura,ynotstartups\/Wanhao,Curahelper\/Cura,totalretribution\/Cura,hmflash\/Cura,Curahelper\/Cura"} {"commit":"6889946ebb1c1559e0e1c7b83e1d7b1d6896e0b0","old_file":"tests\/test_train_dictionary.py","new_file":"tests\/test_train_dictionary.py","old_contents":"import unittest\n\nimport zstd\n\n\nclass TestTrainDictionary(unittest.TestCase):\n def test_no_args(self):\n with self.assertRaises(TypeError):\n zstd.train_dictionary()\n\n def test_bad_args(self):\n with self.assertRaises(TypeError):\n zstd.train_dictionary(8192, u'foo')\n\n with self.assertRaises(ValueError):\n zstd.train_dictionary(8192, [u'foo'])\n\n def test_basic(self):\n samples = []\n for i in range(128):\n samples.append(b'foo' * 64)\n samples.append(b'bar' * 64)\n samples.append(b'foobar' * 64)\n samples.append(b'baz' * 64)\n samples.append(b'foobaz' * 64)\n samples.append(b'bazfoo' * 64)\n\n d = zstd.train_dictionary(8192, samples)\n self.assertLessEqual(len(d), 8192)\n\n dict_id = zstd.dictionary_id(d)\n self.assertIsInstance(dict_id, int)\n","new_contents":"import sys\nimport unittest\n\nimport zstd\n\n\nif sys.version_info[0] >= 3:\n int_type = int\nelse:\n int_type = long\n\n\nclass TestTrainDictionary(unittest.TestCase):\n def test_no_args(self):\n with self.assertRaises(TypeError):\n zstd.train_dictionary()\n\n def test_bad_args(self):\n with self.assertRaises(TypeError):\n zstd.train_dictionary(8192, u'foo')\n\n with self.assertRaises(ValueError):\n zstd.train_dictionary(8192, [u'foo'])\n\n def test_basic(self):\n samples = []\n for i in range(128):\n samples.append(b'foo' * 64)\n samples.append(b'bar' * 64)\n samples.append(b'foobar' * 64)\n samples.append(b'baz' * 64)\n samples.append(b'foobaz' * 64)\n samples.append(b'bazfoo' * 64)\n\n d = zstd.train_dictionary(8192, samples)\n self.assertLessEqual(len(d), 8192)\n\n dict_id = zstd.dictionary_id(d)\n self.assertIsInstance(dict_id, int_type)\n","subject":"Check for appropriate long type on Python 2","message":"Check for appropriate long type on Python 2\n\nThe extension always returns a long, which is not an \"int\" on\nPython 2. Fix the test.","lang":"Python","license":"bsd-3-clause","repos":"terrelln\/python-zstandard,terrelln\/python-zstandard,terrelln\/python-zstandard,indygreg\/python-zstandard,terrelln\/python-zstandard,indygreg\/python-zstandard,indygreg\/python-zstandard,indygreg\/python-zstandard"} {"commit":"47fe1412857dbc251ff89004798d5507b0e70b25","old_file":"boundary\/plugin_get.py","new_file":"boundary\/plugin_get.py","old_contents":"#\n# Copyright 2014-2015 Boundary, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\nfrom boundary import ApiCli\n\nclass PluginGet (ApiCli):\n \n def __init__(self):\n ApiCli.__init__(self)\n self.method = \"GET\"\n self.path=\"v1\/plugins\"\n self.pluginName = None\n \n def addArguments(self):\n ApiCli.addArguments(self)\n self.parser.add_argument('-n', '--plugin-Name', dest='pluginName',action='store',required=True,help='Plugin name')\n \n def getArguments(self):\n '''\n Extracts the specific arguments of this CLI\n '''\n ApiCli.getArguments(self)\n if self.args.pluginName != None:\n self.pluginName = self.args.pluginName\n \n self.path = \"v1\/plugins\/{0}\".format(self.pluginName)\n \n def getDescription(self):\n return \"Get the details of a plugin in a Boundary account\"\n ","new_contents":"#\n# Copyright 2014-2015 Boundary, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\nfrom boundary import ApiCli\n\n\nclass PluginGet(ApiCli):\n def __init__(self):\n ApiCli.__init__(self)\n self.method = \"GET\"\n self.path = \"v1\/plugins\"\n self.pluginName = None\n\n def addArguments(self):\n ApiCli.addArguments(self)\n\n self.parser.add_argument('-n', '--plugin-Name', dest='pluginName', action='store', metavar='plugin_name',\n required=True, help='Plugin name')\n\n def getArguments(self):\n \"\"\"\n Extracts the specific arguments of this CLI\n \"\"\"\n ApiCli.getArguments(self)\n if self.args.pluginName is not None:\n self.pluginName = self.args.pluginName\n\n self.path = \"v1\/plugins\/{0}\".format(self.pluginName)\n\n def getDescription(self):\n return \"Get the details of a plugin in a Boundary account\"\n","subject":"Reformat code to PEP-8 standards","message":"Reformat code to PEP-8 standards\n","lang":"Python","license":"apache-2.0","repos":"wcainboundary\/boundary-api-cli,boundary\/pulse-api-cli,jdgwartney\/pulse-api-cli,wcainboundary\/boundary-api-cli,jdgwartney\/boundary-api-cli,boundary\/boundary-api-cli,jdgwartney\/boundary-api-cli,boundary\/pulse-api-cli,jdgwartney\/pulse-api-cli,boundary\/boundary-api-cli"} {"commit":"27d2cd57337497abb9d106fdb033c26771e481e4","old_file":"rmgpy\/data\/__init__.py","new_file":"rmgpy\/data\/__init__.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n################################################################################\n#\n# RMG - Reaction Mechanism Generator\n#\n# Copyright (c) 2002-2010 Prof. William H. Green (whgreen@mit.edu) and the\n# RMG Team (rmg_dev@mit.edu)\n#\n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the 'Software'),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and\/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED 'AS IS', WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n# DEALINGS IN THE SOFTWARE.\n#\n################################################################################\n\nimport os.path\n\ndef getDatabaseDirectory():\n return os.path.abspath(os.path.join(os.path.dirname(__file__), '..', '..', 'database'))\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\n################################################################################\n#\n# RMG - Reaction Mechanism Generator\n#\n# Copyright (c) 2002-2010 Prof. William H. Green (whgreen@mit.edu) and the\n# RMG Team (rmg_dev@mit.edu)\n#\n# Permission is hereby granted, free of charge, to any person obtaining a\n# copy of this software and associated documentation files (the 'Software'),\n# to deal in the Software without restriction, including without limitation\n# the rights to use, copy, modify, merge, publish, distribute, sublicense,\n# and\/or sell copies of the Software, and to permit persons to whom the\n# Software is furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED 'AS IS', WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING\n# FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER\n# DEALINGS IN THE SOFTWARE.\n#\n################################################################################\n\n","subject":"Remove getDatabaseDirectory() function from rmgpy.data","message":"Remove getDatabaseDirectory() function from rmgpy.data\n\nThis function is not being used anywhere, and also has been\nreplaced by the settings in rmgpy, which searches for and\nsaves a database directory\n","lang":"Python","license":"mit","repos":"pierrelb\/RMG-Py,pierrelb\/RMG-Py,nickvandewiele\/RMG-Py,nickvandewiele\/RMG-Py,nyee\/RMG-Py,nyee\/RMG-Py,chatelak\/RMG-Py,chatelak\/RMG-Py"} {"commit":"f188f2eb81c1310b9862b435a492b4ce6d0fac2d","old_file":"python3\/aniso8601\/resolution.py","new_file":"python3\/aniso8601\/resolution.py","old_contents":"# -*- coding: utf-8 -*-\n\n# This software may be modified and distributed under the terms\n# of the BSD license. See the LICENSE file for details.\n\nfrom enum import Enum\n\nclass DateResolution(Enum):\n Year, Month, Week, Weekday, Day, Ordinal = range(6)\n\nclass TimeResolution(Enum):\n Seconds, Minutes, Hours = range(3)\n","new_contents":"# -*- coding: utf-8 -*-\n\n# This software may be modified and distributed under the terms\n# of the BSD license. See the LICENSE file for details.\n\nclass DateResolution(object):\n Year, Month, Week, Weekday, Day, Ordinal = list(range(6))\n\nclass TimeResolution(object):\n Seconds, Minutes, Hours = list(range(3))\n","subject":"Remove use of enum in Python3","message":"Remove use of enum in Python3\n","lang":"Python","license":"bsd-3-clause","repos":"3stack-software\/python-aniso8601-relativedelta"} {"commit":"dfef23d834ab67acf91dcefd6fe39e089c71fb9a","old_file":"quantized_mesh_tile\/__init__.py","new_file":"quantized_mesh_tile\/__init__.py","old_contents":"","new_contents":"\"\"\"\nThis module provides high level utility functions to encode and decode a terrain tile.\n\nReference\n---------\n\"\"\"\n\nfrom .terrain import TerrainTile\nfrom .topology import TerrainTopology\n\n\ndef encode(geometries, bounds=[], watermask=[], hasLighting=False, gzipped=False):\n \"\"\"\n Function to convert geometries in a quantized-mesh encoded string buffer.\n\n Arguments:\n\n ``geometries``\n\n A list of shapely polygon geometries representing 3 dimensional triangles.\n or\n A list of WKT or WKB Polygons representing 3 dimensional triangles.\n or\n A list of triplet of vertices using the following structure:\n ``(((lon0\/lat0\/height0),(...),(lon2,lat2,height2)),(...))``\n\n ``bounds``\n\n The bounds of the terrain tile. (west, south, east, north)\n If not defined, the bounds will be computed from the provided geometries.\n\n Default is `[]`.\n\n ``hasLighting``\n\n Indicate whether unit vectors should be computed for the lighting extension.\n\n Default is `False`.\n\n ``watermask``\n\n A water mask list (Optional). Adds rendering water effect.\n The water mask list is either one byte, `[0]` for land and `[255]` for\n water, either a list of 256*256 values ranging from 0 to 255.\n Values in the mask are defined from north-to-south and west-to-east.\n Per default no watermask is applied. Note that the water mask effect depends on\n the texture of the raster layer drapped over your terrain.\n\n Default is `[]`.\n\n\n ``gzipped``\n\n Indicate if the tile content is gzipped.\n\n Default is `False`.\n\n \"\"\"\n topology = TerrainTopology(geometries=geometries, hasLighting=hasLighting)\n if len(bounds) == 4:\n west, south, east, north = bounds\n tile = TerrainTile(watermask=watermask,\n west=west, south=south, east=east, north=north, topology=topology)\n else:\n tile = TerrainTile(watermask=watermask, topology=topology)\n return tile.toStringIO(gzipped=gzipped)\n\n\ndef decode(filePath, bounds, hasLighting=False, hasWatermask=False, gzipped=False):\n \"\"\"\n Function to convert a quantized-mesh terrain tile file into a\n :class:`quantized_mesh_tile.terrain.TerrainTile` instance.\n\n Arguments:\n\n ``filePath``\n\n An absolute or relative path to write the terrain tile. (Required)\n \n ``bounds``\n\n The bounds of the terrain tile. (west, south, east, north) (Required).\n\n ``hasLighting``\n\n Indicate whether the tile has the lighting extension.\n\n Default is `False`.\n\n ``hasWatermask``\n\n Indicate whether the tile has the water-mask extension.\n\n Default is `False`.\n\n \"\"\"\n west, south, east, north = bounds\n tile = TerrainTile(west=west, south=south, east=east, north=north)\n tile.fromFile(\n filePath, hasLighting=hasLighting, hasWatermask=hasWatermask, gzipped=gzipped)\n return tile\n","subject":"Add higher level functions encode and decode","message":"Add higher level functions encode and decode\n","lang":"Python","license":"mit","repos":"loicgasser\/quantized-mesh-tile"} {"commit":"b9ba8a929a539f24d674aed7d7ee98b490a6fcd3","old_file":"mopidy\/__init__.py","new_file":"mopidy\/__init__.py","old_contents":"from mopidy import settings as raw_settings\n\ndef get_version():\n return u'0.1.dev'\n\ndef get_mpd_protocol_version():\n return u'0.16.0'\n\nclass SettingsError(Exception):\n pass\n\nclass Settings(object):\n def __getattr__(self, attr):\n if attr.isupper() and not hasattr(raw_settings, attr):\n raise SettingsError(u'Setting \"%s\" is not set.' % attr)\n value = getattr(raw_settings, attr)\n if type(value) != bool and not value:\n raise SettingsError(u'Setting \"%s\" is empty.' % attr)\n return value\n\nsettings = Settings()\n","new_contents":"from mopidy import settings as raw_settings\n\ndef get_version():\n return u'0.1.0a0.dev0'\n\ndef get_mpd_protocol_version():\n return u'0.16.0'\n\nclass SettingsError(Exception):\n pass\n\nclass Settings(object):\n def __getattr__(self, attr):\n if attr.isupper() and not hasattr(raw_settings, attr):\n raise SettingsError(u'Setting \"%s\" is not set.' % attr)\n value = getattr(raw_settings, attr)\n if type(value) != bool and not value:\n raise SettingsError(u'Setting \"%s\" is empty.' % attr)\n return value\n\nsettings = Settings()\n","subject":"Switch to a StrictVersion-compatible version number","message":"Switch to a StrictVersion-compatible version number\n","lang":"Python","license":"apache-2.0","repos":"mokieyue\/mopidy,swak\/mopidy,pacificIT\/mopidy,quartz55\/mopidy,swak\/mopidy,SuperStarPL\/mopidy,mopidy\/mopidy,priestd09\/mopidy,jmarsik\/mopidy,diandiankan\/mopidy,diandiankan\/mopidy,vrs01\/mopidy,glogiotatidis\/mopidy,adamcik\/mopidy,rawdlite\/mopidy,pacificIT\/mopidy,dbrgn\/mopidy,mokieyue\/mopidy,pacificIT\/mopidy,ali\/mopidy,hkariti\/mopidy,jmarsik\/mopidy,swak\/mopidy,ali\/mopidy,hkariti\/mopidy,tkem\/mopidy,mokieyue\/mopidy,ZenithDK\/mopidy,glogiotatidis\/mopidy,bencevans\/mopidy,glogiotatidis\/mopidy,bencevans\/mopidy,bacontext\/mopidy,ZenithDK\/mopidy,SuperStarPL\/mopidy,ali\/mopidy,hkariti\/mopidy,adamcik\/mopidy,glogiotatidis\/mopidy,rawdlite\/mopidy,abarisain\/mopidy,liamw9534\/mopidy,bencevans\/mopidy,mokieyue\/mopidy,kingosticks\/mopidy,tkem\/mopidy,bacontext\/mopidy,ZenithDK\/mopidy,rawdlite\/mopidy,jmarsik\/mopidy,swak\/mopidy,vrs01\/mopidy,abarisain\/mopidy,dbrgn\/mopidy,dbrgn\/mopidy,priestd09\/mopidy,jmarsik\/mopidy,adamcik\/mopidy,SuperStarPL\/mopidy,bacontext\/mopidy,SuperStarPL\/mopidy,bacontext\/mopidy,jodal\/mopidy,jcass77\/mopidy,tkem\/mopidy,diandiankan\/mopidy,woutervanwijk\/mopidy,mopidy\/mopidy,kingosticks\/mopidy,vrs01\/mopidy,hkariti\/mopidy,quartz55\/mopidy,dbrgn\/mopidy,jcass77\/mopidy,ZenithDK\/mopidy,jodal\/mopidy,priestd09\/mopidy,pacificIT\/mopidy,rawdlite\/mopidy,liamw9534\/mopidy,quartz55\/mopidy,vrs01\/mopidy,bencevans\/mopidy,ali\/mopidy,diandiankan\/mopidy,woutervanwijk\/mopidy,quartz55\/mopidy,jcass77\/mopidy,jodal\/mopidy,mopidy\/mopidy,kingosticks\/mopidy,tkem\/mopidy"} {"commit":"4b5a39c6bbc82572f67ea03236490e52049adf52","old_file":"tests\/query_test\/test_scan_range_lengths.py","new_file":"tests\/query_test\/test_scan_range_lengths.py","old_contents":"#!\/usr\/bin\/env python\n# Copyright (c) 2012 Cloudera, Inc. All rights reserved.\n# Validates running with different scan range length values\n#\nimport pytest\nfrom copy import copy\nfrom tests.common.test_vector import TestDimension\nfrom tests.common.impala_test_suite import ImpalaTestSuite, ALL_NODES_ONLY\n\n# We use very small scan ranges to exercise corner cases in the HDFS scanner more\n# thoroughly. In particular, it will exercise:\n# 1. scan range with no tuple\n# 2. tuple that span across multiple scan ranges\nMAX_SCAN_RANGE_LENGTHS = [1, 2, 5]\n\nclass TestScanRangeLengths(ImpalaTestSuite):\n @classmethod\n def get_workload(cls):\n return 'functional-query'\n\n @classmethod\n def add_test_dimensions(cls):\n super(TestScanRangeLengths, cls).add_test_dimensions()\n cls.TestMatrix.add_dimension(\n TestDimension('max_scan_range_length', *MAX_SCAN_RANGE_LENGTHS))\n\n def test_scan_ranges(self, vector):\n if vector.get_value('table_format').file_format != 'text':\n pytest.xfail(reason='IMP-636')\n elif vector.get_value('table_format').compression_codec != 'none':\n pytest.xfail(reason='IMPALA-122')\n\n vector.get_value('exec_option')['max_scan_range_length'] =\\\n vector.get_value('max_scan_range_length')\n self.run_test_case('QueryTest\/hdfs-tiny-scan', vector)\n","new_contents":"#!\/usr\/bin\/env python\n# Copyright (c) 2012 Cloudera, Inc. All rights reserved.\n# Validates running with different scan range length values\n#\nimport pytest\nfrom copy import copy\nfrom tests.common.test_vector import TestDimension\nfrom tests.common.impala_test_suite import ImpalaTestSuite, ALL_NODES_ONLY\n\n# We use very small scan ranges to exercise corner cases in the HDFS scanner more\n# thoroughly. In particular, it will exercise:\n# 1. scan range with no tuple\n# 2. tuple that span across multiple scan ranges\nMAX_SCAN_RANGE_LENGTHS = [1, 2, 5]\n\nclass TestScanRangeLengths(ImpalaTestSuite):\n @classmethod\n def get_workload(cls):\n return 'functional-query'\n\n @classmethod\n def add_test_dimensions(cls):\n super(TestScanRangeLengths, cls).add_test_dimensions()\n cls.TestMatrix.add_dimension(\n TestDimension('max_scan_range_length', *MAX_SCAN_RANGE_LENGTHS))\n\n def test_scan_ranges(self, vector):\n if vector.get_value('table_format').file_format != 'text':\n pytest.xfail(reason='IMP-636')\n\n vector.get_value('exec_option')['max_scan_range_length'] =\\\n vector.get_value('max_scan_range_length')\n self.run_test_case('QueryTest\/hdfs-tiny-scan', vector)\n","subject":"Fix IMPALA-122: Lzo scanner with small scan ranges.","message":"Fix IMPALA-122: Lzo scanner with small scan ranges.\n\nChange-Id: I5226fd1a1aa368f5b291b78ad371363057ef574e\nReviewed-on: http:\/\/gerrit.ent.cloudera.com:8080\/140\nReviewed-by: Skye Wanderman-Milne <6d4b168ab637b0a20cc9dbf96abb2537f372f946@cloudera.com>\nReviewed-by: Nong Li <99a5e5f8f5911755b88e0b536d46aafa102bed41@cloudera.com>\nTested-by: Nong Li <99a5e5f8f5911755b88e0b536d46aafa102bed41@cloudera.com>\n","lang":"Python","license":"apache-2.0","repos":"tempbottle\/Impala,cchanning\/Impala,kapilrastogi\/Impala,cgvarela\/Impala,brightchen\/Impala,mapr\/impala,caseyching\/Impala,rampage644\/impala-cut,mapr\/impala,caseyching\/Impala,rampage644\/impala-cut,lirui-intel\/Impala,mapr\/impala,ibmsoe\/ImpalaPPC,cchanning\/Impala,gistic\/PublicSpatialImpala,gerashegalov\/Impala,lnliuxing\/Impala,rdblue\/Impala,cgvarela\/Impala,henryr\/Impala,rdblue\/Impala,ImpalaToGo\/ImpalaToGo,ibmsoe\/ImpalaPPC,mapr\/impala,lirui-intel\/Impala,henryr\/Impala,caseyching\/Impala,cloudera\/recordservice,grundprinzip\/Impala,tempbottle\/Impala,cgvarela\/Impala,andybab\/Impala,rdblue\/Impala,bratatidas9\/Impala-1,theyaa\/Impala,andybab\/Impala,placrosse\/ImpalaToGo,grundprinzip\/Impala,theyaa\/Impala,XiaominZhang\/Impala,brightchen\/Impala,lnliuxing\/Impala,cloudera\/recordservice,rampage644\/impala-cut,gistic\/PublicSpatialImpala,gerashegalov\/Impala,lirui-intel\/Impala,ibmsoe\/ImpalaPPC,henryr\/Impala,AtScaleInc\/Impala,bowlofstew\/Impala,kapilrastogi\/Impala,ibmsoe\/ImpalaPPC,caseyching\/Impala,gerashegalov\/Impala,andybab\/Impala,mapr\/impala,rampage644\/impala-cut,lnliuxing\/Impala,cloudera\/recordservice,gerashegalov\/Impala,henryr\/Impala,bowlofstew\/Impala,ImpalaToGo\/ImpalaToGo,cgvarela\/Impala,cgvarela\/Impala,AtScaleInc\/Impala,theyaa\/Impala,caseyching\/Impala,XiaominZhang\/Impala,gistic\/PublicSpatialImpala,scalingdata\/Impala,gistic\/PublicSpatialImpala,rdblue\/Impala,cchanning\/Impala,bowlofstew\/Impala,brightchen\/Impala,grundprinzip\/Impala,lirui-intel\/Impala,bratatidas9\/Impala-1,gistic\/PublicSpatialImpala,lirui-intel\/Impala,placrosse\/ImpalaToGo,tempbottle\/Impala,lnliuxing\/Impala,brightchen\/Impala,bowlofstew\/Impala,placrosse\/ImpalaToGo,cchanning\/Impala,brightchen\/Impala,XiaominZhang\/Impala,AtScaleInc\/Impala,gerashegalov\/Impala,theyaa\/Impala,cloudera\/recordservice,scalingdata\/Impala,XiaominZhang\/Impala,AtScaleInc\/Impala,cloudera\/recordservice,caseyching\/Impala,tempbottle\/Impala,rampage644\/impala-cut,kapilrastogi\/Impala,ImpalaToGo\/ImpalaToGo,XiaominZhang\/Impala,AtScaleInc\/Impala,XiaominZhang\/Impala,bowlofstew\/Impala,bowlofstew\/Impala,gerashegalov\/Impala,andybab\/Impala,bratatidas9\/Impala-1,ImpalaToGo\/ImpalaToGo,placrosse\/ImpalaToGo,ibmsoe\/ImpalaPPC,scalingdata\/Impala,brightchen\/Impala,henryr\/Impala,cchanning\/Impala,brightchen\/Impala,kapilrastogi\/Impala,grundprinzip\/Impala,rdblue\/Impala,theyaa\/Impala,scalingdata\/Impala,cloudera\/recordservice,cgvarela\/Impala,lnliuxing\/Impala,lnliuxing\/Impala,kapilrastogi\/Impala,placrosse\/ImpalaToGo,bowlofstew\/Impala,rampage644\/impala-cut,theyaa\/Impala,cloudera\/recordservice,grundprinzip\/Impala,scalingdata\/Impala,gerashegalov\/Impala,rdblue\/Impala,gistic\/PublicSpatialImpala,henryr\/Impala,ImpalaToGo\/ImpalaToGo,ImpalaToGo\/ImpalaToGo,AtScaleInc\/Impala,ibmsoe\/ImpalaPPC,rdblue\/Impala,bratatidas9\/Impala-1,ibmsoe\/ImpalaPPC,lirui-intel\/Impala,bratatidas9\/Impala-1,cchanning\/Impala,scalingdata\/Impala,placrosse\/ImpalaToGo,tempbottle\/Impala,andybab\/Impala,grundprinzip\/Impala,kapilrastogi\/Impala,caseyching\/Impala,bratatidas9\/Impala-1,tempbottle\/Impala,XiaominZhang\/Impala,theyaa\/Impala,lirui-intel\/Impala,bratatidas9\/Impala-1,cchanning\/Impala,lnliuxing\/Impala,cgvarela\/Impala,kapilrastogi\/Impala,andybab\/Impala,tempbottle\/Impala"} {"commit":"ce875a972eb3efaa5201ba1a72ae1d8d6754cfe0","old_file":"python-pscheduler\/pscheduler\/pscheduler\/db.py","new_file":"python-pscheduler\/pscheduler\/pscheduler\/db.py","old_contents":"\"\"\"\nFunctions for connecting to the pScheduler database\n\"\"\"\n\nimport psycopg2\n\nfrom filestring import string_from_file\n\ndef pg_connection(dsn='', autocommit=True):\n \"\"\"\n Connect to the database, and return a handle to it\n\n Arguments:\n\n dsn - A data source name to use in connecting to the database. If\n the string begins with an '@', the remainder will be treated as\n the path to a file where the value can be retrieved.\n\n autocommit - Whether or not commits are done automatically when\n quesies are issued.\n \"\"\"\n\n dsn = string_from_file(dsn)\n\n pg = psycopg2.connect(dsn)\n if autocommit:\n pg.set_isolation_level(psycopg2.extensions.ISOLATION_LEVEL_AUTOCOMMIT)\n\n return pg\n\n\ndef pg_cursor(dsn='', autocommit=True):\n \"\"\"\n Connect to the database, and return a cursor.\n\n Arguments:\n\n dsn - A data source name to use in connecting to the database. If\n the string begins with an '@', the remainder will be treated as\n the path to a file where the value can be retrieved.\n\n autocommit - Whether or not commits are done automatically when\n quesies are issued.\n \"\"\"\n\n pg = pg_connection(dsn, autocommit)\n return pg.cursor()\n\n\n# TODO: Need a routine that does the select wait currently\n# rubberstamped into the services to do timed waits for notifications.\n","new_contents":"\"\"\"\nFunctions for connecting to the pScheduler database\n\"\"\"\n\nimport os\nimport psycopg2\nimport sys\n\nfrom filestring import string_from_file\n\ndef pg_connection(dsn='', autocommit=True, name=None):\n \"\"\"\n Connect to the database, and return a handle to it\n\n Arguments:\n\n dsn - A data source name to use in connecting to the database. If\n the string begins with an '@', the remainder will be treated as\n the path to a file where the value can be retrieved.\n\n autocommit - Whether or not commits are done automatically when\n quesies are issued.\n \"\"\"\n\n dsn = string_from_file(dsn)\n\n if name is None:\n name = os.path.basename(sys.argv[0])\n\n dsn += \" application_name=%s\" % (name)\n\n pg = psycopg2.connect(dsn)\n if autocommit:\n pg.set_isolation_level(psycopg2.extensions.ISOLATION_LEVEL_AUTOCOMMIT)\n\n return pg\n\n\ndef pg_cursor(dsn='', autocommit=True, name=None):\n \"\"\"\n Connect to the database, and return a cursor.\n\n Arguments:\n\n dsn - A data source name to use in connecting to the database. If\n the string begins with an '@', the remainder will be treated as\n the path to a file where the value can be retrieved.\n\n autocommit - Whether or not commits are done automatically when\n quesies are issued.\n \"\"\"\n\n pg = pg_connection(dsn, autocommit, name)\n return pg.cursor()\n\n\n# TODO: Need a routine that does the select wait currently\n# rubberstamped into the services to do timed waits for notifications.\n","subject":"Add application name to database connection for great debugging.","message":"Add application name to database connection for great debugging.\n","lang":"Python","license":"apache-2.0","repos":"mfeit-internet2\/pscheduler-dev,perfsonar\/pscheduler,perfsonar\/pscheduler,perfsonar\/pscheduler,perfsonar\/pscheduler,mfeit-internet2\/pscheduler-dev"} {"commit":"b34634c0c9a8db389ed48b50ca4b2e4b92105f93","old_file":"node\/dictionary.py","new_file":"node\/dictionary.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom nodes import Node\nimport json\n\nclass Dictionary(Node):\n char = \".d\"\n args = 0\n results = 1\n \n def __init__(self, word_ids:Node.IntList):\n if not hasattr(Dictionary, \"word_list\"):\n Dictionary.word_list = init_words()\n self.words = \" \".join(Dictionary.word_list[i] for i in word_ids)\n \n def func(self):\n return self.words\n \n def compress(inp):\n words = init_words()\n inp_words = [word.lower()for word in inp.split(\" \")]\n rtn = chr(len(inp_words))\n for word in inp_words:\n assert(word in words)\n rtn += chr(words.index(word))\n return rtn\n \ndef init_words(dict_file = \"dictionary.json\"):\n words_f = open(dict_file)\n words = json.load(words_f)\n words_f.close()\n return words\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom nodes import Node\nimport json\n\nclass Dictionary(Node):\n char = \".d\"\n args = 0\n results = 1\n \n def __init__(self, word_ids:Node.IntList):\n if not hasattr(Dictionary, \"word_list\"):\n Dictionary.word_list = init_words()\n self.words = \" \".join(Dictionary.word_list[i] for i in word_ids)\n \n def func(self):\n return self.words\n \n def compress(inp):\n words = init_words()\n inp_words = [word.lower()for word in inp.split(\" \")]\n rtn = chr(len(inp_words))\n for word in inp_words:\n if word not in words:\n rtn += \"Word %s not in wordlist\" % word\n else:\n rtn += chr(words.index(word))\n return rtn\n \ndef init_words(dict_file = \"dictionary.json\"):\n words_f = open(dict_file)\n words = json.load(words_f)\n words_f.close()\n return words\n","subject":"Add some exception handling for dict","message":"Add some exception handling for dict\n","lang":"Python","license":"mit","repos":"muddyfish\/PYKE,muddyfish\/PYKE"} {"commit":"404b7af74fb65299aa9c14e0e40541e3a4a68285","old_file":"setuptools\/command\/bdist_wininst.py","new_file":"setuptools\/command\/bdist_wininst.py","old_contents":"from distutils.command.bdist_wininst import bdist_wininst as _bdist_wininst\n\nclass bdist_wininst(_bdist_wininst):\n def reinitialize_command(self, command, reinit_subcommands=0):\n cmd = self.distribution.reinitialize_command(\n command, reinit_subcommands)\n if command in ('install', 'install_lib'):\n cmd.install_lib = None # work around distutils bug\n return cmd\n\n def run(self):\n self._is_running = True\n try:\n _bdist_wininst.run(self)\n finally:\n self._is_running = False\n","new_contents":"from distutils.command.bdist_wininst import bdist_wininst as _bdist_wininst\n\nclass bdist_wininst(_bdist_wininst):\n def reinitialize_command(self, command, reinit_subcommands=0):\n \"\"\"\n Supplement reinitialize_command to work around\n http:\/\/bugs.python.org\/issue20819\n \"\"\"\n cmd = self.distribution.reinitialize_command(\n command, reinit_subcommands)\n if command in ('install', 'install_lib'):\n cmd.install_lib = None\n return cmd\n\n def run(self):\n self._is_running = True\n try:\n _bdist_wininst.run(self)\n finally:\n self._is_running = False\n","subject":"Update workaround to reference filed ticket.","message":"Update workaround to reference filed ticket.\n","lang":"Python","license":"mit","repos":"pypa\/setuptools,pypa\/setuptools,pypa\/setuptools"} {"commit":"efe06967b4896c7d2d4c88fbda96a0504959594b","old_file":"opps\/core\/admin.py","new_file":"opps\/core\/admin.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nfrom django.contrib import admin\n\n\nclass PublishableAdmin(admin.ModelAdmin):\n \"\"\"\n Overrides standard admin.ModelAdmin save_model method\n It sets user (author) based on data from requet.\n \"\"\"\n def save_model(self, request, obj, form, change):\n if getattr(obj, 'pk', None) is None:\n obj.user = request.user\n obj.save()\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nfrom django.contrib import admin\n\n\nclass PublishableAdmin(admin.ModelAdmin):\n \"\"\"\n Overrides standard admin.ModelAdmin save_model method\n It sets user (author) based on data from requet.\n \"\"\"\n list_display = ['title', 'channel', 'date_available', 'published']\n list_filter = ['date_available', 'published', 'channel']\n search_fields = ['title', 'slug', 'headline', 'channel']\n exclude = ('user',)\n date_hierarchy = ('date_available')\n\n def save_model(self, request, obj, form, change):\n if getattr(obj, 'pk', None) is None:\n obj.user = request.user\n obj.save()\n","subject":"Add basic attr on PublishableAdmin","message":"Add basic attr on PublishableAdmin\n","lang":"Python","license":"mit","repos":"williamroot\/opps,williamroot\/opps,jeanmask\/opps,williamroot\/opps,YACOWS\/opps,jeanmask\/opps,opps\/opps,YACOWS\/opps,jeanmask\/opps,williamroot\/opps,YACOWS\/opps,YACOWS\/opps,opps\/opps,jeanmask\/opps,opps\/opps,opps\/opps"} {"commit":"927172b383e1c62b9aab34f38ef95e96ed277cbe","old_file":"conda_env\/specs\/yaml_file.py","new_file":"conda_env\/specs\/yaml_file.py","old_contents":"from .. import env\nfrom ..exceptions import EnvironmentFileNotFound\n\n\nclass YamlFileSpec(object):\n _environment = None\n\n def __init__(self, filename=None, **kwargs):\n self.filename = filename\n self.msg = None\n\n def can_handle(self):\n try:\n self._environment = env.from_file(self.filename)\n return True\n except EnvironmentFileNotFound, e:\n self.msg = e.message\n return False\n\n @property\n def environment(self):\n if not self._environment:\n self.can_handle()\n return self._environment\n","new_contents":"from .. import env\nfrom ..exceptions import EnvironmentFileNotFound\n\n\nclass YamlFileSpec(object):\n _environment = None\n\n def __init__(self, filename=None, **kwargs):\n self.filename = filename\n self.msg = None\n\n def can_handle(self):\n try:\n self._environment = env.from_file(self.filename)\n return True\n except EnvironmentFileNotFound as e:\n self.msg = e.message\n return False\n\n @property\n def environment(self):\n if not self._environment:\n self.can_handle()\n return self._environment\n","subject":"Update Python 2-style exception handling to 'as'","message":"Update Python 2-style exception handling to 'as'\n","lang":"Python","license":"bsd-3-clause","repos":"isaac-kit\/conda-env,asmeurer\/conda-env,mikecroucher\/conda-env,isaac-kit\/conda-env,conda\/conda-env,ESSS\/conda-env,ESSS\/conda-env,dan-blanchard\/conda-env,nicoddemus\/conda-env,asmeurer\/conda-env,conda\/conda-env,phobson\/conda-env,nicoddemus\/conda-env,mikecroucher\/conda-env,dan-blanchard\/conda-env,phobson\/conda-env"} {"commit":"db751eabb690af3b2b4712f46a41b41c1e0499a2","old_file":"lbrynet\/__init__.py","new_file":"lbrynet\/__init__.py","old_contents":"import logging\n\n__version__ = \"0.17.1\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","new_contents":"import logging\n\n__version__ = \"0.17.2rc1\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","subject":"Bump version 0.17.1 --> 0.17.2rc1","message":"Bump version 0.17.1 --> 0.17.2rc1\n\nSigned-off-by: Jack Robison <40884020c67726395ea162083a125620dc32cdab@lbry.io>\n","lang":"Python","license":"mit","repos":"lbryio\/lbry,lbryio\/lbry,lbryio\/lbry"} {"commit":"b41bb9beee1cc183ef7663efd771323ab6fa8f18","old_file":"lbrynet\/__init__.py","new_file":"lbrynet\/__init__.py","old_contents":"import logging\n\n__version__ = \"0.19.1rc1\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","new_contents":"import logging\n\n__version__ = \"0.19.1rc2\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","subject":"Bump version 0.19.1rc1 --> 0.19.1rc2","message":"Bump version 0.19.1rc1 --> 0.19.1rc2\n\nSigned-off-by: Jack Robison <40884020c67726395ea162083a125620dc32cdab@lbry.io>\n","lang":"Python","license":"mit","repos":"lbryio\/lbry,lbryio\/lbry,lbryio\/lbry"} {"commit":"d86cfb740f4119049a7ac293037f9cba12a3516b","old_file":"sqliteschema\/_const.py","new_file":"sqliteschema\/_const.py","old_contents":"# encoding: utf-8\n\n\"\"\"\n.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>\n\"\"\"\n\n\nMAX_VERBOSITY_LEVEL = 100\n\n# https:\/\/www.sqlite.org\/fileformat2.html\nSQLITE_SYSTEM_TABLES = [\n \"sqlite_master\",\n \"sqlite_sequence\",\n \"sqlite_stat1\",\n \"sqlite_stat2\",\n \"sqlite_stat3\",\n \"sqlite_stat4\",\n]\n\nSQLITE_SYSTEM_TABLE_LIST = SQLITE_SYSTEM_TABLES # deprecated\n\n\nclass SchemaHeader(object):\n ATTR_NAME = \"Attribute\"\n DATA_TYPE = \"Type\"\n PRIMARY_KEY = \"PRIMARY KEY\"\n NOT_NULL = \"NOT NULL\"\n UNIQUE = \"UNIQUE\"\n INDEX = \"Index\"\n","new_contents":"# encoding: utf-8\n\n\"\"\"\n.. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com>\n\"\"\"\n\n\nMAX_VERBOSITY_LEVEL = 100\n\n# https:\/\/www.sqlite.org\/fileformat2.html\nSQLITE_SYSTEM_TABLES = (\n \"sqlite_master\",\n \"sqlite_sequence\",\n \"sqlite_stat1\",\n \"sqlite_stat2\",\n \"sqlite_stat3\",\n \"sqlite_stat4\",\n)\n\nSQLITE_SYSTEM_TABLE_LIST = SQLITE_SYSTEM_TABLES # deprecated\n\n\nclass SchemaHeader(object):\n ATTR_NAME = \"Attribute\"\n DATA_TYPE = \"Type\"\n PRIMARY_KEY = \"PRIMARY KEY\"\n NOT_NULL = \"NOT NULL\"\n UNIQUE = \"UNIQUE\"\n INDEX = \"Index\"\n","subject":"Change a constant type from list to tuple","message":"Change a constant type from list to tuple\n","lang":"Python","license":"mit","repos":"thombashi\/sqliteschema"} {"commit":"eac54b3080c37d2530077f23b0c022ed818ca9a4","old_file":"util\/fixedpoint-qtcreator.py","new_file":"util\/fixedpoint-qtcreator.py","old_contents":"from dumper import *\n\ndef qdump__FixedPoint(d, value):\n d.putNumChild(3)\n raw = [ value[\"v\"][\"s\"][i].integer() for i in range( value[\"v\"][\"numWords\"].integer() ) ]\n ss = value[\"v\"][\"storageSize\"].integer()\n exp = [raw[i] * 2**(i * ss) for i in range(len(raw)) ]\n d.putValue(sum(exp) * 2**-value[\"fractionalWidth\"].integer())\n if d.isExpanded():\n with Children(d):\n d.putSubItem(\"fractionalWidth\", value[\"fractionalWidth\"])\n d.putSubItem(\"integerWidth\", value[\"integerWidth\"])\n d.putSubItem(\"v\", value[\"v\"])\n\ndef qdump__MultiwordInteger(d, value):\n d.putNumChild(3)\n raw = [ value[\"s\"][i].integer() for i in range( value[\"numWords\"].integer() ) ]\n exp = [ raw[i] * 2**(i * value[\"storageSize\"].integer()) for i in range(len(raw)) ]\n d.putValue(sum(exp))\n if d.isExpanded():\n with Children(d):\n d.putSubItem(\"numWords\", value[\"numWords\"])\n d.putSubItem(\"storageSize\", value[\"storageSize\"])\n d.putSubItem(\"s\", value[\"s\"])\n\n","new_contents":"from dumper import *\n\ndef qdump__FixedPoint(d, value):\n d.putNumChild(3)\n raw = [ value[\"v\"][\"s\"][i].integer() for i in range( value[\"v\"][\"numWords\"].integer() ) ]\n ss = value[\"v\"][\"storageSize\"].integer()\n exp = [raw[i] * 2**(i * ss) for i in range(len(raw)) ]\n if raw[-1] >= 2**(ss-1):\n exp += [ -2**(ss * len(raw)) ]\n d.putValue(sum(exp) * 2**-value[\"fractionalWidth\"].integer())\n if d.isExpanded():\n with Children(d):\n d.putSubItem(\"fractionalWidth\", value[\"fractionalWidth\"])\n d.putSubItem(\"integerWidth\", value[\"integerWidth\"])\n d.putSubItem(\"v\", value[\"v\"])\n\ndef qdump__MultiwordInteger(d, value):\n d.putNumChild(3)\n raw = [ value[\"s\"][i].integer() for i in range( value[\"numWords\"].integer() ) ]\n exp = [ raw[i] * 2**(i * value[\"storageSize\"].integer()) for i in range(len(raw)) ]\n d.putValue(sum(exp))\n if d.isExpanded():\n with Children(d):\n d.putSubItem(\"numWords\", value[\"numWords\"])\n d.putSubItem(\"storageSize\", value[\"storageSize\"])\n d.putSubItem(\"s\", value[\"s\"])\n\n","subject":"Support negative numbers in qtcreator debugging","message":"Support negative numbers in qtcreator debugging\n","lang":"Python","license":"mit","repos":"Cat-Ion\/FixedPoint,Cat-Ion\/FixedPoint"} {"commit":"218b0f9a42c8d3421f80a3b2b77c9f7f3334722d","old_file":"test_publisher.py","new_file":"test_publisher.py","old_contents":"import publisher\n\ntest_pdf_filename = \"test\/test.pdf\"\ntest_css_filename = \"test\/test.css\"\ntest_md_filename = \"test\/test.md\"\ntest_html_filename = \"test\/test.html\"\n\ntest_md = \"# Test heading\\n\\n- test item 1\\n- test item 2\"\n\ndef from_html_file():\n print publisher.md_to_html(publisher.from_file(test_md_filename))\n\ndef md_to_html():\n print publisher.md_to_html(test_source)\n\ndef md_and_css_to_html():\n html_source = publisher.md_and_css_to_html(publisher.from_file(test_md_filename),\n publisher.from_file(test_css_filename))\n print html_source\n publisher.to_file(html_source, test_html_filename)\n\ndef from_md_file_to_pdf_file():\n test_html = publisher.md_to_html(publisher.from_file(\"README.md\"))\n print publisher.html_to_pdf_file(test_html, test_pdf_filename, [test_css_filename])\n\nmd_and_css_to_html()\n","new_contents":"import publisher\n\ntest_pdf_filename = \"test\/test.pdf\"\ntest_css_filename = \"test\/test.css\"\ntest_md_filename = \"test\/test.md\"\ntest_html_filename = \"test\/test.html\"\ntest_sender = \"cpg@yakko.cs.wmich.edu\"\ntest_recipient = \"cpgillem@gmail.com\"\n\ntest_md = \"# Test heading\\n\\n- test item 1\\n- test item 2\"\n\ndef from_html_file():\n print publisher.md_to_html(publisher.from_file(test_md_filename))\n\ndef md_to_html():\n print publisher.md_to_html(test_source)\n\ndef md_and_css_to_html():\n html_source = publisher.md_and_css_to_html(publisher.from_file(test_md_filename),\n publisher.from_file(test_css_filename))\n print html_source\n publisher.to_file(html_source, test_html_filename)\n\ndef from_md_file_to_pdf_file():\n test_html = publisher.md_to_html(publisher.from_file(\"README.md\"))\n print publisher.html_to_pdf_file(test_html, test_pdf_filename, [test_css_filename])\n\ndef from_md_to_html_email():\n test_email = publisher.md_to_html_email(publisher.from_file(test_md_filename), \n publisher.from_file(test_css_filename))\n print test_email\n\n# The test case currently in use\nfrom_md_to_html_email()\n","subject":"Add test case for HTML email messages.","message":"Add test case for HTML email messages.\n","lang":"Python","license":"mit","repos":"cpgillem\/markdown_publisher,cpgillem\/markdown_publisher"} {"commit":"348c28bacececb787ab73c9716dc515d0fabbe4b","old_file":"armstrong\/hatband\/widgets\/visualsearch.py","new_file":"armstrong\/hatband\/widgets\/visualsearch.py","old_contents":"from django.forms import Widget\nfrom django.template.loader import render_to_string\n\nfrom ..utils import static_url\n\n\nclass GenericKeyWidget(Widget):\n template = \"admin\/hatband\/widgets\/generickey.html\"\n\n class Media:\n js = (static_url(\"visualsearch\/dependencies.js\"),\n static_url(\"visualsearch\/visualsearch.js\"),\n static_url(\"generickey.js\"),\n )\n\n css = {\n \"all\": (static_url(\"visualsearch\/visualsearch.css\"),\n static_url(\"hatband\/css\/generickey.css\"),\n )\n }\n\n\n def __init__(self, object_id_name=\"object_id\",\n content_type_name=\"content_type\", *args, **kwargs):\n super(GenericKeyWidget, self).__init__(*args, **kwargs)\n self.object_id_name = object_id_name\n self.content_type_name = content_type_name\n\n def render(self, name, value, attrs=None):\n if value is None:\n value = ''\n final_attrs = self.build_attrs(attrs, name=name)\n final_attrs[\"value\"] = value\n final_attrs[\"is_templated\"] = final_attrs[\"id\"].find(\"__prefix__\") > -1\n final_attrs[\"object_id_name\"] = self.object_id_name\n final_attrs[\"content_type_name\"] = self.content_type_name\n return render_to_string(self.template, final_attrs)\n","new_contents":"from django.forms import Widget\nfrom django.template.loader import render_to_string\n\nfrom ..utils import static_url\n\n\nclass GenericKeyWidget(Widget):\n template = \"admin\/hatband\/widgets\/generickey.html\"\n\n class Media:\n js = (static_url(\"visualsearch\/dependencies.js\"),\n static_url(\"visualsearch\/visualsearch.js\"),\n static_url(\"generickey.js\"),\n )\n\n css = {\n \"all\": (static_url(\"visualsearch\/visualsearch.css\"),\n static_url(\"hatband\/css\/generickey.css\"),\n )\n }\n\n\n def __init__(self, object_id_name=\"object_id\",\n content_type_name=\"content_type\", *args, **kwargs):\n super(GenericKeyWidget, self).__init__(*args, **kwargs)\n self.object_id_name = object_id_name\n self.content_type_name = content_type_name\n\n def render(self, name, value, attrs=None):\n if value is None:\n value = ''\n final_attrs = self.build_attrs(attrs, name=name)\n final_attrs.update({\n \"value\": value,\n \"is_templated\": final_attrs[\"id\"].find(\"__prefix__\") > -1,\n \"object_id_name\": self.object_id_name,\n \"content_type_name\": self.content_type_name,\n })\n return render_to_string(self.template, final_attrs)\n","subject":"Clean up this code a bit (no functional change)","message":"Clean up this code a bit (no functional change)\n","lang":"Python","license":"apache-2.0","repos":"armstrong\/armstrong.hatband,texastribune\/armstrong.hatband,armstrong\/armstrong.hatband,armstrong\/armstrong.hatband,texastribune\/armstrong.hatband,texastribune\/armstrong.hatband"} {"commit":"6b5c32960565775d8b94825087c503e58f5eed27","old_file":"openslides\/users\/migrations\/0003_group.py","new_file":"openslides\/users\/migrations\/0003_group.py","old_contents":"# Generated by Django 1.10.5 on 2017-01-11 21:45\nfrom __future__ import unicode_literals\n\nimport django.db.models.deletion\nfrom django.db import migrations, models\n\nimport openslides.users.models\nimport openslides.utils.models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('auth', '0008_alter_user_username_max_length'),\n ('users', '0002_user_misc_default_groups'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Group',\n fields=[(\n 'group_ptr',\n models.OneToOneField(\n auto_created=True,\n on_delete=django.db.models.deletion.CASCADE,\n parent_link=True,\n primary_key=True,\n serialize=False,\n to='auth.Group'))],\n options={\n 'default_permissions': (),\n },\n bases=(openslides.utils.models.RESTModelMixin, 'auth.group'),\n managers=[\n ('objects', openslides.users.models.GroupManager()),\n ],\n ),\n ]\n","new_contents":"# Generated by Django 1.10.5 on 2017-01-11 21:45\nfrom __future__ import unicode_literals\n\nimport django.db.models.deletion\nfrom django.db import migrations, models\n\nimport openslides.users.models\nimport openslides.utils.models\n\n\ndef create_openslides_groups(apps, schema_editor):\n \"\"\"\n Creates the users.models.Group objects for each existing\n django.contrib.auth.models.Group object.\n \"\"\"\n # We get the model from the versioned app registry;\n # if we directly import it, it will be the wrong version.\n DjangoGroup = apps.get_model('auth', 'Group')\n Group = apps.get_model('users', 'Group')\n for group in DjangoGroup.objects.all():\n Group.objects.create(group_ptr_id=group.pk, name=group.name)\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('auth', '0008_alter_user_username_max_length'),\n ('users', '0002_user_misc_default_groups'),\n ]\n\n operations = [\n migrations.CreateModel(\n name='Group',\n fields=[(\n 'group_ptr',\n models.OneToOneField(\n auto_created=True,\n on_delete=django.db.models.deletion.CASCADE,\n parent_link=True,\n primary_key=True,\n serialize=False,\n to='auth.Group'))],\n options={\n 'default_permissions': (),\n },\n bases=(openslides.utils.models.RESTModelMixin, 'auth.group'),\n managers=[\n ('objects', openslides.users.models.GroupManager()),\n ],\n ),\n migrations.RunPython(\n create_openslides_groups,\n ),\n ]\n","subject":"Fix the migration of the groups.","message":"Fix the migration of the groups.\n\nFixes #2915\n","lang":"Python","license":"mit","repos":"tsiegleauq\/OpenSlides,normanjaeckel\/OpenSlides,normanjaeckel\/OpenSlides,ostcar\/OpenSlides,emanuelschuetze\/OpenSlides,boehlke\/OpenSlides,ostcar\/OpenSlides,tsiegleauq\/OpenSlides,CatoTH\/OpenSlides,emanuelschuetze\/OpenSlides,FinnStutzenstein\/OpenSlides,OpenSlides\/OpenSlides,normanjaeckel\/OpenSlides,boehlke\/OpenSlides,jwinzer\/OpenSlides,FinnStutzenstein\/OpenSlides,CatoTH\/OpenSlides,tsiegleauq\/OpenSlides,CatoTH\/OpenSlides,jwinzer\/OpenSlides,boehlke\/OpenSlides,jwinzer\/OpenSlides,jwinzer\/OpenSlides,emanuelschuetze\/OpenSlides,normanjaeckel\/OpenSlides,OpenSlides\/OpenSlides,CatoTH\/OpenSlides,FinnStutzenstein\/OpenSlides,FinnStutzenstein\/OpenSlides,boehlke\/OpenSlides,ostcar\/OpenSlides,jwinzer\/OpenSlides,emanuelschuetze\/OpenSlides"} {"commit":"8d7f3320a9d3fd3b7365cad7631835a0a46f374e","old_file":"planner\/signals.py","new_file":"planner\/signals.py","old_contents":"from django.db.models.signals import m2m_changed\nfrom django.dispatch import receiver\nfrom django.core.exceptions import ValidationError\nfrom django.utils.translation import ugettext_lazy as _\nfrom .models import Step\n\n\n@receiver(m2m_changed, sender=Step.passengers.through)\ndef check_passengers(sender, **kwargs):\n step = kwargs['instance']\n if step.passengers.count() >= 8:\n raise ValidationError(_(\"You exceeded passenger maximum number\"))\n","new_contents":"from django.db.models.signals import m2m_changed\nfrom django.dispatch import receiver\nfrom .models import Step\n\n\n@receiver(m2m_changed, sender=Step.passengers.through)\ndef check_passengers(sender, **kwargs):\n step = kwargs['instance']\n if kwargs['action'] == 'post_add':\n if step.passengers.count() >= step.trip.max_num_passengers:\n step.trip.is_joinable = False\n elif kwargs['action'] == 'post_remove':\n step.trip.is_joinable = True\n","subject":"Make is_joinable automatic based of passenger number","message":"Make is_joinable automatic based of passenger number\n","lang":"Python","license":"mit","repos":"livingsilver94\/getaride,livingsilver94\/getaride,livingsilver94\/getaride"} {"commit":"2247162e277f8d09cc951442673d71cd0a8ece65","old_file":"active_link\/templatetags\/active_link_tags.py","new_file":"active_link\/templatetags\/active_link_tags.py","old_contents":"from django import VERSION as DJANGO_VERSION\nfrom django import template\nfrom django.conf import settings\nif DJANGO_VERSION[0] == 1 and DJANGO_VERSION[1] <= 9:\n from django.core.urlresolvers import reverse\nelse:\n from django.urls import reverse\n\nregister = template.Library()\n\n\n@register.simple_tag(takes_context=True)\ndef active_link(context, viewname, css_class=None, strict=None):\n \"\"\"\n Renders the given CSS class if the request path matches the path of the view.\n :param context: The context where the tag was called. Used to access the request object.\n :param viewname: The name of the view (include namespaces if any).\n :param css_class: The CSS class to render.\n :param strict: If True, the tag will perform an exact match with the request path.\n :return:\n \"\"\"\n if css_class is None:\n css_class = getattr(settings, 'ACTIVE_LINK_CSS_CLASS', 'active')\n\n if strict is None:\n strict = getattr(settings, 'ACTIVE_LINK_STRICT', False)\n\n request = context.get('request')\n if request is None:\n # Can't work without the request object.\n return ''\n path = reverse(viewname)\n if strict:\n active = request.path == path\n else:\n active = request.path.find(path) == 0\n if active:\n return css_class\n return ''\n","new_contents":"from django import VERSION as DJANGO_VERSION\nfrom django import template\nfrom django.conf import settings\nif DJANGO_VERSION[0] == 1 and DJANGO_VERSION[1] <= 9:\n from django.core.urlresolvers import reverse\nelse:\n from django.urls import reverse\n\nregister = template.Library()\n\n\n@register.simple_tag(takes_context=True)\ndef active_link(context, viewname, css_class=None, strict=None, *args, **kwargs):\n \"\"\"\n Renders the given CSS class if the request path matches the path of the view.\n :param context: The context where the tag was called. Used to access the request object.\n :param viewname: The name of the view (include namespaces if any).\n :param css_class: The CSS class to render.\n :param strict: If True, the tag will perform an exact match with the request path.\n :return:\n \"\"\"\n if css_class is None:\n css_class = getattr(settings, 'ACTIVE_LINK_CSS_CLASS', 'active')\n\n if strict is None:\n strict = getattr(settings, 'ACTIVE_LINK_STRICT', False)\n\n request = context.get('request')\n if request is None:\n # Can't work without the request object.\n return ''\n path = reverse(viewname, args=args, kwargs=kwargs)\n if strict:\n active = request.path == path\n else:\n active = request.path.find(path) == 0\n if active:\n return css_class\n return ''\n","subject":"Add ability to reverse views with args and kwargs","message":"Add ability to reverse views with args and kwargs","lang":"Python","license":"bsd-3-clause","repos":"valerymelou\/django-active-link"} {"commit":"ba00bececdcca3d1f224128123f12a9f634798b8","old_file":"feedhq\/feeds\/management\/commands\/updatefeeds.py","new_file":"feedhq\/feeds\/management\/commands\/updatefeeds.py","old_contents":"import logging\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.db import connection\nfrom raven import Client\n\nfrom ....tasks import enqueue\nfrom ...models import UniqueFeed, Feed\nfrom ...tasks import update_feed\nfrom ...utils import FeedUpdater\n\nlogger = logging.getLogger('feedupdater')\n\n\nclass Command(BaseCommand):\n \"\"\"Updates the users' feeds\"\"\"\n\n def handle(self, *args, **kwargs):\n if args:\n pk = args[0]\n feed = Feed.objects.get(pk=pk)\n feed.etag = ''\n return FeedUpdater(feed.url).update(use_etags=False)\n\n # Making a list of unique URLs. Makes one call whatever the number of\n # subscribers is.\n urls = Feed.objects.filter(muted=False).values_list('url', flat=True)\n unique_urls = {}\n map(unique_urls.__setitem__, urls, [])\n\n for url in unique_urls:\n try:\n try:\n unique = UniqueFeed.objects.get(url=url)\n if unique.should_update():\n enqueue(update_feed, url)\n except UniqueFeed.DoesNotExist:\n enqueue(update_feed, url)\n except Exception: # We don't know what to expect, and anyway\n # we're reporting the exception\n if settings.DEBUG or not hasattr(settings, 'SENTRY_DSN'):\n raise\n else:\n client = Client(dsn=settings.SENTRY_DSN)\n client.captureException()\n connection.close()\n","new_contents":"import logging\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\nfrom django.db import connection\nfrom raven import Client\n\nfrom ....tasks import enqueue\nfrom ...models import UniqueFeed, Feed\nfrom ...tasks import update_feed\nfrom ...utils import FeedUpdater\n\nlogger = logging.getLogger('feedupdater')\n\n\nclass Command(BaseCommand):\n \"\"\"Updates the users' feeds\"\"\"\n\n def handle(self, *args, **kwargs):\n if args:\n pk = args[0]\n feed = Feed.objects.get(pk=pk)\n feed.etag = ''\n return FeedUpdater(feed.url).update(use_etags=False)\n\n # Making a list of unique URLs. Makes one call whatever the number of\n # subscribers is.\n urls = Feed.objects.filter(muted=False).values_list('url', flat=True)\n unique_urls = {}\n map(unique_urls.__setitem__, urls, [])\n\n for url in unique_urls:\n try:\n try:\n unique = UniqueFeed.objects.get(url=url)\n if unique.should_update():\n enqueue(update_feed, url, timeout=20)\n except UniqueFeed.DoesNotExist:\n enqueue(update_feed, url, timeout=20)\n except Exception: # We don't know what to expect, and anyway\n # we're reporting the exception\n if settings.DEBUG or not hasattr(settings, 'SENTRY_DSN'):\n raise\n else:\n client = Client(dsn=settings.SENTRY_DSN)\n client.captureException()\n connection.close()\n","subject":"Set RQ timeout when enqueuing","message":"Set RQ timeout when enqueuing\n","lang":"Python","license":"bsd-3-clause","repos":"vincentbernat\/feedhq,feedhq\/feedhq,vincentbernat\/feedhq,feedhq\/feedhq,feedhq\/feedhq,rmoorman\/feedhq,rmoorman\/feedhq,feedhq\/feedhq,vincentbernat\/feedhq,rmoorman\/feedhq,rmoorman\/feedhq,vincentbernat\/feedhq,feedhq\/feedhq,vincentbernat\/feedhq,rmoorman\/feedhq"} {"commit":"1bad824786204353ff4f5b955ae687f088f80837","old_file":"employees\/tests.py","new_file":"employees\/tests.py","old_contents":"from django.test import TestCase\n\n# Create your tests here.\n","new_contents":"from .models import Employee\nfrom .serializers import EmployeeSerializer, EmployeeAvatarSerializer, EmployeeListSerializer\nfrom categories.serializers import CategorySerializer\nfrom django.core.urlresolvers import reverse\nfrom django.core.paginator import Paginator\nfrom rest_framework import status\nfrom rest_framework.authtoken.models import Token\nfrom rest_framework.pagination import PageNumberPagination\nfrom rest_framework.test import APITestCase, APIClient\n\n\nclass EmployeeTestCase(APITestCase):\n def setUp(self):\n Employee.objects.create_superuser('user', 'user@email.com', 'userpassword')\n\n def test_employee_creation(self):\n # token = Token.objects.get(user__username='user')\n # user = APIClient()\n # user.credentials(HTTP_AUTHORIZATION='Token ' + token.key)\n # employee1 = Employee.objects.get(username='user')\n # print user.username\n # #user.login(username=employee1.username, password=employee1.password)\n employee1 = Employee.objects.get(email='user@email.com')\n self.assertEqual(employee1.username, 'user')\n\n def test_employee_list(self):\n employees = Employee.objects.all()\n paginator = Paginator(employees, 20)\n print paginator.page(1)\n print paginator.page(1).object_list\n serializer = EmployeeListSerializer(employees, many=True)\n url = reverse('employees:employee_list')\n response = self.client.get(url, format='json')\n self.assertEqual(response.status_code, status.HTTP_200_OK)","subject":"Add draft testcases for employees views","message":"Add draft testcases for employees views\n","lang":"Python","license":"apache-2.0","repos":"belatrix\/BackendAllStars"} {"commit":"eac90ef4d470923bb823f99dc85984faac733f08","old_file":"pysuru\/services.py","new_file":"pysuru\/services.py","old_contents":"# coding: utf-8\nimport json\nfrom collections import namedtuple\n\nfrom pysuru.base import BaseAPI, ObjectMixin\n\n\nSERVICE_INSTANCE_ATTRS = (\n 'name',\n 'description',\n 'type',\n 'plan',\n 'teamOwner',\n)\n\n\n_ServiceInstance = namedtuple('ServiceInstance', SERVICE_INSTANCE_ATTRS)\n\n\nclass ServiceInstance(_ServiceInstance, ObjectMixin):\n pass\n\n\nclass ServiceInstanceAPI(BaseAPI):\n def filter_by_app(self, name):\n http_response = self.request('GET', '\/services\/instances?app=' + name)\n response = json.loads(http_response.data.decode('utf-8'))\n services = []\n for service_data in response:\n for index, instance in enumerate(service_data['instances']):\n data = {\n 'name': instance,\n 'type': service_data['service'],\n 'plan': service_data['plans'][index],\n }\n services.append(ServiceInstance.create(**data))\n return services\n\n def add(self, data):\n http_response = self.post_json('\/services\/instances', data)\n response = json.loads(http_response.data.decode('utf-8'))\n if response.status == 409:\n raise ServiceAlreadyExists()\n elif response.status == 200:\n return True\n else:\n return False\n\n\nclass ServiceAlreadyExists(Exception):\n pass\n","new_contents":"# coding: utf-8\nimport json\nfrom collections import namedtuple\n\nfrom pysuru.base import BaseAPI, ObjectMixin\n\n\nSERVICE_INSTANCE_ATTRS = (\n 'name',\n 'description',\n 'type',\n 'plan',\n)\n\n\n_ServiceInstance = namedtuple('ServiceInstance', SERVICE_INSTANCE_ATTRS)\n\n\nclass ServiceInstance(_ServiceInstance, ObjectMixin):\n pass\n\n\nclass ServiceInstanceAPI(BaseAPI):\n def filter_by_app(self, name):\n http_response = self.request('GET', '\/services\/instances?app=' + name)\n response = json.loads(http_response.data.decode('utf-8'))\n services = []\n for service_data in response:\n for index, instance in enumerate(service_data['instances']):\n data = {\n 'name': instance,\n 'type': service_data['service'],\n 'plan': service_data['plans'][index],\n }\n services.append(ServiceInstance.create(**data))\n return services\n\n def add(self, data):\n http_response = self.post_json('\/services\/instances', data)\n response = json.loads(http_response.data.decode('utf-8'))\n if response.status == 409:\n raise ServiceAlreadyExists()\n elif response.status == 200:\n return True\n else:\n return False\n\n\nclass ServiceAlreadyExists(Exception):\n pass\n","subject":"Remove (currently) unused service instance field","message":"Remove (currently) unused service instance field\n","lang":"Python","license":"mit","repos":"rcmachado\/pysuru"} {"commit":"1a320cadb37de27964f5973e2860804df3a5e479","old_file":"agir\/authentication\/tasks.py","new_file":"agir\/authentication\/tasks.py","old_contents":"from celery import shared_task\nfrom django.conf import settings\nfrom django.utils import timezone\n\nfrom agir.people.actions.mailing import send_mosaico_email\n\n\ndef interleave_spaces(s, n=3):\n return ' '.join([s[i:i+n] for i in range(0, len(s), n)])\n\n\n@shared_task\ndef send_login_email(email, short_code, expiry_time):\n utc_expiry_time = timezone.make_aware(timezone.datetime.fromtimestamp(expiry_time), timezone.utc)\n local_expiry_time = timezone.localtime(utc_expiry_time)\n\n send_mosaico_email(\n code='LOGIN_MESSAGE',\n subject=\"Connexion à agir.lafranceinsoumise.fr\",\n from_email=settings.EMAIL_FROM,\n bindings={\n 'CODE': interleave_spaces(short_code),\n 'EXPIRY_TIME': local_expiry_time.strftime(\"%H:%M\")\n },\n recipients=[email]\n )\n","new_contents":"from celery import shared_task\nfrom django.conf import settings\nfrom django.utils import timezone\n\nfrom agir.people.actions.mailing import send_mosaico_email\n\n\ndef interleave_spaces(s, n=3):\n return ' '.join([s[i:i+n] for i in range(0, len(s), n)])\n\n\n@shared_task\ndef send_login_email(email, short_code, expiry_time):\n utc_expiry_time = timezone.make_aware(timezone.datetime.utcfromtimestamp(expiry_time), timezone.utc)\n local_expiry_time = timezone.localtime(utc_expiry_time)\n\n send_mosaico_email(\n code='LOGIN_MESSAGE',\n subject=\"Connexion à agir.lafranceinsoumise.fr\",\n from_email=settings.EMAIL_FROM,\n bindings={\n 'CODE': interleave_spaces(short_code),\n 'EXPIRY_TIME': local_expiry_time.strftime(\"%H:%M\")\n },\n recipients=[email]\n )\n","subject":"Send correct expiration timing in login email","message":"Send correct expiration timing in login email\n","lang":"Python","license":"agpl-3.0","repos":"lafranceinsoumise\/api-django,lafranceinsoumise\/api-django,lafranceinsoumise\/api-django,lafranceinsoumise\/api-django"} {"commit":"487897e4b515a4c514fa0c91dec80d981c3bb98b","old_file":"tools\/telemetry\/telemetry\/core\/profile_types.py","new_file":"tools\/telemetry\/telemetry\/core\/profile_types.py","old_contents":"# Copyright (c) 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nimport os\n\nPROFILE_TYPE_MAPPING = {\n 'typical_user': 'chrome\/test\/data\/extensions\/profiles\/content_scripts1',\n 'power_user': 'chrome\/test\/data\/extensions\/profiles\/content_scripts10',\n}\nPROFILE_TYPES = PROFILE_TYPE_MAPPING.keys()\n\ndef GetProfileDir(profile_type):\n path = os.path.abspath(os.path.join(os.path.dirname(__file__),\n '..', '..', '..', '..', *PROFILE_TYPE_MAPPING[profile_type].split('\/')))\n assert os.path.exists(path)\n return path\n","new_contents":"# Copyright (c) 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\nimport os\n\nPROFILE_TYPE_MAPPING = {\n 'typical_user': 'chrome\/test\/data\/extensions\/profiles\/content_scripts1',\n 'power_user': 'chrome\/test\/data\/extensions\/profiles\/extension_webrequest',\n}\nPROFILE_TYPES = PROFILE_TYPE_MAPPING.keys()\n\ndef GetProfileDir(profile_type):\n path = os.path.abspath(os.path.join(os.path.dirname(__file__),\n '..', '..', '..', '..', *PROFILE_TYPE_MAPPING[profile_type].split('\/')))\n assert os.path.exists(path)\n return path\n","subject":"Use correct profile for power_user.","message":"[Telemetry] Use correct profile for power_user.\n\nTEST=None\nBUG=None\nNOTRY=True\n\n\nReview URL: https:\/\/chromiumcodereview.appspot.com\/12775015\n\ngit-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@188294 0039d316-1c4b-4281-b951-d872f2087c98\n","lang":"Python","license":"bsd-3-clause","repos":"patrickm\/chromium.src,mohamed--abdel-maksoud\/chromium.src,M4sse\/chromium.src,Just-D\/chromium-1,Pluto-tv\/chromium-crosswalk,krieger-od\/nwjs_chromium.src,TheTypoMaster\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,hgl888\/chromium-crosswalk,chuan9\/chromium-crosswalk,dednal\/chromium.src,chuan9\/chromium-crosswalk,timopulkkinen\/BubbleFish,hgl888\/chromium-crosswalk,mohamed--abdel-maksoud\/chromium.src,littlstar\/chromium.src,pozdnyakov\/chromium-crosswalk,chuan9\/chromium-crosswalk,TheTypoMaster\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,pozdnyakov\/chromium-crosswalk,markYoungH\/chromium.src,axinging\/chromium-crosswalk,mogoweb\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,M4sse\/chromium.src,dushu1203\/chromium.src,hgl888\/chromium-crosswalk-efl,krieger-od\/nwjs_chromium.src,Chilledheart\/chromium,hujiajie\/pa-chromium,bright-sparks\/chromium-spacewalk,PeterWangIntel\/chromium-crosswalk,dushu1203\/chromium.src,hgl888\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,hujiajie\/pa-chromium,Jonekee\/chromium.src,PeterWangIntel\/chromium-crosswalk,axinging\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,mohamed--abdel-maksoud\/chromium.src,ltilve\/chromium,hgl888\/chromium-crosswalk,timopulkkinen\/BubbleFish,Chilledheart\/chromium,markYoungH\/chromium.src,hgl888\/chromium-crosswalk,ondra-novak\/chromium.src,hujiajie\/pa-chromium,PeterWangIntel\/chromium-crosswalk,dednal\/chromium.src,chuan9\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,axinging\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,hujiajie\/pa-chromium,anirudhSK\/chromium,fujunwei\/chromium-crosswalk,pozdnyakov\/chromium-crosswalk,Fireblend\/chromium-crosswalk,markYoungH\/chromium.src,hgl888\/chromium-crosswalk-efl,Pluto-tv\/chromium-crosswalk,fujunwei\/chromium-crosswalk,ondra-novak\/chromium.src,patrickm\/chromium.src,anirudhSK\/chromium,PeterWangIntel\/chromium-crosswalk,krieger-od\/nwjs_chromium.src,littlstar\/chromium.src,timopulkkinen\/BubbleFish,krieger-od\/nwjs_chromium.src,patrickm\/chromium.src,fujunwei\/chromium-crosswalk,ChromiumWebApps\/chromium,Pluto-tv\/chromium-crosswalk,TheTypoMaster\/chromium-crosswalk,markYoungH\/chromium.src,ChromiumWebApps\/chromium,hujiajie\/pa-chromium,mohamed--abdel-maksoud\/chromium.src,krieger-od\/nwjs_chromium.src,patrickm\/chromium.src,jaruba\/chromium.src,dednal\/chromium.src,ltilve\/chromium,ondra-novak\/chromium.src,Just-D\/chromium-1,markYoungH\/chromium.src,hujiajie\/pa-chromium,TheTypoMaster\/chromium-crosswalk,M4sse\/chromium.src,patrickm\/chromium.src,fujunwei\/chromium-crosswalk,Just-D\/chromium-1,littlstar\/chromium.src,dushu1203\/chromium.src,crosswalk-project\/chromium-crosswalk-efl,axinging\/chromium-crosswalk,pozdnyakov\/chromium-crosswalk,PeterWangIntel\/chromium-crosswalk,hujiajie\/pa-chromium,mogoweb\/chromium-crosswalk,jaruba\/chromium.src,TheTypoMaster\/chromium-crosswalk,M4sse\/chromium.src,crosswalk-project\/chromium-crosswalk-efl,M4sse\/chromium.src,ltilve\/chromium,Just-D\/chromium-1,pozdnyakov\/chromium-crosswalk,ltilve\/chromium,M4sse\/chromium.src,Just-D\/chromium-1,dednal\/chromium.src,Fireblend\/chromium-crosswalk,Jonekee\/chromium.src,ondra-novak\/chromium.src,mogoweb\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,patrickm\/chromium.src,Pluto-tv\/chromium-crosswalk,Pluto-tv\/chromium-crosswalk,mogoweb\/chromium-crosswalk,littlstar\/chromium.src,PeterWangIntel\/chromium-crosswalk,jaruba\/chromium.src,jaruba\/chromium.src,mohamed--abdel-maksoud\/chromium.src,axinging\/chromium-crosswalk,markYoungH\/chromium.src,hgl888\/chromium-crosswalk-efl,markYoungH\/chromium.src,hujiajie\/pa-chromium,Chilledheart\/chromium,jaruba\/chromium.src,chuan9\/chromium-crosswalk,chuan9\/chromium-crosswalk,Jonekee\/chromium.src,mohamed--abdel-maksoud\/chromium.src,axinging\/chromium-crosswalk,Jonekee\/chromium.src,dushu1203\/chromium.src,Jonekee\/chromium.src,littlstar\/chromium.src,TheTypoMaster\/chromium-crosswalk,hgl888\/chromium-crosswalk-efl,dednal\/chromium.src,anirudhSK\/chromium,timopulkkinen\/BubbleFish,Fireblend\/chromium-crosswalk,hgl888\/chromium-crosswalk,jaruba\/chromium.src,mohamed--abdel-maksoud\/chromium.src,chuan9\/chromium-crosswalk,M4sse\/chromium.src,M4sse\/chromium.src,mogoweb\/chromium-crosswalk,patrickm\/chromium.src,hujiajie\/pa-chromium,ondra-novak\/chromium.src,timopulkkinen\/BubbleFish,PeterWangIntel\/chromium-crosswalk,Fireblend\/chromium-crosswalk,hgl888\/chromium-crosswalk,jaruba\/chromium.src,axinging\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,dednal\/chromium.src,ChromiumWebApps\/chromium,chuan9\/chromium-crosswalk,dushu1203\/chromium.src,Chilledheart\/chromium,chuan9\/chromium-crosswalk,fujunwei\/chromium-crosswalk,timopulkkinen\/BubbleFish,dushu1203\/chromium.src,Jonekee\/chromium.src,ltilve\/chromium,dednal\/chromium.src,ChromiumWebApps\/chromium,M4sse\/chromium.src,hgl888\/chromium-crosswalk,anirudhSK\/chromium,krieger-od\/nwjs_chromium.src,TheTypoMaster\/chromium-crosswalk,dushu1203\/chromium.src,dushu1203\/chromium.src,pozdnyakov\/chromium-crosswalk,jaruba\/chromium.src,fujunwei\/chromium-crosswalk,PeterWangIntel\/chromium-crosswalk,ondra-novak\/chromium.src,ChromiumWebApps\/chromium,ltilve\/chromium,anirudhSK\/chromium,hgl888\/chromium-crosswalk-efl,dednal\/chromium.src,Jonekee\/chromium.src,Chilledheart\/chromium,crosswalk-project\/chromium-crosswalk-efl,Jonekee\/chromium.src,krieger-od\/nwjs_chromium.src,ChromiumWebApps\/chromium,dushu1203\/chromium.src,ltilve\/chromium,ChromiumWebApps\/chromium,pozdnyakov\/chromium-crosswalk,mogoweb\/chromium-crosswalk,dednal\/chromium.src,M4sse\/chromium.src,fujunwei\/chromium-crosswalk,mohamed--abdel-maksoud\/chromium.src,markYoungH\/chromium.src,fujunwei\/chromium-crosswalk,Jonekee\/chromium.src,anirudhSK\/chromium,ltilve\/chromium,ltilve\/chromium,axinging\/chromium-crosswalk,pozdnyakov\/chromium-crosswalk,anirudhSK\/chromium,krieger-od\/nwjs_chromium.src,crosswalk-project\/chromium-crosswalk-efl,Jonekee\/chromium.src,anirudhSK\/chromium,littlstar\/chromium.src,anirudhSK\/chromium,Fireblend\/chromium-crosswalk,mogoweb\/chromium-crosswalk,axinging\/chromium-crosswalk,axinging\/chromium-crosswalk,mogoweb\/chromium-crosswalk,Just-D\/chromium-1,fujunwei\/chromium-crosswalk,krieger-od\/nwjs_chromium.src,Fireblend\/chromium-crosswalk,timopulkkinen\/BubbleFish,ondra-novak\/chromium.src,hujiajie\/pa-chromium,timopulkkinen\/BubbleFish,bright-sparks\/chromium-spacewalk,mohamed--abdel-maksoud\/chromium.src,krieger-od\/nwjs_chromium.src,Jonekee\/chromium.src,Fireblend\/chromium-crosswalk,ChromiumWebApps\/chromium,patrickm\/chromium.src,timopulkkinen\/BubbleFish,bright-sparks\/chromium-spacewalk,Pluto-tv\/chromium-crosswalk,anirudhSK\/chromium,bright-sparks\/chromium-spacewalk,ondra-novak\/chromium.src,jaruba\/chromium.src,anirudhSK\/chromium,ondra-novak\/chromium.src,patrickm\/chromium.src,dednal\/chromium.src,hgl888\/chromium-crosswalk-efl,PeterWangIntel\/chromium-crosswalk,Fireblend\/chromium-crosswalk,Just-D\/chromium-1,ChromiumWebApps\/chromium,hgl888\/chromium-crosswalk-efl,Fireblend\/chromium-crosswalk,dushu1203\/chromium.src,Just-D\/chromium-1,mogoweb\/chromium-crosswalk,pozdnyakov\/chromium-crosswalk,timopulkkinen\/BubbleFish,bright-sparks\/chromium-spacewalk,hgl888\/chromium-crosswalk,Pluto-tv\/chromium-crosswalk,pozdnyakov\/chromium-crosswalk,dednal\/chromium.src,Chilledheart\/chromium,jaruba\/chromium.src,ChromiumWebApps\/chromium,pozdnyakov\/chromium-crosswalk,crosswalk-project\/chromium-crosswalk-efl,Pluto-tv\/chromium-crosswalk,jaruba\/chromium.src,Chilledheart\/chromium,M4sse\/chromium.src,Pluto-tv\/chromium-crosswalk,TheTypoMaster\/chromium-crosswalk,bright-sparks\/chromium-spacewalk,mogoweb\/chromium-crosswalk,ChromiumWebApps\/chromium,hgl888\/chromium-crosswalk-efl,markYoungH\/chromium.src,Chilledheart\/chromium,TheTypoMaster\/chromium-crosswalk,dushu1203\/chromium.src,hujiajie\/pa-chromium,ChromiumWebApps\/chromium,anirudhSK\/chromium,littlstar\/chromium.src,Chilledheart\/chromium,mohamed--abdel-maksoud\/chromium.src,axinging\/chromium-crosswalk,markYoungH\/chromium.src,markYoungH\/chromium.src,timopulkkinen\/BubbleFish,Just-D\/chromium-1,littlstar\/chromium.src,mohamed--abdel-maksoud\/chromium.src,krieger-od\/nwjs_chromium.src"} {"commit":"761e74feac34c198da75f17b6145b6ca37d7afed","old_file":"tests\/__init__.py","new_file":"tests\/__init__.py","old_contents":"import threading\nimport time\n\n\nDEFAULT_SLEEP = 0.01\n\n\nclass CustomError(Exception):\n pass\n\n\ndef defer(callback, *args, **kwargs):\n sleep = kwargs.pop('sleep', DEFAULT_SLEEP)\n expected_return = kwargs.pop('expected_return', None)\n call = kwargs.pop('call', True)\n\n def func():\n time.sleep(sleep)\n if call:\n assert expected_return == callback(*args, **kwargs)\n else:\n print(\"generator is not re-called\")\n\n t = threading.Thread(target=func)\n t.start()\n\n\ndef wait_until_finished(wrapper, timeout=1, sleep=DEFAULT_SLEEP):\n\n start_time = time.time()\n while time.time() < start_time + timeout:\n # Relies on .has_terminated, but shouldn't be a problem\n if wrapper.has_terminated():\n return\n time.sleep(sleep)\n else:\n raise RuntimeError(\"Has not been collected within %ss\" % timeout)\n\n\nclass State(object):\n\n \"\"\"Helper class to keep track of a test's state.\"\"\"\n\n def __init__(self):\n self.reset()\n\n def inc(self):\n self.counter += 1\n\n def reset(self):\n self.counter = 0\n self.run = False\n","new_contents":"import threading\nimport time\n\n\nDEFAULT_SLEEP = 0.01\n\n\nclass CustomError(Exception):\n pass\n\n\ndef defer(callback, *args, **kwargs):\n sleep = kwargs.pop('sleep', DEFAULT_SLEEP)\n expected_return = kwargs.pop('expected_return', None)\n call = kwargs.pop('call', True)\n\n def func():\n time.sleep(sleep)\n if call:\n assert expected_return == callback(*args, **kwargs)\n else:\n print(\"generator is not re-called\")\n\n t = threading.Thread(target=func)\n t.start()\n\n\ndef wait_until_finished(wrapper, timeout=1, sleep=DEFAULT_SLEEP):\n\n start_time = time.time()\n while time.time() < start_time + timeout:\n # Relies on .has_terminated, but shouldn't be a problem\n if wrapper.has_terminated():\n return\n time.sleep(sleep)\n\n raise RuntimeError(\"Has not been collected within %ss\" % timeout)\n\n\nclass State(object):\n\n \"\"\"Helper class to keep track of a test's state.\"\"\"\n\n def __init__(self):\n self.reset()\n\n def inc(self):\n self.counter += 1\n\n def reset(self):\n self.counter = 0\n self.run = False\n","subject":"Fix test case being unable to fail","message":"Fix test case being unable to fail\n","lang":"Python","license":"mit","repos":"FichteFoll\/resumeback"} {"commit":"4a549a5d7509fa23fc6787662b4950291baa3407","old_file":"tests\/conftest.py","new_file":"tests\/conftest.py","old_contents":"import tempfile\n\nimport pytest\n\nfrom coaction import create_app\nfrom coaction.extensions import db as _db\n\n\ndbfile = tempfile.NamedTemporaryFile(delete=False)\ndbfile.close()\n\nTEST_DATABASE_FILE = dbfile.name\nTEST_DATABASE_URI = \"sqlite:\/\/\/\" + TEST_DATABASE_FILE\nTESTING = True\nDEBUG = False\nDEBUG_TB_ENABLED = False\nDEBUG_TB_INTERCEPT_REDIRECTS = False\nSQLALCHEMY_DATABASE_URI = TEST_DATABASE_URI\nWTF_CSRF_ENABLED = False\n\n\n\n@pytest.fixture\ndef app():\n app = create_app()\n app.config.from_object(__name__)\n return app\n\n\n@pytest.fixture\ndef db(app, request):\n def teardown():\n _db.drop_all()\n\n _db.app = app\n _db.create_all()\n\n request.addfinalizer(teardown)\n\n _db.app = app\n return _db","new_contents":"import tempfile\n\nimport pytest\n\nfrom toolshed import create_app\nfrom toolshed.extensions import db as _db\n\n\ndbfile = tempfile.NamedTemporaryFile(delete=False)\ndbfile.close()\n\nTEST_DATABASE_FILE = dbfile.name\nTEST_DATABASE_URI = \"postgres:\/\/localhost\/\" + TEST_DATABASE_FILE\nTESTING = True\nDEBUG = False\nDEBUG_TB_ENABLED = False\nDEBUG_TB_INTERCEPT_REDIRECTS = False\nSQLALCHEMY_DATABASE_URI = TEST_DATABASE_URI\nWTF_CSRF_ENABLED = False\n\n\n\n@pytest.fixture\ndef app():\n app = create_app()\n app.config.from_object(__name__)\n return app\n\n\n@pytest.fixture\ndef db(app, request):\n def teardown():\n _db.drop_all()\n\n _db.app = app\n _db.create_all()\n\n request.addfinalizer(teardown)\n\n _db.app = app\n return _db","subject":"Correct imports, switch to postgresql","message":"Correct imports, switch to postgresql\n","lang":"Python","license":"mit","repos":"PythonClutch\/python-clutch,PythonClutch\/python-clutch,PythonClutch\/python-clutch"} {"commit":"d14a34bff8e0462ebc2b8da9bc021f9c6f8f432d","old_file":"libclang_samples\/kernel-sigs.py","new_file":"libclang_samples\/kernel-sigs.py","old_contents":"import pprint\nimport sys\nimport clang.cindex\nfrom clang.cindex import CursorKind\n\n\ndef handle_function_decl(fdecl_cursor):\n children = list(fdecl_cursor.get_children())\n\n # Only interested in functions that have a CUDAGLOBAL_ATTR attached.\n if not any(c.kind == CursorKind.CUDAGLOBAL_ATTR for c in children):\n return\n\n print fdecl_cursor.displayname\n\n # Look at param decls\n for c in children:\n if c.kind == CursorKind.PARM_DECL:\n print '>>', c.spelling, c.type.spelling\n\n\ndef visitor(cursor):\n if cursor.kind == CursorKind.FUNCTION_DECL:\n handle_function_decl(cursor)\n\n for child in cursor.get_children():\n visitor(child)\n\nindex = clang.cindex.Index.create()\n# Parse as CUDA\ntu = index.parse(sys.argv[1], args=['-x', 'cuda'])\n\ndiagnostics = list(tu.diagnostics)\nif len(diagnostics) > 0:\n print 'There were parse errors'\n pprint.pprint(diagnostics)\nelse:\n visitor(tu.cursor)\n","new_contents":"import pprint\nimport sys\nimport clang.cindex\nfrom clang.cindex import CursorKind\n\n\ndef handle_function_decl(fdecl_cursor):\n children = list(fdecl_cursor.get_children())\n\n # Only interested in functions that have a CUDAGLOBAL_ATTR attached.\n if not any(c.kind == CursorKind.CUDAGLOBAL_ATTR for c in children):\n return\n\n print fdecl_cursor.displayname\n\n # Look at param decls\n for c in children:\n if c.kind == CursorKind.PARM_DECL:\n print '>>', c.spelling, c.type.spelling\n\n\nindex = clang.cindex.Index.create()\n# Parse as CUDA\ntu = index.parse(sys.argv[1], args=['-x', 'cuda'])\n\ndiagnostics = list(tu.diagnostics)\nif len(diagnostics) > 0:\n print 'There were parse errors'\n pprint.pprint(diagnostics)\nelse:\n for c in tu.cursor.walk_preorder():\n if c.kind == CursorKind.FUNCTION_DECL:\n handle_function_decl(c)\n","subject":"Use walk_preorder instead of manual visiting","message":"Use walk_preorder instead of manual visiting\n","lang":"Python","license":"unlicense","repos":"eliben\/llvm-clang-samples,eliben\/llvm-clang-samples,eliben\/llvm-clang-samples,eliben\/llvm-clang-samples,eliben\/llvm-clang-samples,eliben\/llvm-clang-samples"} {"commit":"dda9904a756e309047bebcbfecd2120383a257cc","old_file":"django_countries\/settings.py","new_file":"django_countries\/settings.py","old_contents":"from django.conf import settings\n\n\ndef _build_flag_url():\n if hasattr(settings, 'COUNTRIES_FLAG_URL'):\n url = settings.COUNTRIES_FLAG_URL\n else:\n url = 'flags\/%(code)s.gif'\n prefix = getattr(settings, 'STATIC_URL', '') or settings.MEDIA_URL\n if not prefix.endswith('\/'):\n prefix = '%s\/' % prefix\n return '%s%s' % (prefix, url)\n\n\nFLAG_URL = _build_flag_url()\n","new_contents":"from django.conf import settings\n\n\ndef _build_flag_url():\n if hasattr(settings, 'COUNTRIES_FLAG_URL'):\n url = settings.COUNTRIES_FLAG_URL\n else:\n url = 'flags\/%(code)s.gif'\n\n prefix = getattr(settings, 'STATIC_URL', '') or \\\n getattr(settings, 'STATICFILES_URL', '') or \\\n settings.MEDIA_URL\n\n if not prefix.endswith('\/'):\n prefix = '%s\/' % prefix\n return '%s%s' % (prefix, url)\n\n\nFLAG_URL = _build_flag_url()\n","subject":"Add django 1.3 staticfiles compatibility","message":"Add django 1.3 staticfiles compatibility\n","lang":"Python","license":"mit","repos":"degenhard\/django-countries"} {"commit":"4b65ab0fbc5839be9a49dd235549a13996a56108","old_file":"tests\/tabular_output\/test_tabulate_adapter.py","new_file":"tests\/tabular_output\/test_tabulate_adapter.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"Test the tabulate output adapter.\"\"\"\n\nfrom __future__ import unicode_literals\nfrom textwrap import dedent\n\nfrom cli_helpers.tabular_output import tabulate_adapter\n\n\ndef test_tabulate_wrapper():\n \"\"\"Test the *output_formatter.tabulate_wrapper()* function.\"\"\"\n data = [['abc', 1], ['d', 456]]\n headers = ['letters', 'number']\n output = tabulate_adapter.adapter(data, headers, table_format='psql')\n assert output == dedent('''\\\n +-----------+----------+\n | letters | number |\n |-----------+----------|\n | abc | 1 |\n | d | 456 |\n +-----------+----------+''')\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"Test the tabulate output adapter.\"\"\"\n\nfrom __future__ import unicode_literals\nfrom textwrap import dedent\n\nfrom cli_helpers.tabular_output import tabulate_adapter\n\n\ndef test_tabulate_wrapper():\n \"\"\"Test the *output_formatter.tabulate_wrapper()* function.\"\"\"\n data = [['abc', 1], ['d', 456]]\n headers = ['letters', 'number']\n output = tabulate_adapter.adapter(data, headers, table_format='psql')\n assert output == dedent('''\\\n +-----------+----------+\n | letters | number |\n |-----------+----------|\n | abc | 1 |\n | d | 456 |\n +-----------+----------+''')\n","subject":"Fix tabulate adapter test with numparse on.","message":"Fix tabulate adapter test with numparse on.\n","lang":"Python","license":"bsd-3-clause","repos":"dbcli\/cli_helpers,dbcli\/cli_helpers"} {"commit":"d67099ce7d30e31b98251f7386b33caaa5199a01","old_file":"censusreporter\/config\/prod\/wsgi.py","new_file":"censusreporter\/config\/prod\/wsgi.py","old_contents":"import os\nfrom django.core.wsgi import get_wsgi_application\n\nimport newrelic.agent\nnewrelic.agent.initialize('\/var\/www-data\/censusreporter\/conf\/newrelic.ini')\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"config.prod.settings\")\napplication = get_wsgi_application()\n","new_contents":"import os\nfrom django.core.wsgi import get_wsgi_application\n\nimport newrelic.agent\nnewrelic.agent.initialize(os.path.join(os.path.abspath(os.path.dirname(__file__)), '..\/..\/..\/conf\/newrelic.ini'))\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"config.prod.settings\")\napplication = get_wsgi_application()\n","subject":"Correct location of newrelic config","message":"Correct location of newrelic config\n","lang":"Python","license":"mit","repos":"sseguku\/simplecensusug,Code4SA\/censusreporter,Code4SA\/censusreporter,Code4SA\/censusreporter,sseguku\/simplecensusug,4bic\/censusreporter,sseguku\/simplecensusug,4bic\/censusreporter,Code4SA\/censusreporter,4bic\/censusreporter"} {"commit":"f69e6555387d4cb6828baaac3ce1a17217577b48","old_file":"faddsdata\/format_definitions\/__init__.py","new_file":"faddsdata\/format_definitions\/__init__.py","old_contents":"from apt import APT_RECORDS, ATT_RECORDS, RWY_RECORDS, RMK_RECORDS, APT_RECORD_MAP\nfrom arb import ARB_RECORDS\nfrom awos import AWOS_RECORDS\n","new_contents":"from apt import APT_RECORDS, ATT_RECORDS, RWY_RECORDS, RMK_RECORDS, APT_RECORD_MAP\nfrom awos import AWOS_RECORDS\n","subject":"Remove import for lib not commited yet.","message":"Remove import for lib not commited yet.\n","lang":"Python","license":"bsd-3-clause","repos":"adamfast\/faddsdata"} {"commit":"76177710323f1ccb408fd006ef7b87ff36743f8c","old_file":"telostats\/settings\/heroku.py","new_file":"telostats\/settings\/heroku.py","old_contents":"from __future__ import absolute_import\n\nimport dj_database_url\nimport urlparse\n\nfrom os import environ\nfrom .base import *\n\nENV = 'HEROKU'\n\n# Store files on S3\nDEFAULT_FILE_STORAGE = 'storages.backends.s3boto.S3BotoStorage'\n\nredis_url = urlparse.urlparse(environ.get('REDISTOGO_URL'))\nCACHES = {\n 'default': {\n 'BACKEND': 'redis_cache.RedisCache',\n 'LOCATION': '%s:%s' % (redis_url.hostname, redis_url.port),\n 'OPTIONS': {\n 'PASSWORD': redis_url.password,\n 'DB': 1,\n },\n },\n}\n\n# Grab database info\nDATABASES = {\n 'default': dj_database_url.config()\n}\n\n# Setup sentry \/ raven\nSENTRY_DSN = environ.get('SENTRY_DSN')\nINSTALLED_APPS += (\n 'raven.contrib.django',\n)\n","new_contents":"from __future__ import absolute_import\n\nimport dj_database_url\nimport urlparse\n\nfrom os import environ\nfrom .base import *\n\nENV = 'HEROKU'\n\n# Store files on S3\nDEFAULT_FILE_STORAGE = 'storages.backends.s3boto.S3BotoStorage'\n\nredis_url = urlparse.urlparse(environ.get('REDISTOGO_URL'))\nCACHES = {\n 'default': {\n 'BACKEND': 'redis_cache.RedisCache',\n 'LOCATION': '%s:%s' % (redis_url.hostname, redis_url.port),\n 'OPTIONS': {\n 'PASSWORD': redis_url.password,\n 'DB': 0,\n },\n },\n}\n\n# Grab database info\nDATABASES = {\n 'default': dj_database_url.config()\n}\n\n# Setup sentry \/ raven\nSENTRY_DSN = environ.get('SENTRY_DSN')\nINSTALLED_APPS += (\n 'raven.contrib.django',\n)\n","subject":"Fix wrong Redis DB setting","message":"Fix wrong Redis DB setting\n","lang":"Python","license":"bsd-3-clause","repos":"idan\/telostats,idan\/telostats,idan\/telostats"} {"commit":"5550217c028c422e7dc2d54c3b8b61ea43cfc26f","old_file":"dosagelib\/__pyinstaller\/hook-dosagelib.py","new_file":"dosagelib\/__pyinstaller\/hook-dosagelib.py","old_contents":"# SPDX-License-Identifier: MIT\n# Copyright (C) 2016-2020 Tobias Gruetzmacher\nfrom PyInstaller.utils.hooks import collect_submodules, copy_metadata\n\nhiddenimports = collect_submodules('dosagelib.plugins')\ndatas = copy_metadata('dosage')\n","new_contents":"# SPDX-License-Identifier: MIT\n# Copyright (C) 2016-2022 Tobias Gruetzmacher\nfrom PyInstaller.utils.hooks import collect_data_files, collect_submodules, copy_metadata\n\nhiddenimports = collect_submodules('dosagelib.plugins')\ndatas = copy_metadata('dosage') + collect_data_files('dosagelib')\n","subject":"Make sure data files are included","message":"PyInstaller: Make sure data files are included\n","lang":"Python","license":"mit","repos":"webcomics\/dosage,webcomics\/dosage"} {"commit":"76af6248c479127f1a212c331a2278f69484bcbc","old_file":"project\/api\/management\/commands\/rebuild_data.py","new_file":"project\/api\/management\/commands\/rebuild_data.py","old_contents":"# Django\nfrom django.apps import apps\nfrom django.core.management.base import BaseCommand\nfrom django.utils import timezone\nimport datetime\n\nclass Command(BaseCommand):\n help = \"Command to rebuild denorms.\"\n\n def add_arguments(self, parser):\n parser.add_argument(\n '--days',\n type=int,\n dest='days',\n nargs='?',\n const=1,\n help='Number of days to update.',\n )\n\n parser.add_argument(\n '--hours',\n type=int,\n dest='hours',\n nargs='?',\n const=1,\n help='Number of hours to update.',\n )\n\n parser.add_argument(\n '--minutes',\n type=int,\n dest='minutes',\n nargs='?',\n const=1,\n help='Number of hours to update.',\n )\n\n def handle(self, *args, **options):\n # Set Cursor\n if options['days']:\n cursor = timezone.now() - datetime.timedelta(days=options['days'], hours=1)\n elif options['hours']:\n cursor = timezone.now() - datetime.timedelta(hours=options['hours'], minutes=5)\n elif options['minutes']:\n cursor = timezone.now() - datetime.timedelta(minutes=options['minutes'], seconds=5)\n else:\n cursor = None\n Group = apps.get_model('api.group')\n Group.objects.denormalize(cursor=cursor)\n # Group.objects.sort_tree()\n Group.objects.update_seniors()\n Award = apps.get_model('api.award')\n Award.objects.sort_tree()\n return\n","new_contents":" # Django\nfrom django.apps import apps\nfrom django.core.management.base import BaseCommand\nfrom django.utils import timezone\nimport datetime\n\nclass Command(BaseCommand):\n help = \"Command to rebuild denorms.\"\n\n def add_arguments(self, parser):\n parser.add_argument(\n '--days',\n type=int,\n dest='days',\n nargs='?',\n const=1,\n help='Number of days to update.',\n )\n\n parser.add_argument(\n '--hours',\n type=int,\n dest='hours',\n nargs='?',\n const=1,\n help='Number of hours to update.',\n )\n\n parser.add_argument(\n '--minutes',\n type=int,\n dest='minutes',\n nargs='?',\n const=1,\n help='Number of hours to update.',\n )\n\n def handle(self, *args, **options):\n # Set Cursor\n if options['days']:\n cursor = timezone.now() - datetime.timedelta(days=options['days'], hours=1)\n elif options['hours']:\n cursor = timezone.now() - datetime.timedelta(hours=options['hours'], minutes=5)\n elif options['minutes']:\n cursor = timezone.now() - datetime.timedelta(minutes=options['minutes'], seconds=5)\n else:\n cursor = None\n Group = apps.get_model('api.group')\n # Group.objects.denormalize(cursor=cursor)\n # Group.objects.sort_tree()\n # Group.objects.update_seniors()\n Award = apps.get_model('api.award')\n Award.objects.sort_tree()\n return\n","subject":"Disable group rebuild until fix indexing","message":"Disable group rebuild until fix indexing\n","lang":"Python","license":"bsd-2-clause","repos":"dbinetti\/barberscore-django,barberscore\/barberscore-api,barberscore\/barberscore-api,dbinetti\/barberscore-django,barberscore\/barberscore-api,dbinetti\/barberscore,dbinetti\/barberscore,barberscore\/barberscore-api"} {"commit":"84490442de881788a7f83bc18ec4eedb7f6edb99","old_file":"tests\/bugs\/test-200908231005.py","new_file":"tests\/bugs\/test-200908231005.py","old_contents":"import pyxb.binding.generate\nimport pyxb.binding.datatypes as xs\nimport pyxb.binding.basis\nimport pyxb.utils.domutils\n\nimport os.path\nxsd='''<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<xs:schema xmlns:xs=\"http:\/\/www.w3.org\/2001\/XMLSchema\">\n <xs:element name=\"structure\">\n <xs:complexType><xs:anyAttribute processContents=\"lax\"\/><\/xs:complexType>\n <\/xs:element>\n<\/xs:schema>'''\n\n#file('schema.xsd', 'w').write(xsd)\ncode = pyxb.binding.generate.GeneratePython(schema_text=xsd)\nfile('code.py', 'w').write(code)\n\nrv = compile(code, 'test', 'exec')\neval(rv)\n\nfrom pyxb.exceptions_ import *\n\nimport unittest\n\nAttributeNamespace = pyxb.namespace.NamespaceInstance('URN:attr:200908231005')\n\nclass TestTrac_200908231005 (unittest.TestCase):\n def testParsing (self):\n xmls = '<structure xmlns:attr=\"%s\" attr:field=\"value\"\/>' % (AttributeNamespace.uri(),)\n instance = CreateFromDocument(xmls)\n wam = instance.wildcardAttributeMap()\n self.assertEqual(1, len(wam))\n self.assertEqual('value', wam.get(AttributeNamespace.createExpandedName('field')))\n\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"import pyxb.binding.generate\nimport pyxb.binding.datatypes as xs\nimport pyxb.binding.basis\nimport pyxb.utils.domutils\n\nimport os.path\nxsd='''<?xml version=\"1.0\" encoding=\"UTF-8\"?>\n<xs:schema xmlns:xs=\"http:\/\/www.w3.org\/2001\/XMLSchema\">\n <xs:element name=\"structure\">\n <xs:complexType><xs:anyAttribute processContents=\"lax\"\/><\/xs:complexType>\n <\/xs:element>\n<\/xs:schema>'''\n\n#file('schema.xsd', 'w').write(xsd)\ncode = pyxb.binding.generate.GeneratePython(schema_text=xsd)\n#file('code.py', 'w').write(code)\n\nrv = compile(code, 'test', 'exec')\neval(rv)\n\nfrom pyxb.exceptions_ import *\n\nimport unittest\n\nAttributeNamespace = pyxb.namespace.NamespaceInstance('URN:attr:200908231005')\n\nclass TestTrac_200908231005 (unittest.TestCase):\n def testParsing (self):\n xmls = '<structure xmlns:attr=\"%s\" attr:field=\"value\"\/>' % (AttributeNamespace.uri(),)\n instance = CreateFromDocument(xmls)\n wam = instance.wildcardAttributeMap()\n self.assertEqual(1, len(wam))\n self.assertEqual('value', wam.get(AttributeNamespace.createExpandedName('field')))\n\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Stop writing generated code unnecessarily","message":"Stop writing generated code unnecessarily\n","lang":"Python","license":"apache-2.0","repos":"jonfoster\/pyxb-upstream-mirror,pabigot\/pyxb,pabigot\/pyxb,jonfoster\/pyxb2,balanced\/PyXB,jonfoster\/pyxb-upstream-mirror,jonfoster\/pyxb-upstream-mirror,jonfoster\/pyxb1,jonfoster\/pyxb2,CantemoInternal\/pyxb,balanced\/PyXB,balanced\/PyXB,jonfoster\/pyxb2,CantemoInternal\/pyxb,jonfoster\/pyxb1,CantemoInternal\/pyxb"} {"commit":"ae1a2b73d0c571c49726528f9b8730c9e02ce35f","old_file":"tests\/integration\/test_webui.py","new_file":"tests\/integration\/test_webui.py","old_contents":"import requests\nimport pytest\n\n\nclass TestWebUI(object):\n def get_page(self, page):\n return requests.get('http:\/\/nginx' + page)\n\n pages = [\n {\n 'page': '\/',\n 'matching_text': 'Diamond',\n },\n {\n 'page': '\/scoreboard',\n },\n {\n 'page': '\/login',\n 'matching_text': 'Please sign in',\n },\n {\n 'page': '\/about',\n 'matching_text': 'Use the following credentials to login',\n },\n {\n 'page': '\/overview',\n },\n {\n 'page': '\/api\/overview\/data'\n }\n ]\n\n @pytest.mark.parametrize(\"page_data\", pages)\n def test_page(self, page_data):\n resp = self.get_page(page_data['page'])\n assert resp.status_code == 200\n if 'matching_text' in page_data:\n assert page_data['matching_text'] in resp.text\n","new_contents":"import requests\nimport pytest\n\n\nclass TestWebUI(object):\n def get_page(self, page):\n return requests.get('https:\/\/nginx\/{0}'.format(page), verify=False)\n\n pages = [\n {\n 'page': '',\n 'matching_text': 'Diamond',\n },\n {\n 'page': 'scoreboard',\n },\n {\n 'page': 'login',\n 'matching_text': 'Please sign in',\n },\n {\n 'page': 'about',\n 'matching_text': 'Use the following credentials to login',\n },\n {\n 'page': 'overview',\n },\n {\n 'page': 'api\/overview\/data'\n }\n ]\n\n @pytest.mark.parametrize(\"page_data\", pages)\n def test_page(self, page_data):\n resp = self.get_page(page_data['page'])\n assert resp.status_code == 200\n if 'matching_text' in page_data:\n assert page_data['matching_text'] in resp.text\n","subject":"Fix webui integration tests to use https","message":"Fix webui integration tests to use https\n","lang":"Python","license":"mit","repos":"pwnbus\/scoring_engine,pwnbus\/scoring_engine,pwnbus\/scoring_engine,pwnbus\/scoring_engine"} {"commit":"afb8aadcc1dbea109c7882c1a1d65fc328372a74","old_file":"resources\/Dependencies\/DecoraterBotCore\/Core.py","new_file":"resources\/Dependencies\/DecoraterBotCore\/Core.py","old_contents":"# coding=utf-8\n\"\"\"\nDecoraterBotCore\n~~~~~~~~~~~~~~~~~~~\n\nCore to DecoraterBot\n\n:copyright: (c) 2015-2018 AraHaan\n:license: MIT, see LICENSE for more details.\n\n\"\"\"\nfrom DecoraterBotUtils.utils import BaseClient, config\n\n\n__all__ = ['main', 'BotClient']\n\n\nclass BotClient(BaseClient):\n \"\"\"\n Bot Main client Class.\n This is where the Events are Registered.\n \"\"\"\n def __init__(self, **kwargs):\n super(BotClient, self).__init__(**kwargs)\n\n\ndef main():\n \"\"\"\n EntryPoint to DecoraterBot.\n \"\"\"\n BotClient(command_prefix=config.bot_prefix,\n description=config.description,\n pm_help=False)\n","new_contents":"# coding=utf-8\n\"\"\"\nDecoraterBotCore\n~~~~~~~~~~~~~~~~~~~\n\nCore to DecoraterBot\n\n:copyright: (c) 2015-2018 AraHaan\n:license: MIT, see LICENSE for more details.\n\n\"\"\"\nfrom DecoraterBotUtils.utils import BotClient, config\n\n\n__all__ = ['main']\n\n\ndef main():\n \"\"\"\n EntryPoint to DecoraterBot.\n \"\"\"\n BotClient(command_prefix=config.bot_prefix,\n description=config.description,\n pm_help=False)\n","subject":"Update to use new client class name on the DecoraterBotUtils.utils module.","message":"Update to use new client class name on the DecoraterBotUtils.utils module.\n\n","lang":"Python","license":"mit","repos":"DecoraterBot-devs\/DecoraterBot"} {"commit":"f862823b24bfe59f6fb15fc417aea8ebacdc092d","old_file":"autoresponse\/__init__.py","new_file":"autoresponse\/__init__.py","old_contents":"","new_contents":"import twisted.python.failure\nimport twisted.web.error\nimport scrapy.item\n\n\nclass Autoresponder(object):\n # The purpose of Autoresponder is to be initialized during a run\n # of a test case and used to iterate over scrapy Request objects\n # until finally there are no more Requests to execute.\n #\n # Instead of passing the Requests to the 'net, it handles them\n # through the configuration you pass it.\n #\n # Successful (status=200) Responses can be generated by configuring\n # url2filename.\n #\n # Error responses (status usually 4xx) can be generated by configuring\n # url2errors.\n def __init__(self, url2filename, url2errors=None):\n self.url2filename = url2filename\n if url2errors is None:\n url2errors = {}\n self.url2errors = url2errors\n\n @staticmethod\n def manufacture_http_failure(status_code):\n error = twisted.web.error.Error(code=status_code)\n failure = twisted.python.failure.Failure(\n exc_value=error, exc_type=twisted.web.error.Error)\n return failure\n\n def respond_recursively(self, request_iterable):\n items = []\n work_queue = []\n work_queue.extend(request_iterable)\n while work_queue:\n thing = work_queue.pop(0)\n if isinstance(thing, scrapy.item.Item):\n items.append(thing)\n continue\n\n request = thing\n if request.url in self.url2filename:\n raise NotImplemented\n\n if request.url in self.url2errors:\n status_code = self.url2errors[request.url]\n failure = (\n Autoresponder.manufacture_http_failure(status_code))\n results = request.errback(failure)\n if results:\n if isinstance(results, scrapy.item.Item):\n results = [results]\n work_queue.extend(results)\n return items\n","subject":"Add an implementation that works for testing errbacks","message":"Add an implementation that works for testing errbacks\n","lang":"Python","license":"apache-2.0","repos":"paulproteus\/autoresponse"} {"commit":"ca6508bf6e8a1fd2032baaacb81e9131c088c6f8","old_file":"UCP\/news_event\/serializers.py","new_file":"UCP\/news_event\/serializers.py","old_contents":"from django.contrib.auth.models import User\n\nfrom rest_framework import serializers\n\nfrom login.models import UserProfile\nfrom discussion.serializers import UserShortSerializer, UserProfileShortSerializer\nfrom news_event.models import Event, News\n\n\nclass EventSerializer(serializers.ModelSerializer):\n \n class Meta:\n model = Event\n fields = ('id', 'title', 'description','posted_at', 'image')\n \n \nclass NewsSerializer(serializers.ModelSerializer):\n \n class Meta:\n model = News\n fields = ('id', 'title', 'description')","new_contents":"from django.contrib.auth.models import User\n\nfrom rest_framework import serializers\n\nfrom login.models import UserProfile\nfrom discussion.serializers import UserShortSerializer, UserProfileShortSerializer\nfrom news_event.models import Event, News\n\n\nclass EventSerializer(serializers.ModelSerializer):\n \n class Meta:\n model = Event\n fields = ('id', 'title', 'description','posted_at', 'image', 'venue')\n \n \nclass NewsSerializer(serializers.ModelSerializer):\n \n class Meta:\n model = News\n fields = ('id', 'title', 'description')","subject":"Add venue to event serializer","message":"Add venue to event serializer\n","lang":"Python","license":"bsd-3-clause","repos":"BuildmLearn\/University-Campus-Portal-UCP,BuildmLearn\/University-Campus-Portal-UCP,BuildmLearn\/University-Campus-Portal-UCP"} {"commit":"77b943b8bace86d82f76dfa159c5de48d9379dbb","old_file":"svenv\/blog\/settings.py","new_file":"svenv\/blog\/settings.py","old_contents":"BLOG_TITLE = 'Svenv.nl'\nBLOG_DESCRIPTION = 'Blogposts about tech related subject like Unix, Linux, Docker and programming.'\n\nBASE_URL = 'http:\/\/svenv.nl'\n\nCONTACT_PAGE_PATH = 'contact'\nCONTACT_THANK_YOU_PAGE = '\/thankyou'\nEMAIL = 'svenvandescheur@gmail.com'\nEMAIL_FROM = 'noreply@svenv.nl'\nSMTP_HOST = 'localhost'\n\nREST_FRAMEWORK = {\n 'DEFAULT_FILTER_BACKENDS': ('rest_framework.filters.OrderingFilter',),\n 'DEFAULT_PAGINATION_CLASS': 'rest_framework.pagination.PageNumberPagination',\n 'PAGE_SIZE': 12,\n 'DEFAULT_PERMISSION_CLASSES': [\n 'rest_framework.permissions.DjangoModelPermissionsOrAnonReadOnly'\n ],\n}","new_contents":"BLOG_TITLE = 'Svenv.nl'\nBLOG_DESCRIPTION = 'Blogposts about tech related subject like Unix, Linux, Docker and programming.'\n\nBASE_URL = 'https:\/\/svenv.nl'\n\nCONTACT_PAGE_PATH = 'contact'\nCONTACT_THANK_YOU_PAGE = '\/thankyou'\nEMAIL = 'svenvandescheur@gmail.com'\nEMAIL_FROM = 'noreply@svenv.nl'\nSMTP_HOST = 'localhost'\n\nREST_FRAMEWORK = {\n 'DEFAULT_FILTER_BACKENDS': ('rest_framework.filters.OrderingFilter',),\n 'DEFAULT_PAGINATION_CLASS': 'rest_framework.pagination.PageNumberPagination',\n 'PAGE_SIZE': 12,\n 'DEFAULT_PERMISSION_CLASSES': [\n 'rest_framework.permissions.DjangoModelPermissionsOrAnonReadOnly'\n ],\n}","subject":"Set base url to https","message":"Set base url to https\n","lang":"Python","license":"mit","repos":"svenvandescheur\/svenv.nl-app,svenvandescheur\/svenv.nl-app,svenvandescheur\/svenv.nl-app,svenvandescheur\/svenv.nl-app"} {"commit":"958426d43a5aa8e153e0999417377b54be67f04b","old_file":"ctypeslib\/experimental\/byref_at.py","new_file":"ctypeslib\/experimental\/byref_at.py","old_contents":"# hack a byref_at function\n\nfrom ctypes import *\n\ntry:\n set\nexcept NameError:\n from sets import Set as set\n\ndef _determine_layout():\n result = set()\n for obj in (c_int(), c_longlong(), c_float(), c_double(), (c_int * 32)()):\n ref = byref(obj)\n result.add((c_void_p * 32).from_address(id(ref))[:].index(id(obj)) * sizeof(c_void_p))\n if len(result) != 1:\n raise RuntimeError, \"cannot determine byref() object layout\"\n return result.pop()\n\noffset = _determine_layout()\n\n__all__ = [\"byref_at\"]\n","new_contents":"from ctypes import *\n\n\"\"\"\nstruct tagPyCArgObject {\n\tPyObject_HEAD\n\tffi_type *pffi_type;\n\tchar tag;\n\tunion {\n\t\tchar c;\n\t\tchar b;\n\t\tshort h;\n\t\tint i;\n\t\tlong l;\n#ifdef HAVE_LONG_LONG\n\t\tPY_LONG_LONG q;\n#endif\n\t\tdouble d;\n\t\tfloat f;\n\t\tvoid *p;\n\t} value;\n\tPyObject *obj;\n\tint size; \/* for the 'V' tag *\/\n};\n\"\"\"\n\nclass value(Union):\n _fields_ = [(\"c\", c_char),\n (\"h\", c_short),\n (\"i\", c_int),\n (\"l\", c_long),\n (\"q\", c_longlong),\n (\"d\", c_double),\n (\"f\", c_float),\n (\"p\", c_void_p)]\n\n# Thanks to Lenard Lindstrom for this tip: The sizeof(PyObject_HEAD)\n# is the same as object.__basicsize__.\n\nclass PyCArgObject(Structure):\n _fields_ = [(\"PyObject_HEAD\", c_byte * object.__basicsize__),\n (\"pffi_type\", c_void_p),\n (\"value\", value),\n (\"obj\", c_void_p),\n (\"size\", c_int)]\n _anonymous_ = [\"value\"]\n\nassert sizeof(PyCArgObject) == type(byref(c_int())).__basicsize__\n\nprint \"sizeof(PyCArgObject)\", sizeof(PyCArgObject)\nfor name in \"c h i l q d f p\".split():\n print name, getattr(PyCArgObject, name)\n","subject":"Define the structure of PyCArgObject in ctypes.","message":"Define the structure of PyCArgObject in ctypes.\n\ngit-svn-id: ac2c3632cb6543e7ab5fafd132c7fe15057a1882@52772 6015fed2-1504-0410-9fe1-9d1591cc4771\n","lang":"Python","license":"mit","repos":"trolldbois\/ctypeslib,luzfcb\/ctypeslib,luzfcb\/ctypeslib,trolldbois\/ctypeslib,trolldbois\/ctypeslib,luzfcb\/ctypeslib"} {"commit":"93650f0ae762949affcee4bfcf4a94093fe651d5","old_file":"neuralmonkey\/runners\/perplexity_runner.py","new_file":"neuralmonkey\/runners\/perplexity_runner.py","old_contents":"from typing import Dict, List\n\nfrom typeguard import check_argument_types\nimport tensorflow as tf\nimport numpy as np\n\nfrom neuralmonkey.decoders.autoregressive import AutoregressiveDecoder\nfrom neuralmonkey.decorators import tensor\nfrom neuralmonkey.runners.base_runner import BaseRunner\n\n\nclass PerplexityRunner(BaseRunner[AutoregressiveDecoder]):\n\n # pylint: disable=too-few-public-methods\n # Pylint issue here: https:\/\/github.com\/PyCQA\/pylint\/issues\/2607\n class Executable(BaseRunner.Executable[\"PerplexityRunner\"]):\n\n def collect_results(self, results: List[Dict]) -> None:\n perplexities = np.mean(\n [2 ** res[\"xents\"] for res in results], axis=0)\n xent = float(np.mean([res[\"xents\"] for res in results]))\n self.set_runner_result(outputs=perplexities.tolist(),\n losses=[xent])\n # pylint: enable=too-few-public-methods\n\n def __init__(self,\n output_series: str,\n decoder: AutoregressiveDecoder) -> None:\n check_argument_types()\n BaseRunner[AutoregressiveDecoder].__init__(\n self, output_series, decoder)\n\n @tensor\n def fetches(self) -> Dict[str, tf.Tensor]:\n return {\"xents\": self.decoder.train_xents}\n\n @property\n def loss_names(self) -> List[str]:\n return [\"xents\"]\n","new_contents":"from typing import Dict, List\n\nfrom typeguard import check_argument_types\nimport tensorflow as tf\nimport numpy as np\n\nfrom neuralmonkey.decoders.autoregressive import AutoregressiveDecoder\nfrom neuralmonkey.decorators import tensor\nfrom neuralmonkey.runners.base_runner import BaseRunner\n\n\nclass PerplexityRunner(BaseRunner[AutoregressiveDecoder]):\n\n # pylint: disable=too-few-public-methods\n # Pylint issue here: https:\/\/github.com\/PyCQA\/pylint\/issues\/2607\n class Executable(BaseRunner.Executable[\"PerplexityRunner\"]):\n\n def collect_results(self, results: List[Dict]) -> None:\n perplexities = np.mean(\n [2 ** res[\"xents\"] for res in results], axis=0)\n xent = float(np.mean([res[\"xents\"] for res in results]))\n self.set_runner_result(outputs=perplexities.tolist(),\n losses=[xent])\n # pylint: enable=too-few-public-methods\n\n def __init__(self,\n output_series: str,\n decoder: AutoregressiveDecoder) -> None:\n check_argument_types()\n BaseRunner[AutoregressiveDecoder].__init__(\n self, output_series, decoder)\n\n @tensor\n def fetches(self) -> Dict[str, tf.Tensor]:\n return {\"xents\": self.decoder.train_xents}\n\n @property\n def loss_names(self) -> List[str]:\n return [\"xent\"]\n","subject":"Revert \"small fix in perplexity runner\"","message":"Revert \"small fix in perplexity runner\"\n\nThis reverts commit b195416761f12df496baa389df4686b2cf60c675.\n","lang":"Python","license":"bsd-3-clause","repos":"ufal\/neuralmonkey,ufal\/neuralmonkey,ufal\/neuralmonkey,ufal\/neuralmonkey,ufal\/neuralmonkey"} {"commit":"db537ab80444b9e4cc22f332577c2cba640fca0a","old_file":"tasks\/factory_utils.py","new_file":"tasks\/factory_utils.py","old_contents":"from factory import enums\nfrom collections import namedtuple\nimport gc\n\n\n# Factoryboy uses \"__\" and Salesforce uses \"__\". Luckily Factoryboy makes\n# theirs easy to override!\nenums.SPLITTER = \"____\"\n\n\n# More flexible than FactoryBoy's sequences because you can create and \n# destroy them where-ever you want.\nclass Adder:\n def __init__(self, x=0):\n self.x = x\n\n def __call__(self, value):\n self.x += value\n return int(self.x)\n\n def reset(self, x):\n self.x = x\n\n# Boilerplate that every factory would need to deal with.\ndef SessionBase(session):\n class BaseMeta:\n sqlalchemy_session = session\n sqlalchemy_session_persistence = \"commit\"\n\n return BaseMeta\n\n# Thin collector for the factories and a place to try to achieve better \n# scalability than the create_batch function from FactoryBoy.\nclass Factories:\n unflushed_record_counter = 0\n\n def __init__(self, session, namespace):\n self.session = session\n self.factory_classes = {\n key: value for key, value in namespace.items() if hasattr(value, \"generate_batch\")\n }\n\n def create_batch(self, classname, batchsize, **kwargs):\n cls = self.factory_classes.get(classname, None)\n assert cls, f\"Cannot find a factory class named {classname}. Did you misspell it?\"\n for _ in range(batchsize):\n cls.create(**kwargs)\n","new_contents":"from factory import enums\nfrom collections import namedtuple\nimport gc\n\n\n# Factoryboy uses \"__\" and Salesforce uses \"__\". Luckily Factoryboy makes\n# theirs easy to override!\nenums.SPLITTER = \"____\"\n\n\n# More flexible than FactoryBoy's sequences because you can create and \n# destroy them where-ever you want.\nclass Adder:\n def __init__(self, x=0):\n self.x = x\n\n def __call__(self, value):\n self.x += value\n return int(self.x)\n\n def reset(self, x):\n self.x = x\n\n# Boilerplate that every factory would need to deal with.\ndef SessionBase(session):\n class BaseMeta:\n sqlalchemy_session = session\n sqlalchemy_session_persistence = \"commit\"\n\n return BaseMeta\n\n# Thin collector for the factories and a place to try to achieve better \n# scalability than the create_batch function from FactoryBoy.\nclass Factories:\n unflushed_record_counter = 0\n\n def __init__(self, session, namespace):\n self.session = session\n self.factory_classes = {\n key: value for key, value in namespace.items() if hasattr(value, \"generate_batch\")\n }\n\n def create_batch(self, classname, batchsize, **kwargs):\n cls = self.factory_classes.get(classname, None)\n assert cls, f\"Cannot find a factory class named {classname}. Did you misspell it?\"\n for _ in range(batchsize):\n cls.create(**kwargs)\n\n def __getitem__(self, name):\n return self.factory_classes[name]\n","subject":"Make it easy to get a single item.","message":"Make it easy to get a single item.\n","lang":"Python","license":"bsd-3-clause","repos":"SalesforceFoundation\/Cumulus,SalesforceFoundation\/Cumulus,SalesforceFoundation\/Cumulus,SalesforceFoundation\/Cumulus"} {"commit":"180062c4d1159185ab113e98f41bb219d52086e8","old_file":"test.py","new_file":"test.py","old_contents":"from pyserializable import serialize, deserialize, autoserialized\nfrom pyserializable.util import repr_func\n\n\n@autoserialized\nclass Color:\n serial_format = 'r=uint:8, g=uint:8, b=uint:8, a=uint:8'\n serial_attr_converters = {'r': [int, str]}\n __repr__ = repr_func('r', 'g', 'b', 'a')\n\n\n@autoserialized\nclass Tile:\n serial_format = 'enabled=uint:1, color=Color, elite=uint:1'\n serial_fmt_converters = {'uint:1': [int, bool]}\n __repr__ = repr_func('enabled', 'color', 'elite')\n\nt = Tile()\nt.enabled = False\nt.elite = True\nt.color = Color()\nt.color.r = '201'\nt.color.g = 202\nt.color.b = 203\nt.color.a = 204\n\ndata = serialize(t)\n\n# Deserialize based on class\nt2 = deserialize(Tile, data)\n\n#Deserialize into existing instance\nt3 = Tile()\ndeserialize(t3, data)\n\nprint(t)\nprint(t2)\nprint(t3)\n","new_contents":"from pyserializable import serialize, deserialize, autoserialized\nfrom pyserializable.util import repr_func\n\n\n@autoserialized\nclass Color(object):\n serial_format = 'r=uint:8, g=uint:8, b=uint:8, a=uint:8'\n serial_attr_converters = {'r': [int, str]}\n __repr__ = repr_func('r', 'g', 'b', 'a')\n\n\n@autoserialized\nclass Tile(object):\n serial_format = 'enabled=uint:1, color=Color, elite=uint:1'\n serial_fmt_converters = {'uint:1': [int, bool]}\n __repr__ = repr_func('enabled', 'color', 'elite')\n\nt = Tile()\nt.enabled = False\nt.elite = True\nt.color = Color()\nt.color.r = '201'\nt.color.g = 202\nt.color.b = 203\nt.color.a = 204\n\ndata = serialize(t)\n\n# Deserialize based on class\nt2 = deserialize(Tile, data)\n\n#Deserialize into existing instance\nt3 = Tile()\ndeserialize(t3, data)\n\nprint(t)\nprint(t2)\nprint(t3)\n","subject":"Fix base class for python 2.x","message":"Fix base class for python 2.x\n","lang":"Python","license":"mit","repos":"numberoverzero\/origami"} {"commit":"42f67bdbf94a1a186518788f9685786b5c767eec","old_file":"performance\/web.py","new_file":"performance\/web.py","old_contents":"import requests\nfrom time import time\n\n\nclass Client:\n def __init__(self, host, requests, do_requests_counter):\n self.host = host\n self.requests = requests\n self.counter = do_requests_counter\n\n\nclass Request:\n GET = 'get'\n POST = 'post'\n\n def __init__(self, url, type=GET, data=None):\n self.url = url\n self.type = type\n self.data = data\n\n def do(self):\n try:\n data = ''\n if isinstance(self.data, RequestData):\n data = self.data.for_type(type=self.type)\n self.started = time()\n response = getattr(requests, self.type)(\n url=self.url,\n data=data\n )\n self.finished = time()\n self.status_code = response.status_code\n except AttributeError:\n raise RequestTypeError(type=self.type)\n\n def get_response_time(self):\n try:\n return self.finished - self.started\n except AttributeError:\n raise RequestTimeError\n\n\nclass RequestData:\n def __init__(self, data=None):\n self.data = data\n\n def for_type(self, type=Request.GET):\n if type is Request.GET:\n return data\n\n\nclass RequestTypeError(Exception):\n def __init__(self, type):\n self.type = type\n\n def __str__(self):\n return 'Invalid request type \"%s\"' % self.type\n\n\nclass RequestTimeError(Exception):\n pass\n","new_contents":"import requests\nfrom time import time\n\n\nclass Client:\n def __init__(self, host, requests, do_requests_counter):\n self.host = host\n self.requests = requests\n self.counter = do_requests_counter\n\n\nclass Request:\n GET = 'get'\n POST = 'post'\n\n def __init__(self, url, type=GET, data=None):\n self.url = url\n self.type = type\n self.data = data\n\n def do(self):\n try:\n data = ''\n if isinstance(self.data, RequestData):\n data = self.data.for_type(type=self.type)\n started = time()\n response = getattr(requests, self.type)(\n url=self.url,\n data=data\n )\n finished = time()\n return finished - started\n except AttributeError:\n raise RequestTypeError(type=self.type)\n\n\nclass RequestData:\n def __init__(self, data=None):\n self.data = data\n\n def for_type(self, type=Request.GET):\n if type is Request.GET:\n return data\n\n\nclass RequestTypeError(Exception):\n def __init__(self, type):\n self.type = type\n\n def __str__(self):\n return 'Invalid request type \"%s\"' % self.type\n\n\nclass RequestTimeError(Exception):\n pass\n","subject":"Update Request, only return response time","message":"Update Request, only return response time\n\nRequest: remove status_code and get_response_time() and only return the\nresponse time on do() function\n","lang":"Python","license":"mit","repos":"BakeCode\/performance-testing,BakeCode\/performance-testing"} {"commit":"24fa27e05e0ed58e955ed6365de101b2e9653a7b","old_file":"cli.py","new_file":"cli.py","old_contents":"#!\/usr\/bin\/env python\nimport sys,os\nfrom copy import deepcopy\nfrom scrabble import make_board,top_moves,read_dictionary\n\ndef show_board(board,play=None):\n if not play:\n for row in board:\n print ''.join(row)\n else:\n b = deepcopy(board)\n for x,r,c in play:\n b[r][c] = x.lower()\n show_board(b)\n\nif __name__ == '__main__':\n assert len(sys.argv) >= 3, 'Usage: .\/scrabble.py boardfile hand [num_moves]'\n board = make_board(open(sys.argv[1]))\n hand = sys.argv[2].upper()\n path = os.path.dirname(sys.argv[0])\n num_moves = int(sys.argv[3]) if len(sys.argv) > 3 else 20\n for score,words,play in top_moves(board,read_dictionary(path),hand,num_moves):\n print score, ', '.join(words)\n show_board(board,play)\n print ''\n","new_contents":"#!\/usr\/bin\/env python\nimport os\nfrom copy import deepcopy\nfrom optparse import OptionParser\nfrom scrabble import make_board,top_moves,read_dictionary\n\n\ndef show_board(board, play=None):\n if not play:\n for row in board:\n print ''.join(row)\n else:\n b = deepcopy(board)\n for x,r,c in play:\n b[r][c] = x.lower()\n show_board(b)\n\n\ndef main():\n op = OptionParser(usage='%prog [-n 20] boardfile hand')\n op.add_option('-n', '--num-plays', type=int, default=20,\n help='Number of possible plays to display')\n opts, args = op.parse_args()\n if len(args) != 2:\n op.error('Must provide boardfile and hand as arguments')\n board = make_board(open(args[0]))\n hand = args[1].upper()\n word_list = read_dictionary(os.path.dirname(__file__))\n for score, words, play in top_moves(board, word_list, hand, opts.num_plays):\n print score, ', '.join(words)\n show_board(board, play)\n print ''\n\nif __name__ == '__main__':\n main()\n\n","subject":"Use optparse instead of hand-rolled options","message":"Use optparse instead of hand-rolled options\n","lang":"Python","license":"mit","repos":"perimosocordiae\/wwf,perimosocordiae\/wwf"} {"commit":"1bd540f43c25dec125085acee7bbe0904363c204","old_file":"test.py","new_file":"test.py","old_contents":"import unittest\n\nfrom enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen\n\n\nclass RotorTestCase(unittest.TestCase):\n def test_rotor_encoding(self):\n rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')\n self.assertEqual('E', rotor.encode('A'))\n\n\n def test_rotor_reverse_encoding(self):\n rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')\n self.assertEqual('U', rotor.encode_reverse('A'))\n\n\ndef run_tests():\n runner = unittest.TextTestRunner()\n suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)\n runner.run(suite)\n\n\nif __name__ == '__main__': # pragma: no cover\n run_tests()","new_contents":"import unittest\n\nfrom enigma import Enigma, Steckerbrett, Umkehrwalze, Walzen\n\n\nclass RotorTestCase(unittest.TestCase):\n def test_rotor_encoding(self):\n rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')\n self.assertEqual('E', rotor.encode('A'))\n\n\n def test_rotor_reverse_encoding(self):\n rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q')\n self.assertEqual('U', rotor.encode_reverse('A'))\n\n def test_rotor_different_setting(self):\n rotor = Walzen(wiring='EKMFLGDQVZNTOWYHXUSPAIBRCJ', notch='Q',\n setting='B')\n self.assertEqual('K', rotor.encode('A'))\n self.assertEqual('K', rotor.encode_reverse('A'))\n\n\ndef run_tests():\n runner = unittest.TextTestRunner()\n suite = unittest.TestLoader().loadTestsFromTestCase(RotorTestCase)\n runner.run(suite)\n\n\nif __name__ == '__main__': # pragma: no cover\n run_tests()","subject":"Test if rotor encodes with different setting properly","message":"Test if rotor encodes with different setting properly\n","lang":"Python","license":"mit","repos":"ranisalt\/enigma"} {"commit":"0b2d523665f3989d9375cd83970400ace0e40336","old_file":"sites\/docs\/conf.py","new_file":"sites\/docs\/conf.py","old_contents":"# Obtain shared config values\nimport os, sys\nsys.path.append(os.path.abspath('..'))\nsys.path.append(os.path.abspath('..\/..'))\nfrom shared_conf import *\n\n# Enable autodoc, intersphinx\nextensions.extend(['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'])\n\n# Autodoc settings\nautodoc_default_flags = ['members', 'special-members']\n","new_contents":"# Obtain shared config values\nimport os, sys\nsys.path.append(os.path.abspath('..'))\nsys.path.append(os.path.abspath('..\/..'))\nfrom shared_conf import *\n\n# Enable autodoc, intersphinx\nextensions.extend(['sphinx.ext.autodoc', 'sphinx.ext.intersphinx'])\n\n# Autodoc settings\nautodoc_default_flags = ['members', 'special-members']\n\n# Intersphinx connection to stdlib\nintersphinx_mapping = {\n 'python': ('http:\/\/docs.python.org\/2.6', None),\n}\n","subject":"Connect to Python stdlib intersphinx","message":"Connect to Python stdlib intersphinx\n","lang":"Python","license":"lgpl-2.1","repos":"torkil\/paramiko,ameily\/paramiko,SebastianDeiss\/paramiko,dorianpula\/paramiko,remram44\/paramiko,zarr12steven\/paramiko,mirrorcoder\/paramiko,fvicente\/paramiko,davidbistolas\/paramiko,CptLemming\/paramiko,redixin\/paramiko,digitalquacks\/paramiko,reaperhulk\/paramiko,thisch\/paramiko,rcorrieri\/paramiko,mhdaimi\/paramiko,paramiko\/paramiko,jaraco\/paramiko,anadigi\/paramiko,Automatic\/paramiko,toby82\/paramiko,varunarya10\/paramiko,thusoy\/paramiko,selboo\/paramiko,jorik041\/paramiko,esc\/paramiko,dlitz\/paramiko,zpzgone\/paramiko"} {"commit":"890c0d919aba273489c27243c513c822399c5d35","old_file":"datasurvey\/packages.py","new_file":"datasurvey\/packages.py","old_contents":"import gzip\nimport bz2\nimport rarfile\nimport zipfile\nimport tarfile\n\nclass PackageHandler:\n def __init__(self, buffer):\n self.buffer = buffer\n\n def __iter__(self):\n for name in self.archive.namelist():\n with self.archive.open(name) as ar:\n # TODO: Handle archives\n outbuf = ar.read(1000)\n\n yield (name, outbuf)\n\nclass ZipHandler(PackageHandler):\n MIMETYPES = [\"application\/zip\"]\n def __init__(self, buffer):\n PackageHandler.__init__(self, buffer)\n self.archive = zipfile.ZipFile(buffer)\n\nclass RarHandler(PackageHandler):\n MIMETYPES = [\"application\/x-rar-compressed\"]\n def __init__(self, buffer):\n PackageHandler.__init__(self, buffer)\n self.archive = zipfile.RarFile(buffer)\n\n\n\npackage_handlers = {}\n\ndef register_handler(handler):\n for mt in handler.MIMETYPES:\n package_handlers[mt] = handler\n\nregister_handler(ZipHandler)\nregister_handler(RarHandler)\n","new_contents":"import gzip\nimport bz2\nimport rarfile\nimport zipfile\nimport tarfile\n\nclass PackageHandler:\n def __init__(self, buffer):\n self.buffer = buffer\n\n def __iter__(self):\n for name in self.archive.namelist():\n info = self.archive.getinfo(name)\n if hasattr(info, 'isdir') and info.isdir():\n continue\n if name[-1] == \"\/\":\n continue\n with self.archive.open(name) as ar:\n # TODO: Handle archives\n outbuf = ar.read(1000)\n\n yield (name, outbuf)\n\nclass ZipHandler(PackageHandler):\n MIMETYPES = [\"application\/zip\"]\n def __init__(self, buffer):\n PackageHandler.__init__(self, buffer)\n self.archive = zipfile.ZipFile(buffer)\n\nclass RarHandler(PackageHandler):\n MIMETYPES = [\"application\/x-rar-compressed\"]\n def __init__(self, buffer):\n PackageHandler.__init__(self, buffer)\n self.archive = zipfile.RarFile(buffer)\n\n\n\npackage_handlers = {}\n\ndef register_handler(handler):\n for mt in handler.MIMETYPES:\n package_handlers[mt] = handler\n\nregister_handler(ZipHandler)\nregister_handler(RarHandler)\n","subject":"Fix recursion issue with zip\/rar files","message":"Fix recursion issue with zip\/rar files\n","lang":"Python","license":"mit","repos":"occrp\/datasurvey"} {"commit":"bb7de7e76302fbd3eeeeb740d00c234faadef4ef","old_file":"tests\/test_nonsensefilter.py","new_file":"tests\/test_nonsensefilter.py","old_contents":"from unittest import TestCase\n\nfrom spicedham.nonsensefilter import NonsenseFilter\n\nclass TestNonsenseFilter(TestCase):\n \n # TODO: This test will likely fail spectacularly because of a lack of\n # training.\n def test_classify(self):\n nonsense = NonsenseFilter()\n nonsense.filter_match = 1\n nonsense.filter_miss = 0\n reverse = lambda x: x[::-1]\n match_message = map(reverse, ['supposedly', 'nonsense', 'words'])\n miss_message = ['Firefox']\n self.assertEqual(nonsense.classify(match_message), 1)\n self.assertEqual(nonsense.classify(miss_message), 0)\n\n","new_contents":"from tests.test_classifierbase import TestClassifierBase\n\nfrom spicedham.backend import load_backend\nfrom spicedham.nonsensefilter import NonsenseFilter\n\nclass TestNonsenseFilter(TestClassifierBase):\n\n def test_train(self):\n backend = load_backend()\n nonsense = NonsenseFilter()\n alphabet = map(chr, range(97, 123))\n reversed_alphabet = reversed(alphabet)\n self._training(nonsense, alphabet, reversed_alphabet)\n for letter in alphabet:\n self.assertEqual(True,\n backend.get_key(nonsense.__class__.__name__, letter))\n \n def test_classify(self):\n nonsense = NonsenseFilter()\n nonsense.filter_match = 1\n nonsense.filter_miss = 0\n alphabet = map(chr, range(97, 123))\n reversed_alphabet = reversed(alphabet)\n self._training(nonsense, alphabet, reversed_alphabet)\n match_message = ['not', 'in', 'training', 'set']\n miss_message = ['a']\n self.assertEqual(nonsense.classify(match_message), 1)\n self.assertEqual(nonsense.classify(miss_message), 0)\n","subject":"Add a base class and a test_train function","message":"Add a base class and a test_train function\n\nOverall, fix a very incomplete test.\n","lang":"Python","license":"mpl-2.0","repos":"mozilla\/spicedham,mozilla\/spicedham"} {"commit":"4672e447617e754d6b4d229ce775fbf9ee0b35aa","old_file":"tests\/test_requesthandler.py","new_file":"tests\/test_requesthandler.py","old_contents":"from unittest import TestCase\nfrom ppp_datamodel.communication import Request\nfrom ppp_datamodel import Triple, Resource, Missing\nfrom ppp_libmodule.tests import PPPTestCase\n\nfrom ppp_spell_checker import app\n\nclass RequestHandlerTest(PPPTestCase(app)):\n def testCorrectSentence(self):\n original = 'What is the birth date of George Washington'\n j = {'id': '1', 'language': 'en', 'measures': {}, 'trace': [],\n 'tree': {'type': 'sentence', 'value': original}}\n answer = self.request(j)\n self.assertEquals(len(answer), 0)\n\n def testWrongSentence(self):\n original = 'What is the bitrh date of George Washington'\n expected = 'What is the birth date of George Washington'\n j = {'id': '1', 'language': 'en', 'measures': {}, 'trace': [],\n 'tree': {'type': 'sentence', 'value': original}}\n answer = self.request(j)\n self.assertEquals(len(answer), 1)\n self.assertIsInstance(answer[0].tree, Resource)\n result = answer[0].tree.__getattr__('value')\n self.assertEqual(result, expected)\n","new_contents":"from unittest import TestCase\nfrom ppp_datamodel.communication import Request\nfrom ppp_datamodel import Triple, Resource, Missing\nfrom ppp_libmodule.tests import PPPTestCase\n\nfrom ppp_spell_checker import app\n\nclass RequestHandlerTest(PPPTestCase(app)):\n def testCorrectSentence(self):\n original = 'What is the birth date of George Washington'\n j = {'id': '1', 'language': 'en', 'measures': {}, 'trace': [],\n 'tree': {'type': 'sentence', 'value': original}}\n answer = self.request(j)\n self.assertEquals(len(answer), 0)\n\n def testWrongSentence(self):\n original = 'What is the bitrh date of George Washington'\n expected = 'What is the birth date of George Washington'\n j = {'id': '1', 'language': 'en', 'measures': {}, 'trace': [],\n 'tree': {'type': 'sentence', 'value': original}}\n answer = self.request(j)\n self.assertEquals(len(answer), 1)\n self.assertIsInstance(answer[0].tree, Resource)\n result = answer[0].tree.__getattr__('value')\n self.assertEqual(result, expected)\n\n def testIrrelevantInput(self):\n original = 'What is the birth date of George Washington'\n j = {'id': '1', 'language': 'en', 'measures': {}, 'trace': [],\n 'tree': {'type': 'resource', 'value': original}}\n answer = self.request(j)\n self.assertEquals(len(answer), 0)\n","subject":"Add test for irrelevant input.","message":"Add test for irrelevant input.\n","lang":"Python","license":"mit","repos":"ProjetPP\/PPP-Spell-Checker,ProjetPP\/PPP-Spell-Checker"} {"commit":"978dd0161552458331870af0b524cdcff25fd71d","old_file":"furious\/handlers\/__init__.py","new_file":"furious\/handlers\/__init__.py","old_contents":"#\n# Copyright 2012 WebFilings, LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nimport json\n\nimport logging\n\nfrom ..async import Async\nfrom ..processors import run_job\n\n\ndef process_async_task(headers, request_body):\n \"\"\"Process an Async task and execute the requested function.\"\"\"\n async_options = json.loads(request_body)\n work = Async.from_dict(async_options)\n\n logging.info(work._function_path)\n\n run_job(work)\n\n return 200, work._function_path\n\n\n\n","new_contents":"#\n# Copyright 2012 WebFilings, LLC\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n\nimport json\n\nimport logging\n\nfrom ..async import Async\nfrom ..processors import run_job\n\n\ndef process_async_task(headers, request_body):\n \"\"\"Process an Async task and execute the requested function.\"\"\"\n async_options = json.loads(request_body)\n async = Async.from_dict(async_options)\n\n logging.info(work._function_path)\n\n with context.job_context_from_async(async):\n run_job(async)\n\n return 200, work._function_path\n\n\n\n","subject":"Adjust async run handler to use the JobContext manager.","message":"Adjust async run handler to use the JobContext manager.\n","lang":"Python","license":"apache-2.0","repos":"mattsanders-wf\/furious,beaulyddon-wf\/furious,Workiva\/furious,andreleblanc-wf\/furious,rosshendrickson-wf\/furious,andreleblanc-wf\/furious,beaulyddon-wf\/furious,robertkluin\/furious,rosshendrickson-wf\/furious,Workiva\/furious,mattsanders-wf\/furious"} {"commit":"713b91c4d7dc3737223bc70aa329ec9de2c48fb8","old_file":"mycli\/packages\/special\/utils.py","new_file":"mycli\/packages\/special\/utils.py","old_contents":"import os\nimport subprocess\n\ndef handle_cd_command(arg):\n \"\"\"Handles a `cd` shell command by calling python's os.chdir.\"\"\"\n CD_CMD = 'cd'\n directory = ''\n error = False\n\n tokens = arg.split(CD_CMD + ' ')\n directory = tokens[-1]\n\n try:\n os.chdir(directory)\n subprocess.call(['pwd'])\n return True, None\n except OSError as e:\n return False, e.strerror\n","new_contents":"import os\nimport subprocess\n\ndef handle_cd_command(arg):\n \"\"\"Handles a `cd` shell command by calling python's os.chdir.\"\"\"\n CD_CMD = 'cd'\n tokens = arg.split(CD_CMD + ' ')\n directory = tokens[-1] if len(tokens) > 1 else None\n if not directory:\n return False, \"No folder name was provided.\"\n try:\n os.chdir(directory)\n subprocess.call(['pwd'])\n return True, None\n except OSError as e:\n return False, e.strerror\n","subject":"Add validation for 'cd' command argument","message":"Add validation for 'cd' command argument\n","lang":"Python","license":"bsd-3-clause","repos":"mdsrosa\/mycli,mdsrosa\/mycli"} {"commit":"75b1fdc9f290c85b4d469cdce5e5d1154aed4881","old_file":"indra\/tests\/test_util.py","new_file":"indra\/tests\/test_util.py","old_contents":"from __future__ import absolute_import, print_function, unicode_literals\nfrom builtins import dict, str\nimport xml.etree.ElementTree as ET\nfrom indra.util import UnicodeXMLTreeBuilder as UTB\nfrom indra.util import unicode_strs\nfrom io import BytesIO\n\ndef test_unicode_tree_builder():\n xml = u'<html><bar>asdf<\/bar><\/html>'.encode('utf-8')\n xml_io = BytesIO(xml)\n tree = ET.parse(xml_io, parser=UTB())\n bar = tree.find('.\/\/bar')\n assert unicode_strs(bar)\n\n","new_contents":"from __future__ import absolute_import, print_function, unicode_literals\nfrom builtins import dict, str\n\nimport json\nimport xml.etree.ElementTree as ET\nfrom indra.util import UnicodeXMLTreeBuilder as UTB, kappy_json_to_graph\nfrom indra.util import unicode_strs\nfrom io import BytesIO\n\n\ndef test_unicode_tree_builder():\n xml = u'<html><bar>asdf<\/bar><\/html>'.encode('utf-8')\n xml_io = BytesIO(xml)\n tree = ET.parse(xml_io, parser=UTB())\n bar = tree.find('.\/\/bar')\n assert unicode_strs(bar)\n\n\ndef test_kappy_influence_json_to_graph():\n with open('kappy_influence.json', 'r') as f:\n imap = json.load(f)\n graph = kappy_json_to_graph(imap)\n assert graph is not None, 'No graph produced.'\n n_nodes = len(graph.nodes)\n n_edges = len(graph.edges)\n assert n_nodes == 4, \\\n 'Wrong number (%d vs. %d) of nodes on the graph.' % (n_nodes, 4)\n assert n_edges == 6, \\\n \"Wrong number (%d vs. %d) of edges on graph.\" % (n_edges, 4)\n\n","subject":"Implement test for basic graph.","message":"Implement test for basic graph.\n","lang":"Python","license":"bsd-2-clause","repos":"pvtodorov\/indra,johnbachman\/indra,sorgerlab\/belpy,pvtodorov\/indra,sorgerlab\/indra,sorgerlab\/indra,pvtodorov\/indra,bgyori\/indra,sorgerlab\/indra,sorgerlab\/belpy,johnbachman\/indra,pvtodorov\/indra,bgyori\/indra,johnbachman\/belpy,johnbachman\/belpy,sorgerlab\/belpy,johnbachman\/belpy,bgyori\/indra,johnbachman\/indra"} {"commit":"052367e1239e918cdcb9106b4494a48e34e92643","old_file":"pychecker2\/File.py","new_file":"pychecker2\/File.py","old_contents":"from pychecker2.util import type_filter\n\nfrom compiler import ast\n\nclass File:\n def __init__(self, name):\n self.name = name\n self.parseTree = None\n self.scopes = {}\n self.root_scope = None\n self.warnings = []\n\n def __cmp__(self, other):\n return cmp(self.name, other.name)\n\n def warning(self, line, warn, *args):\n\ttry:\n\t line = line.lineno\n except AttributeError:\n\t pass\n self.warnings.append( (line, warn, args) )\n\n def scope_filter(self, type):\n return [(n, s)\n for n, s in self.scopes.iteritems() if isinstance(n, type)\n ]\n\n def function_scopes(self):\n return self.scope_filter(ast.Function)\n\n def class_scopes(self):\n return self.scope_filter(ast.Class)\n","new_contents":"from pychecker2.util import parents\n\nfrom compiler import ast\n\nclass File:\n def __init__(self, name):\n self.name = name\n self.parseTree = None\n self.scopes = {}\n self.root_scope = None\n self.warnings = []\n\n def __cmp__(self, other):\n return cmp(self.name, other.name)\n\n def warning(self, line, warn, *args):\n lineno = line\n\ttry:\n\t lineno = line.lineno\n except AttributeError:\n\t pass\n if not lineno:\n try:\n for p in parents(line):\n if p.lineno:\n lineno = p.lineno\n break\n except AttributeError:\n pass\n self.warnings.append( (lineno, warn, args) )\n\n def scope_filter(self, type):\n return [(n, s)\n for n, s in self.scopes.iteritems() if isinstance(n, type)\n ]\n\n def function_scopes(self):\n return self.scope_filter(ast.Function)\n\n def class_scopes(self):\n return self.scope_filter(ast.Class)\n","subject":"Add more ways to suck line numbers from nodes","message":"Add more ways to suck line numbers from nodes\n","lang":"Python","license":"bsd-3-clause","repos":"akaihola\/PyChecker,thomasvs\/pychecker,akaihola\/PyChecker,thomasvs\/pychecker"} {"commit":"68dee1cfef3a765a82fe7b51395219f8e1270a12","old_file":"reason\/__init__.py","new_file":"reason\/__init__.py","old_contents":"import os\nfrom flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_login import LoginManager\nfrom datetime import datetime\nfrom flask.ext.moment import Moment\nfrom flask_debugtoolbar import DebugToolbarExtension\n\n\nbasedir = os.path.abspath(os.path.dirname(__file__))\napp = Flask(__name__)\napp.config['SECRET_KEY'] = os.urandom(60)\n\nuserDb = 'sqlite:\/\/\/' + os.path.join(basedir, 'db\/reasonUser.db')\nspdxDb = 'postgresql:\/\/spdx:spdx@host:5432\/spdx'\n\ndatabases = {\n 'userDb': userDb,\n 'spdxDb': spdxDb \n}\n\n\n\napp.config['SQLALCHEMY_BINDS'] = databases\n# Debug configuration\napp.config['DEBUG'] = False\ndb = SQLAlchemy(app)\n\nlogin_manager = LoginManager()\nlogin_manager.session_protection = \"strong\"\nlogin_manager.login_view = \"login\"\nlogin_manager.init_app(app)\n# Debug Configuration\n#DEBUG_TB_INTERCEPT_REDIRECTS = False\n#toolbar = DebugToolbarExtension(app)\n\nALLOWED_EXTENSIONS = set(['xml', 'jar'])\nmoment = Moment(app)\n\n\n","new_contents":"import os\nfrom flask import Flask\nfrom flask_sqlalchemy import SQLAlchemy\nfrom flask_login import LoginManager\nfrom datetime import datetime\nfrom flask.ext.moment import Moment\nfrom flask_debugtoolbar import DebugToolbarExtension\n\n\nbasedir = os.path.abspath(os.path.dirname(__file__))\napp = Flask(__name__)\napp.config['SECRET_KEY'] = os.urandom(60)\n\nnvdDb = 'sqlite:\/\/\/' + os.path.join(basedir, 'db\/nvd.vulnerabilities.db')\nuserDb = 'sqlite:\/\/\/' + os.path.join(basedir, 'db\/reasonUser.db')\nspdxDb = 'postgresql:\/\/spdx:spdx@host:5432\/spdx'\n\n# Initial URI was for searching from NVD Database, Made changes to connect to Users db using binds\n# App.config['SQLALCHEMY_DATABASE_URI'] = 'sqlite:\/\/\/' + os.path.join(basedir, 'nvd.vulnerabilities.db')\n\ndatabases = {\n 'nvdDb': nvdDb,\n 'userDb': userDb,\n 'spdxDb': spdxDb\n \n}\n\n\n\napp.config['SQLALCHEMY_BINDS'] = databases\n# Debug configuration\napp.config['DEBUG'] = False\ndb = SQLAlchemy(app)\n\nlogin_manager = LoginManager()\nlogin_manager.session_protection = \"strong\"\nlogin_manager.login_view = \"login\"\nlogin_manager.init_app(app)\n# Debug Configuration\n#DEBUG_TB_INTERCEPT_REDIRECTS = False\n#toolbar = DebugToolbarExtension(app)\n\nALLOWED_EXTENSIONS = set(['xml', 'jar'])\nmoment = Moment(app)\n\n\n","subject":"Revert \"Removed connections to nvd database\"","message":"Revert \"Removed connections to nvd database\"\n\nThis reverts commit 4b2edb4bc6d9c525d0d5c825174fbd5dbfac0ac1.\n","lang":"Python","license":"mit","repos":"pombredanne\/reason,pombredanne\/reason,pombredanne\/reason"} {"commit":"381bb4e11cf6951d819fa2cf298e2cc558464fc9","old_file":"utils\/nflc-get-categories.py","new_file":"utils\/nflc-get-categories.py","old_contents":"#!\/usr\/bin\/env python3\n\nimport argparse\nimport json\nfrom urllib.request import urlopen\n\n\ndef get_data(domain):\n response = urlopen('http:\/\/{}\/media\/nflc-playlist-video.json'.format(domain)).read()\n return json.loads(response.decode('utf-8'))\n\n\ndef main():\n parser = argparse.ArgumentParser(description='Get the category names and IDs for a NFLC site')\n parser.add_argument('domain', type=str, nargs=1, help='Domain name to query for')\n args = parser.parse_args()\n\n data = get_data(args.domain[0])\n result = {}\n strip_left = [\n 'Podcast - ',\n 'Video - Show - ',\n 'Video - Shows - ',\n 'Video - ',\n 'Videos - Show - ',\n 'Videos - ',\n ]\n\n for category_id, category in data.items():\n name = category['name']\n for strip in strip_left:\n if name.startswith(strip):\n name = name[(len(strip)):]\n\n result[name.strip()] = category_id\n\n for category_name in sorted(result):\n print('({}, \"{}\"),'.format(result[category_name], category_name))\n\n\nif __name__ == '__main__':\n main()\n","new_contents":"#!\/usr\/bin\/env python3\n\nimport argparse\nimport json\nfrom urllib.request import urlopen\n\n\ndef get_data(domain):\n response = urlopen('http:\/\/{}\/media\/nflc-playlist-video.json'.format(domain)).read()\n return json.loads(response.decode('utf-8'))\n\n\ndef main():\n parser = argparse.ArgumentParser(description='Get the category names and IDs for a NFLC site')\n parser.add_argument('domain', type=str, nargs=1, help='Domain name to query for')\n args = parser.parse_args()\n\n data = get_data(args.domain[0])\n result = {}\n strip_left = [\n 'Podcast - ',\n 'Video - ',\n 'Videos - ',\n 'Show - ',\n 'Shows - ',\n ]\n\n for category_id, category in data.items():\n name = category['name']\n for strip in strip_left:\n if name.startswith(strip):\n name = name[(len(strip)):]\n\n result[name.strip()] = category_id\n\n for category_name in sorted(result, key=str.lower):\n print('({}, \"{}\"),'.format(result[category_name], category_name))\n\n\nif __name__ == '__main__':\n main()\n","subject":"Simplify NFLC utils script stripping and sort categories case-insensitive","message":"Simplify NFLC utils script stripping and sort categories case-insensitive\n","lang":"Python","license":"mit","repos":"Tenzer\/plugin.video.nfl-teams"} {"commit":"a0c2e64c92d89276d73b5e4ca31e10a352ab37f1","old_file":"analyser\/api.py","new_file":"analyser\/api.py","old_contents":"import os\nimport requests\n\nfrom flask import Blueprint\n\nfrom utils.decorators import validate, require\nfrom utils.validators import validate_url\n\nfrom .parser import Parser\n\nendpoint = Blueprint('analyse_url', __name__)\n\n\n@endpoint.route('analyse\/', methods=['POST'])\n@require('url')\n@validate({\n 'url': validate_url\n})\ndef analyse_url(url):\n name, ext = os.path.splitext(url)\n parse = Parser(ext=ext[1:])\n\n response = requests.get(url, stream=True)\n fields = []\n for chunk in response.iter_lines(1024):\n fields = parse(chunk)\n if fields:\n break\n\n print fields\n\n return url\n","new_contents":"import os\nimport json\n\nimport requests\nimport rethinkdb as r\n\nfrom flask import Blueprint, current_app\n\nfrom utils.decorators import validate, require\nfrom utils.validators import validate_url\n\nfrom krunchr.vendors.rethinkdb import db\n\nfrom .parser import Parser\nfrom .tasks import get_file\n\nendpoint = Blueprint('analyse_url', __name__)\n\n\n@endpoint.route('analyse\/', methods=['POST'])\n@require('url')\n@validate({\n 'url': validate_url\n})\ndef analyse_url(url):\n name, ext = os.path.splitext(url)\n parse = Parser(ext=ext[1:])\n\n response = requests.get(url, stream=True)\n fields = []\n for chunk in response.iter_lines(1024):\n fields = parse(chunk)\n if fields:\n break\n\n task_id = get_file.delay(url, current_app.config['DISCO_FILES']).task_id\n r.table('jobs').insert({\n 'url': url,\n 'task_id': task_id,\n 'state': 'starting'\n }).run(db.conn)\n return json.dumps(fields)\n","subject":"Put job id in rethink db","message":"Put job id in rethink db\n","lang":"Python","license":"apache-2.0","repos":"vtemian\/kruncher"} {"commit":"de69c4048fe8533185a4eca6f98c7d74967618bf","old_file":"opentreemap\/opentreemap\/util.py","new_file":"opentreemap\/opentreemap\/util.py","old_contents":"from django.views.decorators.csrf import csrf_exempt\nimport json\n\n\ndef route(**kwargs):\n @csrf_exempt\n def routed(request, *args2, **kwargs2):\n method = request.method\n req_method = kwargs[method]\n return req_method(request, *args2, **kwargs2)\n return routed\n\n\ndef json_from_request(request):\n body = request.body\n\n if body:\n return json.loads(body)\n else:\n return None\n\n\ndef merge_view_contexts(viewfns):\n def wrapped(*args, **kwargs):\n context = {}\n for viewfn in viewfns:\n context.update(viewfn(*args, **kwargs))\n\n return context\n return wrapped\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import print_function\nfrom __future__ import unicode_literals\nfrom __future__ import division\n\nimport json\n\nfrom django.views.decorators.csrf import csrf_exempt\nfrom django.http import HttpResponse, HttpResponseRedirect, Http404\n\n\ndef route(**kwargs):\n @csrf_exempt\n def routed(request, *args2, **kwargs2):\n method = request.method\n\n if method not in kwargs:\n raise Http404()\n else:\n req_method = kwargs[method]\n return req_method(request, *args2, **kwargs2)\n return routed\n\n\ndef json_from_request(request):\n body = request.body\n\n if body:\n return json.loads(body)\n else:\n return None\n\n\ndef merge_view_contexts(viewfns):\n def wrapped(*args, **kwargs):\n context = {}\n for viewfn in viewfns:\n context.update(viewfn(*args, **kwargs))\n\n return context\n return wrapped\n","subject":"Return a 404, not a 500 on a verb mismatch","message":"Return a 404, not a 500 on a verb mismatch\n\nFixes #1101\n","lang":"Python","license":"agpl-3.0","repos":"maurizi\/otm-core,maurizi\/otm-core,clever-crow-consulting\/otm-core,recklessromeo\/otm-core,clever-crow-consulting\/otm-core,recklessromeo\/otm-core,RickMohr\/otm-core,recklessromeo\/otm-core,recklessromeo\/otm-core,maurizi\/otm-core,RickMohr\/otm-core,RickMohr\/otm-core,clever-crow-consulting\/otm-core,RickMohr\/otm-core,clever-crow-consulting\/otm-core,maurizi\/otm-core"} {"commit":"0019cfc3d6bc8b81520e09daa0b84662d76aff93","old_file":"opps\/core\/models\/publishable.py","new_file":"opps\/core\/models\/publishable.py","old_contents":"#!\/usr\/bin\/env python\nfrom django.db import models\nfrom django.conf import settings\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.contrib.sites.models import Site\n\nfrom opps.core.models.date import Date\n\nfrom datetime import datetime\n\n\n\nclass PublishableMnager(models.Manager):\n def all_published(self):\n return super(PublishableMnager, self).get_query_set().filter(\n date_available__lte=datetime.now(), published=True)\n\n\nclass Publishable(Date):\n\n user = models.ForeignKey(settings.AUTH_USER_MODEL)\n site = models.ForeignKey(Site, default=0)\n date_available = models.DateTimeField(_(u\"Date available\"),\n default=datetime.now, null=True)\n published = models.BooleanField(_(u\"Published\"), default=False)\n\n objects = PublishableMnager()\n\n class Meta:\n abstract = True\n\n def is_published(self):\n return self.published and \\\n self.date_available.replace(tzinfo=None) <= datetime.now()\n","new_contents":"#!\/usr\/bin\/env python\nfrom django.db import models\nfrom django.conf import settings\nfrom django.utils.translation import ugettext_lazy as _\nfrom django.contrib.sites.models import Site\n\nfrom opps.core.models.date import Date\n\nfrom datetime import datetime\n\n\n\nclass PublishableManager(models.Manager):\n def all_published(self):\n return super(PublishableManager, self).get_query_set().filter(\n date_available__lte=datetime.now(), published=True)\n\n\nclass Publishable(Date):\n\n user = models.ForeignKey(settings.AUTH_USER_MODEL)\n site = models.ForeignKey(Site, default=0)\n date_available = models.DateTimeField(_(u\"Date available\"),\n default=datetime.now, null=True)\n published = models.BooleanField(_(u\"Published\"), default=False)\n\n objects = PublishableManager()\n\n class Meta:\n abstract = True\n\n def is_published(self):\n return self.published and \\\n self.date_available.replace(tzinfo=None) <= datetime.now()\n","subject":"Fix spelling error on core published PublishableMnager to PublishableManager","message":"Fix spelling error on core published PublishableMnager to PublishableManager\n","lang":"Python","license":"mit","repos":"jeanmask\/opps,YACOWS\/opps,jeanmask\/opps,opps\/opps,opps\/opps,YACOWS\/opps,opps\/opps,williamroot\/opps,williamroot\/opps,jeanmask\/opps,YACOWS\/opps,YACOWS\/opps,williamroot\/opps,opps\/opps,williamroot\/opps,jeanmask\/opps"} {"commit":"ee33022db50a66b6e2db12972a2ed107276cc666","old_file":"apps\/common\/tests\/python\/mediawords\/key_value_store\/test_cached_amazon_s3.py","new_file":"apps\/common\/tests\/python\/mediawords\/key_value_store\/test_cached_amazon_s3.py","old_contents":"from mediawords.key_value_store.cached_amazon_s3 import CachedAmazonS3Store\nfrom .amazon_s3_credentials import (\n TestAmazonS3CredentialsTestCase,\n get_test_s3_credentials,\n)\n\ntest_credentials = get_test_s3_credentials()\n\n\nclass TestCachedAmazonS3StoreTestCase(TestAmazonS3CredentialsTestCase):\n def _initialize_store(self) -> CachedAmazonS3Store:\n return CachedAmazonS3Store(\n access_key_id=test_credentials.access_key_id(),\n secret_access_key=test_credentials.secret_access_key(),\n bucket_name=test_credentials.bucket_name(),\n directory_name=test_credentials.directory_name(),\n cache_table='cache.s3_raw_downloads_cache',\n )\n\n def _expected_path_prefix(self) -> str:\n return 's3:'\n\n def test_key_value_store(self):\n self._test_key_value_store()\n","new_contents":"from mediawords.key_value_store.cached_amazon_s3 import CachedAmazonS3Store\nfrom mediawords.util.text import random_string\nfrom .amazon_s3_credentials import (\n TestAmazonS3CredentialsTestCase,\n get_test_s3_credentials,\n)\n\ntest_credentials = get_test_s3_credentials()\n\n\nclass TestCachedAmazonS3StoreTestCase(TestAmazonS3CredentialsTestCase):\n def _initialize_store(self) -> CachedAmazonS3Store:\n return CachedAmazonS3Store(\n access_key_id=test_credentials.access_key_id(),\n secret_access_key=test_credentials.secret_access_key(),\n bucket_name=test_credentials.bucket_name(),\n directory_name=test_credentials.directory_name() + '\/' + random_string(16),\n cache_table='cache.s3_raw_downloads_cache',\n )\n\n def _expected_path_prefix(self) -> str:\n return 's3:'\n\n def test_key_value_store(self):\n self._test_key_value_store()\n","subject":"Append random string to S3 test directory name to be able to run parallel tests","message":"Append random string to S3 test directory name to be able to run parallel tests\n","lang":"Python","license":"agpl-3.0","repos":"berkmancenter\/mediacloud,berkmancenter\/mediacloud,berkmancenter\/mediacloud,berkmancenter\/mediacloud,berkmancenter\/mediacloud"} {"commit":"94197717719b580aa9b8bf7a6cbe28f95000a2dc","old_file":"gcouchbase\/tests\/test_api.py","new_file":"gcouchbase\/tests\/test_api.py","old_contents":"from couchbase.tests.base import ApiImplementationMixin, SkipTest\ntry:\n import gevent\nexcept ImportError as e:\n raise SkipTest(e)\n\nfrom gcouchbase.bucket import Bucket, GView\nfrom couchbase.tests.importer import get_configured_classes\n\nclass GEventImplMixin(ApiImplementationMixin):\n factory = Bucket\n viewfactor = GView\n should_check_refcount = False\n\n\nskiplist = ('ConnectionIopsTest', 'LockmodeTest', 'ConnectionPipelineTest')\n\nconfigured_classes = get_configured_classes(GEventImplMixin,\n skiplist=skiplist)\nglobals().update(configured_classes)\n","new_contents":"from couchbase.tests.base import ApiImplementationMixin, SkipTest\ntry:\n import gevent\nexcept ImportError as e:\n raise SkipTest(e)\n\nfrom gcouchbase.bucket import Bucket, GView\nfrom couchbase.tests.importer import get_configured_classes\n\nclass GEventImplMixin(ApiImplementationMixin):\n factory = Bucket\n viewfactor = GView\n should_check_refcount = False\n\n\nskiplist = ('ConnectionIopsTest', 'LockmodeTest', 'ConnectionPipelineTest')\n\nconfigured_classes = get_configured_classes(GEventImplMixin,\n skiplist=skiplist)\n\n# View iterator test no longer works because of missing include_docs\ndef viter_skipfn(*args):\n raise SkipTest(\"include_docs not provided on client, \"\n \"and no longer supported by server\")\n\n\nfor n in ('test_include_docs', 'test_row_processor'):\n setattr(configured_classes['ViewIteratorTest_Bucket'], n, viter_skipfn)\n\nglobals().update(configured_classes)\n","subject":"Disable include_docs test for GCouchbase","message":"Disable include_docs test for GCouchbase\n\nThe gevent api cannot cleanly use include_docs here; previously we\nrelied on the server to retrieve this, but the server no longer supports\nthis. Perhaps in some future time we can implement this within\nlibcouchbase itself, but until then, it's left unimplemented.\n\nChange-Id: I1dc5cf86b4b063d251363f3d5fa14c8d29e8e6c1\nReviewed-on: http:\/\/review.couchbase.org\/44149\nTested-by: Mark Nunberg <e7d42768707bf23038325b0b68d4b577e5f6064a@haskalah.org>\nReviewed-by: Volker Mische <fb414f8ac0dbbf87663550ae4ef5fc95b1041941@gmail.com>\n","lang":"Python","license":"apache-2.0","repos":"couchbase\/couchbase-python-client,couchbase\/couchbase-python-client,mnunberg\/couchbase-python-client,mnunberg\/couchbase-python-client"} {"commit":"a33df3c34cf69102831c8878b71266aaa41ac8fb","old_file":"feed\/setup.py","new_file":"feed\/setup.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom setuptools import setup, find_packages\n\nlong_desc = '''\nThis package contains the feed Sphinx extension.\n\nIt creates an RSS feed of recently updated sphinx pages.\n'''\n\nrequires = ['Sphinx>=0.6', 'python-dateutil', 'beautifulsoup>=3.2.0', 'html5lib']\n\nsetup(\n name='feed',\n version='0.2',\n url='http:\/\/bitbucket.org\/birkenfeld\/sphinx-contrib',\n # download_url='http:\/\/pypi.python.org\/pypi\/feed',\n license='BSD',\n author='dan mackinlay',\n author_email='bitbucket@email.possumpalace.org',\n description='Sphinx extension feed',\n long_description=long_desc,\n zip_safe=False,\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Environment :: Console',\n 'Environment :: Web Environment',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Documentation',\n 'Topic :: Utilities',\n ],\n platforms='any',\n packages=find_packages(),\n include_package_data=True,\n install_requires=requires,\n namespace_packages=['sphinxcontrib'],\n)\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom setuptools import setup, find_packages\n\nlong_desc = '''\nThis package contains the feed Sphinx extension.\n\nIt creates an RSS feed of recently updated sphinx pages.\n'''\n\nrequires = ['Sphinx>=0.6', 'python-dateutil', 'html5lib']\n\nsetup(\n name='feed',\n version='0.2',\n url='http:\/\/bitbucket.org\/birkenfeld\/sphinx-contrib',\n # download_url='http:\/\/pypi.python.org\/pypi\/feed',\n license='BSD',\n author='dan mackinlay',\n author_email='bitbucket@email.possumpalace.org',\n description='Sphinx extension feed',\n long_description=long_desc,\n zip_safe=False,\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Environment :: Console',\n 'Environment :: Web Environment',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Documentation',\n 'Topic :: Utilities',\n ],\n platforms='any',\n packages=find_packages(),\n include_package_data=True,\n install_requires=requires,\n namespace_packages=['sphinxcontrib'],\n)\n","subject":"Remove beautifulsoup from the feed extension's requirements.","message":"Remove beautifulsoup from the feed extension's requirements.\n","lang":"Python","license":"bsd-2-clause","repos":"sphinx-contrib\/spelling,sphinx-contrib\/spelling"} {"commit":"53780b6d16d631a3c0e8859ff9771a1379de16f1","old_file":"calaccess_raw\/admin\/tracking.py","new_file":"calaccess_raw\/admin\/tracking.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nCustom administration panels for tracking models.\n\"\"\"\nfrom django.contrib import admin\nfrom calaccess_raw import models\nfrom .base import BaseAdmin\n\n\n@admin.register(models.RawDataVersion)\nclass RawDataVersionAdmin(BaseAdmin):\n \"\"\"\n Custom admin for the RawDataVersion model.\n \"\"\"\n list_display = (\n \"id\",\n \"release_datetime\",\n \"pretty_download_size\",\n \"download_file_count\",\n \"download_record_count\",\n \"clean_file_count\",\n \"clean_record_count\",\n \"pretty_clean_size\",\n \"download_file_count\",\n \"clean_record_count\"\n )\n list_display_links = ('release_datetime',)\n list_filter = (\"release_datetime\",)\n\n\n@admin.register(models.RawDataFile)\nclass RawDataFileAdmin(BaseAdmin):\n \"\"\"\n Custom admin for the RawDataFile model.\n \"\"\"\n list_display = (\n \"id\",\n \"version\",\n \"file_name\",\n \"download_records_count\",\n \"clean_records_count\",\n \"load_records_count\",\n \"error_count\"\n )\n list_display_links = ('id', 'file_name',)\n list_filter = (\"version__release_datetime\",)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nCustom administration panels for tracking models.\n\"\"\"\nfrom django.contrib import admin\nfrom calaccess_raw import models\nfrom .base import BaseAdmin\n\n\n@admin.register(models.RawDataVersion)\nclass RawDataVersionAdmin(BaseAdmin):\n \"\"\"\n Custom admin for the RawDataVersion model.\n \"\"\"\n list_display = (\n \"id\",\n \"release_datetime\",\n \"pretty_download_size\",\n \"download_file_count\",\n \"download_record_count\",\n \"clean_file_count\",\n \"clean_record_count\",\n \"pretty_clean_size\",\n )\n list_display_links = ('release_datetime',)\n list_filter = (\"release_datetime\",)\n\n\n@admin.register(models.RawDataFile)\nclass RawDataFileAdmin(BaseAdmin):\n \"\"\"\n Custom admin for the RawDataFile model.\n \"\"\"\n list_display = (\n \"id\",\n \"version\",\n \"file_name\",\n \"download_records_count\",\n \"clean_records_count\",\n \"load_records_count\",\n \"error_count\"\n )\n list_display_links = ('id', 'file_name',)\n list_filter = (\"version__release_datetime\",)\n","subject":"Cut dupe admin display fields","message":"Cut dupe admin display fields\n","lang":"Python","license":"mit","repos":"california-civic-data-coalition\/django-calaccess-raw-data"} {"commit":"7fc4e7382665cf9eac4d19efcf9641ad57271e87","old_file":"organizer\/models.py","new_file":"organizer\/models.py","old_contents":"from django.db import models\n\n\n# Model Field Reference\n# https:\/\/docs.djangoproject.com\/en\/1.8\/ref\/models\/fields\/\n\n\nclass Tag(models.Model):\n name = models.CharField(\n max_length=31, unique=True)\n slug = models.SlugField(\n max_length=31,\n unique=True,\n help_text='A label for URL config.')\n\n class Meta:\n ordering = ['name']\n\n def __str__(self):\n return self.name\n\n\nclass Startup(models.Model):\n name = models.CharField(\n max_length=31, db_index=True)\n slug = models.SlugField(\n max_length=31,\n unique=True,\n help_text='A label for URL config.')\n description = models.TextField()\n founded_date = models.DateField(\n 'date founded')\n contact = models.EmailField()\n website = models.URLField(max_length=255)\n tags = models.ManyToManyField(Tag)\n\n def __str__(self):\n return self.name\n\n\nclass NewsLink(models.Model):\n title = models.CharField(max_length=63)\n pub_date = models.DateField('date published')\n link = models.URLField(max_length=255)\n startup = models.ForeignKey(Startup)\n\n def __str__(self):\n return \"{}: {}\".format(\n self.startup, self.title)\n","new_contents":"from django.db import models\n\n\n# Model Field Reference\n# https:\/\/docs.djangoproject.com\/en\/1.8\/ref\/models\/fields\/\n\n\nclass Tag(models.Model):\n name = models.CharField(\n max_length=31, unique=True)\n slug = models.SlugField(\n max_length=31,\n unique=True,\n help_text='A label for URL config.')\n\n class Meta:\n ordering = ['name']\n\n def __str__(self):\n return self.name\n\n\nclass Startup(models.Model):\n name = models.CharField(\n max_length=31, db_index=True)\n slug = models.SlugField(\n max_length=31,\n unique=True,\n help_text='A label for URL config.')\n description = models.TextField()\n founded_date = models.DateField(\n 'date founded')\n contact = models.EmailField()\n website = models.URLField(max_length=255)\n tags = models.ManyToManyField(Tag)\n\n class Meta:\n ordering = ['name']\n get_latest_by = 'founded_date'\n\n def __str__(self):\n return self.name\n\n\nclass NewsLink(models.Model):\n title = models.CharField(max_length=63)\n pub_date = models.DateField('date published')\n link = models.URLField(max_length=255)\n startup = models.ForeignKey(Startup)\n\n def __str__(self):\n return \"{}: {}\".format(\n self.startup, self.title)\n","subject":"Declare Meta class in Startup model.","message":"Ch03: Declare Meta class in Startup model. [skip ci]\n","lang":"Python","license":"bsd-2-clause","repos":"jambonrose\/DjangoUnleashed-1.8,jambonrose\/DjangoUnleashed-1.8"} {"commit":"bcb4d14e7be413a08ca9c3f98656ff7b2bcb3d7d","old_file":"test\/test_wikilinks.py","new_file":"test\/test_wikilinks.py","old_contents":"from tiddlywebplugins.markdown import render\nfrom tiddlyweb.model.tiddler import Tiddler\n\n\nsample = \"\"\"# Hello\n \nThis is WikiLink\n\n* List\n* List\n\"\"\"\n\nsample_linked = \"\"\"\n\nThis is WikiLink and this is not: [NotLink](http:\/\/example.com).\n\nThis forthcoming in camel case but actually\na link [label](http:\/\/example.org\/CamelCase)\n\nThis is (HtmlJavascript in parens).\nThis is (parens around HtmlJavascript).\n\n\"\"\"\n\n\ndef test_no_wiki():\n tiddler = Tiddler('hello')\n tiddler.text = sample\n\n environ = {}\n output = render(tiddler, environ)\n assert '<h1>' in output\n assert '<li>' in output\n assert 'href' not in output\n\n environ = {'tiddlyweb.config': {'markdown.wiki_link_base': ''}}\n output = render(tiddler, environ)\n assert 'href' in output\n assert '<a href=\"WikiLink\">' in output\n assert '>WikiLink<\/a>' in output\n\n tiddler.text = sample_linked\n output = render(tiddler, environ)\n assert '\"NotLink\"' not in output\n assert '<a href=\"http:\/\/example.org\/CamelCase\">label<\/a>' in output\n\n print output\n","new_contents":"from tiddlywebplugins.markdown import render\nfrom tiddlyweb.model.tiddler import Tiddler\n\n\nsample = \"\"\"# Hello\n \nThis is WikiLink\n\n* List\n* List\n\"\"\"\n\nsample_linked = \"\"\"\n\nThis is WikiLink and this is not: [NotLink](http:\/\/example.com).\n\nThis forthcoming in camel case but actually\na link [label](http:\/\/example.org\/CamelCase)\n\nThis is (HtmlJavascript in parens).\nThis is (parens around HtmlJavascript).\n\n\"\"\"\n\n\ndef test_no_wiki():\n tiddler = Tiddler('hello')\n tiddler.text = sample\n\n environ = {}\n output = render(tiddler, environ)\n assert '<h1>' in output\n assert '<li>' in output\n assert 'href' not in output\n\n environ = {'tiddlyweb.config': {'markdown.wiki_link_base': ''}}\n output = render(tiddler, environ)\n assert 'href' in output\n assert '<a href=\"WikiLink\">' in output\n assert '>WikiLink<\/a>' in output\n\n tiddler.text = sample_linked\n output = render(tiddler, environ)\n assert '\"NotLink\"' not in output\n assert '<a href=\"http:\/\/example.org\/CamelCase\">label<\/a>' in output\n\n assert '(<a href=\"HtmlJavascript\">HtmlJavascript<\/a> in parens)' in output\n assert '(parens around <a href=\"HtmlJavascript\">HtmlJavascript<\/a>)' in output\n\n","subject":"Write forgotten test replace debugging output.","message":"Write forgotten test replace debugging output.\n","lang":"Python","license":"bsd-2-clause","repos":"tiddlyweb\/tiddlywebplugins.markdown"} {"commit":"8d50846847852741410463d98de2c4f9e5fea844","old_file":"zaqar_ui\/content\/queues\/urls.py","new_file":"zaqar_ui\/content\/queues\/urls.py","old_contents":"# Copyright 2015 IBM Corp.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom django.conf import urls\n\nfrom zaqar_ui.content.queues import views\n\n\nurlpatterns = urls.patterns(\n 'zaqar_ui.content.queues',\n urls.url(r'^$', views.IndexView.as_view(), name='index'),\n)\n","new_contents":"# Copyright 2015 IBM Corp.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom django.conf import urls\n\nfrom zaqar_ui.content.queues import views\n\n\nurlpatterns = [\n urls.url(r'^$', views.IndexView.as_view(), name='index'),\n]\n","subject":"Update URLs to Django 1.8 style","message":"Update URLs to Django 1.8 style\n\ndjango.conf.urls.patterns() is deprecated since 1.8.\nWe should not use patterns(), so this patch updates URLs to\n1.8 style.\n\nChange-Id: I6f2b6f44d843ca5e0cdb5db9828df94fa4df5f88\nCloses-Bug: #1539354\n","lang":"Python","license":"apache-2.0","repos":"openstack\/zaqar-ui,openstack\/zaqar-ui,openstack\/zaqar-ui,openstack\/zaqar-ui"} {"commit":"1d2a62b87b98513bd84a0ae1433781157cb45f70","old_file":"admin.py","new_file":"admin.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom django.contrib import admin\n\nfrom djurk.models import HIT\n\nclass HIT_Admin(admin.ModelAdmin):\n list_display = (\n 'creation_time',\n 'hit_id',\n 'hit_type_id',\n 'title',\n 'reward'\n )\n list_filter = (\n 'creation_time',\n 'hit_status',\n 'hit_review_status',\n )\n search_fields = (\n 'hit_id',\n 'hit_type_id',\n 'title',\n 'description',\n 'keyword',\n )\n\nadmin.site.register(HIT, HIT_Admin)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom django.contrib import admin\n\nfrom djurk.models import HIT\n\nclass HIT_Admin(admin.ModelAdmin):\n date_hierarchy = 'creation_time'\n fieldsets = (\n (None, {\n 'fields': (('hit_id','hit_type_id'),\n ('creation_time', 'hit_status'),\n ('title', 'keywords', 'description'),\n 'reward',\n 'requester_annotation',\n ),\n }),\n ('HIT Details', {\n 'classes': ('collapse',),\n 'fields': (\n 'lifetime_in_seconds',\n 'auto_approval_delay_in_seconds',\n 'number_of_similar_hits',\n 'hit_review_status',\n )\n }),\n ('Assignment Overview', {\n 'classes': ('collapse',),\n 'fields': (\n 'max_assignments',\n 'assignment_duration_in_seconds',\n 'number_of_assignments_pending',\n 'number_of_assignments_available',\n 'number_of_assignments_completed',\n )\n }),\n )\n list_display = (\n 'creation_time',\n 'hit_id',\n 'hit_type_id',\n 'title',\n 'reward'\n )\n readonly_fields = (\n 'creation_time',\n 'hit_id',\n 'hit_type_id',\n 'hit_status',\n 'hit_review_status',\n )\n list_display_links = list_display\n list_filter = (\n 'hit_status',\n 'hit_review_status',\n 'creation_time',\n )\n read_only_fields = ('creation_time',)\n search_fields = (\n 'hit_id',\n 'hit_type_id',\n 'title',\n 'description',\n 'keyword',\n )\n\nadmin.site.register(HIT, HIT_Admin)\n","subject":"Customize HIT Admin to \"fit your brain\"","message":"Customize HIT Admin to \"fit your brain\"\n","lang":"Python","license":"bsd-3-clause","repos":"glenjarvis\/djurk"} {"commit":"053bfa79b54a95d405d6401c86dcce3c6065bf32","old_file":"appium\/conftest.py","new_file":"appium\/conftest.py","old_contents":"import os\nimport pytest\n\nfrom appium import webdriver\n\n\n@pytest.fixture(scope='function')\ndef driver(request):\n return get_driver(request, default_capabilities())\n\n@pytest.fixture(scope='function')\ndef no_reset_driver(request):\n desired_caps = default_capabilities()\n desired_caps['noReset'] = True\n return get_driver(request, desired_caps)\n\ndef get_driver(request, desired_caps):\n _driver = webdriver.Remote(\n command_executor='http:\/\/127.0.0.1:4723\/wd\/hub',\n desired_capabilities=desired_caps)\n\n request.addfinalizer(_driver.quit)\n\n return _driver\n\ndef default_capabilities():\n app = os.path.abspath('aws\/Build\/Products\/Release-iphonesimulator\/AwesomeProject.app')\n\n screenshot_folder = os.getenv('SCREENSHOT_PATH', '')\n\n desired_caps = {}\n\n if screenshot_folder == '':\n desired_caps['platformName'] = 'iOS'\n desired_caps['platformVersion'] = '10.3'\n desired_caps['deviceName'] = 'iPhone Simulator'\n desired_caps['app'] = app\n\n return desired_caps\n","new_contents":"import os\nimport pytest\n\nfrom appium import webdriver\n\n\n@pytest.fixture(scope='function')\ndef driver(request):\n return get_driver(request, default_capabilities())\n\n@pytest.fixture(scope='function')\ndef no_reset_driver(request):\n desired_caps = default_capabilities()\n desired_caps['noReset'] = (runs_on_aws() == True and False or True)\n return get_driver(request, desired_caps)\n\ndef get_driver(request, desired_caps):\n _driver = webdriver.Remote(\n command_executor='http:\/\/127.0.0.1:4723\/wd\/hub',\n desired_capabilities=desired_caps)\n\n request.addfinalizer(_driver.quit)\n\n return _driver\n\ndef runs_on_aws():\n return os.getenv('SCREENSHOT_PATH', False) != False\n\ndef default_capabilities():\n app = os.path.abspath('aws\/Build\/Products\/Release-iphonesimulator\/AwesomeProject.app')\n\n desired_caps = {}\n\n if runs_on_aws() == False:\n desired_caps['platformName'] = 'iOS'\n desired_caps['platformVersion'] = '10.3'\n desired_caps['deviceName'] = 'iPhone Simulator'\n desired_caps['app'] = app\n\n return desired_caps\n","subject":"Remove no reset for devicefarm","message":"Remove no reset for devicefarm\n","lang":"Python","license":"mit","repos":"getsentry\/react-native-sentry,getsentry\/react-native-sentry,getsentry\/react-native-sentry,getsentry\/react-native-sentry,getsentry\/react-native-sentry,getsentry\/react-native-sentry"} {"commit":"8b8e206c21d08fee74fd43dc4b7e4d1d95a93060","old_file":"sconsole\/cmdbar.py","new_file":"sconsole\/cmdbar.py","old_contents":"'''\nDefine the command bar\n'''\n# Import third party libs\nimport urwid\n\n\nclass CommandBar(object):\n '''\n The object to manage the command bar\n '''\n def __init__(self, opts):\n self.opts = opts\n self.tgt_txt = urwid.Text('Target')\n self.tgt_edit = urwid.Edit()\n self.fun_txt = urwid.Text('Function')\n self.fun_edit = urwid.Edit()\n self.arg_txt = urwid.Text('Arguments')\n self.arg_edit = urwid.Edit()\n self.go_button = urwid.Button('GO!')\n self.grid = urwid.GridFlow(\n [self.tgt_txt,\n self.tgt_edit,\n self.fun_txt,\n self.fun_edit,\n self.arg_txt,\n self.arg_edit,\n self.go_button],\n cell_width=10,\n h_sep=1,\n v_sep=1,\n align='left')\n","new_contents":"'''\nDefine the command bar\n'''\n# Import third party libs\nimport urwid\n\n# Import salt libs\nimport salt.client\n\n\nclass CommandBar(object):\n '''\n The object to manage the command bar\n '''\n def __init__(self, opts):\n self.opts = opts\n self.local = salt.client.LocalClient(mopts=opts)\n self.tgt_txt = urwid.Text('Target')\n self.tgt_edit = urwid.Edit()\n self.fun_txt = urwid.Text('Function')\n self.fun_edit = urwid.Edit()\n self.arg_txt = urwid.Text('Arguments')\n self.arg_edit = urwid.Edit()\n self.go_button = urwid.Button('GO!', self.run_command)\n self.grid = urwid.GridFlow(\n [self.tgt_txt,\n self.tgt_edit,\n self.fun_txt,\n self.fun_edit,\n self.arg_txt,\n self.arg_edit,\n self.go_button],\n cell_width=10,\n h_sep=1,\n v_sep=1,\n align='left')\n\n def run_command(self, button, user_data):\n '''\n Execute the corresponding salt command\n '''\n tgt = self.tgt_edit.get_edit_text()\n fun = self.fun_edit.get_edit_text()\n args = self.arg_edit.get_edit_text().split()\n self.local.cmd_async(tgt, fun, args)\n","subject":"Add functionality to the go button","message":"Add functionality to the go button\n","lang":"Python","license":"apache-2.0","repos":"saltstack\/salt-console"} {"commit":"f4170ec0cff71e8bcce834bbf8f4336410d45e76","old_file":"klab\/cluster\/__init__.py","new_file":"klab\/cluster\/__init__.py","old_contents":"#!\/usr\/bin\/env python2\n\ndef is_this_chef():\n from socket import gethostname\n return gethostname() == 'chef.compbio.ucsf.edu'\n\ndef require_chef():\n if not is_this_chef():\n raise SystemExit(\"This script must be run on chef.\")\n\ndef require_qsub():\n import os, subprocess\n\n try:\n command = 'qsub', '-help'\n devnull = open(os.devnull)\n subprocess.Popen(command, stdout=devnull, stderr=devnull).communicate()\n except OSError as e:\n if e.errno == os.errno.ENOENT:\n return False\n\n return True\n\n \n","new_contents":"#!\/usr\/bin\/env python2\n\ndef is_this_chef():\n from socket import gethostname\n return gethostname() == 'chef.compbio.ucsf.edu'\n\ndef require_chef():\n if not is_this_chef():\n raise SystemExit(\"This script must be run on chef.\")\n\ndef require_qsub():\n import os, subprocess\n\n try:\n command = 'qsub', '-help'\n devnull = open(os.devnull)\n subprocess.Popen(command, stdout=devnull, stderr=devnull).communicate()\n except OSError as e:\n if e.errno == os.errno.ENOENT:\n print \"'qsub' not found. Are you logged onto the cluster?\"\n raise SystemExit\n\n \n","subject":"Make require_qsub() crash if qsub isn't found.","message":"Make require_qsub() crash if qsub isn't found.\n","lang":"Python","license":"mit","repos":"Kortemme-Lab\/klab,Kortemme-Lab\/klab,Kortemme-Lab\/klab,Kortemme-Lab\/klab"} {"commit":"65fb294c17130985e6549702a1df728126de8cb0","old_file":"addon.py","new_file":"addon.py","old_contents":"import sys, os\nsys.path.append(os.path.join(os.path.dirname(__file__), 'resources', 'site-packages'))\nfrom xbmctorrent import plugin\n\n\nif __name__ == '__main__':\n plugin.run()\n","new_contents":"import sys, os\nsys.path.insert(0, os.path.join(os.path.dirname(__file__), 'resources', 'site-packages'))\nfrom xbmctorrent import plugin\n\n\nif __name__ == '__main__':\n plugin.run()\n","subject":"Make sure we use our own vendored packages","message":"Make sure we use our own vendored packages\n","lang":"Python","license":"apache-2.0","repos":"neno1978\/xbmctorrent"} {"commit":"0cc89fe31729a485a0e055b343acfde3d71745d7","old_file":"apps\/metricsmanager\/api.py","new_file":"apps\/metricsmanager\/api.py","old_contents":"from rest_framework.views import APIView\nfrom rest_framework.reverse import reverse\nfrom rest_framework.response import Response\nfrom rest_framework import generics, status\nfrom django.core.exceptions import ValidationError\nfrom .models import *\nfrom .serializers import *\nfrom .formula import validate_formula\n\nclass MetricsBase(APIView):\n\n def get(self, request, format=None):\n \"\"\"\n :type request: Request\n :param request:\n :return:\n \"\"\"\n result = {\n \"Metrics\": reverse('metrics-create', request=request)\n }\n\n return Response(result)\n\nclass FormulaValidate(APIView):\n\n def get(self, request):\n if \"formula\" not in request.QUERY_PARAMS:\n return Response(\"No formula provided\")\n try:\n validate_formula(request.QUERY_PARAMS[\"formula\"])\n return Response(status=status.HTTP_204_NO_CONTENT)\n except ValidationError as e:\n return Response({ \"formula\": e.message }, status=status.HTTP_400_BAD_REQUEST)\n\nclass MetricsCreate(generics.CreateAPIView):\n model = Metric\n serializer_class = MetricSerializer\n\nclass MetricsDetail(generics.RetrieveAPIView):\n model = Metric\n serializer_class = MetricSerializer\n","new_contents":"from rest_framework.views import APIView\nfrom rest_framework.reverse import reverse\nfrom rest_framework.response import Response\nfrom rest_framework import generics, status\nfrom django.core.exceptions import ValidationError\nfrom .models import *\nfrom .serializers import *\nfrom .formula import validate_formula\n\nclass MetricsBase(APIView):\n\n def get(self, request, format=None):\n \"\"\"\n :type request: Request\n :param request:\n :return:\n \"\"\"\n result = {\n \"Metrics\": reverse('metrics-create', request=request)\n }\n\n return Response(result)\n\nclass FormulaValidate(APIView):\n\n def get(self, request):\n if \"formula\" not in request.QUERY_PARAMS:\n return Response({ \"formula\": \"Can not be empty\"}, status=status.HTTP_400_BAD_REQUEST)\n try:\n validate_formula(request.QUERY_PARAMS[\"formula\"])\n return Response(status=status.HTTP_204_NO_CONTENT)\n except ValidationError as e:\n return Response({ \"formula\": e.message }, status=status.HTTP_400_BAD_REQUEST)\n\nclass MetricsCreate(generics.CreateAPIView):\n model = Metric\n serializer_class = MetricSerializer\n\nclass MetricsDetail(generics.RetrieveAPIView):\n model = Metric\n serializer_class = MetricSerializer\n","subject":"Fix error format of check formula endpoint","message":"Fix error format of check formula endpoint\n","lang":"Python","license":"agpl-3.0","repos":"mmilaprat\/policycompass-services,almey\/policycompass-services,mmilaprat\/policycompass-services,policycompass\/policycompass-services,almey\/policycompass-services,policycompass\/policycompass-services,almey\/policycompass-services,mmilaprat\/policycompass-services,policycompass\/policycompass-services"} {"commit":"a8601d8a17c9ba8e87b8336870e0d52f79e0ffa2","old_file":"indra\/tests\/test_omnipath.py","new_file":"indra\/tests\/test_omnipath.py","old_contents":"from __future__ import unicode_literals\nfrom builtins import dict, str\nfrom indra.statements import Phosphorylation\nfrom indra.databases import omnipath as op\n\ndef test_query_ptms():\n stmts = op.get_ptms(['Q13873'])\n assert len(stmts) == 1\n assert isinstance(stmts[0], Phosphorylation)\n assert stmts[0].enz.name == 'CSNK2A1'\n assert stmts[0].sub.name == 'BMPR2'\n assert stmts[0].residue == 'S'\n assert stmts[0].position == '757'\n","new_contents":"import requests\nfrom indra.sources.omnipath import OmniPathModificationProcessor,\\\n OmniPathLiganReceptorProcessor\nfrom indra.sources.omnipath.api import op_url\nfrom indra.statements import Agent, Phosphorylation\nfrom indra.preassembler.grounding_mapper import GroundingMapper\n\nBRAF_UPID = 'P15056'\nJAK2_UPID = 'O60674'\nBRAF_AG = Agent(None, db_refs={'UP': BRAF_UPID})\nGroundingMapper.standardize_agent_name(BRAF_AG)\nJAK2_AG = Agent(None, db_refs={'UP': JAK2_UPID})\nGroundingMapper.standardize_agent_name(JAK2_AG)\n\n\ndef test_omnipath_web_api():\n query_url = '%s\/queries'\n res = requests.get(query_url)\n assert res.status_code == 200\n\ndef test_query_ptms():\n stmts = op.get_ptms(['Q13873'])\n assert len(stmts) == 1\n assert isinstance(stmts[0], Phosphorylation)\n assert stmts[0].enz.name == 'CSNK2A1'\n assert stmts[0].sub.name == 'BMPR2'\n assert stmts[0].residue == 'S'\n assert stmts[0].position == '757'\n","subject":"Update imports, test general web api","message":"Update imports, test general web api\n","lang":"Python","license":"bsd-2-clause","repos":"johnbachman\/indra,johnbachman\/indra,johnbachman\/belpy,johnbachman\/indra,johnbachman\/belpy,sorgerlab\/belpy,sorgerlab\/indra,sorgerlab\/indra,bgyori\/indra,johnbachman\/belpy,bgyori\/indra,sorgerlab\/belpy,bgyori\/indra,sorgerlab\/indra,sorgerlab\/belpy"} {"commit":"2083c0079a70783deff54a7acd6f3ef6bba25302","old_file":"tests\/test_pyglmnet.py","new_file":"tests\/test_pyglmnet.py","old_contents":"import numpy as np\nimport scipy.sparse as sps\nfrom sklearn.preprocessing import StandardScaler\n\nfrom numpy.testing import assert_allclose\n\nfrom pyglmnet import GLM\n\n\ndef test_glmnet():\n \"\"\"Test glmnet.\"\"\"\n glm = GLM(distr='poisson')\n scaler = StandardScaler()\n n_samples, n_features = 10000, 100\n density = 0.1\n\n # coefficients\n beta0 = np.random.rand()\n beta = sps.rand(n_features, 1, density=density).toarray()\n\n X_train = np.random.normal(0.0, 1.0, [n_samples, n_features])\n y_train = glm.simulate(beta0, beta, X_train)\n\n X_train = scaler.fit_transform(X_train)\n glm.fit(X_train, y_train)\n\n beta_ = glm.fit_params[-2]['beta'][:]\n assert_allclose(beta[:], beta_, atol=0.1) # check fit\n density_ = np.sum(beta_ > 0.1) \/ float(n_features)\n assert_allclose(density_, density, atol=0.05) # check density\n","new_contents":"import numpy as np\nimport scipy.sparse as sps\nfrom sklearn.preprocessing import StandardScaler\n\nfrom numpy.testing import assert_allclose\n\nfrom pyglmnet import GLM\n\n\ndef test_glmnet():\n \"\"\"Test glmnet.\"\"\"\n glm = GLM(distr='poisson')\n scaler = StandardScaler()\n n_samples, n_features = 10000, 100\n density = 0.1\n\n # coefficients\n beta0 = np.random.rand()\n beta = sps.rand(n_features, 1, density=density).toarray()\n\n X_train = np.random.normal(0.0, 1.0, [n_samples, n_features])\n y_train = glm.simulate(beta0, beta, X_train)\n\n X_train = scaler.fit_transform(X_train)\n glm.fit(X_train, y_train)\n\n beta_ = glm.fit_[-2]['beta'][:]\n assert_allclose(beta[:], beta_, atol=0.1) # check fit\n density_ = np.sum(beta_ > 0.1) \/ float(n_features)\n assert_allclose(density_, density, atol=0.05) # check density\n\ndef test_multinomial_gradient():\n \"\"\"Gradient of intercept params is different\"\"\"\n glm = GLM(distr='multinomial')\n X = np.array([[1,2,3], [4,5,6]])\n y = np.array([1,2])\n beta = np.zeros([4, 2])\n grad_beta0, grad_beta = glm.grad_L2loss(beta[0], beta[1:], 0, X, y)\n assert grad_beta0[0] != grad_beta0[1]\n","subject":"Fix glmnet test and add multinomial gradient test","message":"Fix glmnet test and add multinomial gradient test\n","lang":"Python","license":"mit","repos":"the872\/pyglmnet,glm-tools\/pyglmnet,pavanramkumar\/pyglmnet"} {"commit":"2296ef02345f51666ff6653abe372e7965ef361c","old_file":"categories_i18n\/admin.py","new_file":"categories_i18n\/admin.py","old_contents":"from django.contrib import admin\nfrom mptt.admin import MPTTModelAdmin\nfrom mptt.forms import MPTTAdminForm\nfrom parler.admin import TranslatableAdmin\nfrom .models import Category\nfrom parler.forms import TranslatableModelForm\n\n\nclass CategoryAdminForm(MPTTAdminForm, TranslatableModelForm):\n \"\"\"\n Form for categories, both MPTT + translatable.\n \"\"\"\n pass\n\n\nclass CategoryAdmin(MPTTModelAdmin, TranslatableAdmin):\n \"\"\"\n Admin page for categories.\n \"\"\"\n list_display = ('title', 'slug')\n search_fields = ('translations__title', 'translations__slug')\n form = CategoryAdminForm\n\n fieldsets = (\n (None, {\n 'fields': ('title', 'slug', 'parent'),\n }),\n )\n\n def get_prepopulated_fields(self, request, obj=None):\n # Needed for django-parler\n return {'slug': ('title',)}\n\n\nadmin.site.register(Category, CategoryAdmin)\n","new_contents":"from django.contrib import admin\nfrom mptt.admin import MPTTModelAdmin\nfrom mptt.forms import MPTTAdminForm\nfrom parler.admin import TranslatableAdmin\nfrom .models import Category\nfrom parler.forms import TranslatableModelForm\n\n\nclass CategoryAdminForm(MPTTAdminForm, TranslatableModelForm):\n \"\"\"\n Form for categories, both MPTT + translatable.\n \"\"\"\n pass\n\n\nclass CategoryAdmin(MPTTModelAdmin, TranslatableAdmin):\n \"\"\"\n Admin page for categories.\n \"\"\"\n list_display = ('title', 'slug')\n mptt_indent_field = 'title' # be explicit for MPTT\n search_fields = ('translations__title', 'translations__slug')\n form = CategoryAdminForm\n\n fieldsets = (\n (None, {\n 'fields': ('title', 'slug', 'parent'),\n }),\n )\n\n def get_prepopulated_fields(self, request, obj=None):\n # Needed for django-parler\n return {'slug': ('title',)}\n\n\nadmin.site.register(Category, CategoryAdmin)\n","subject":"Set `mptt_indent_field` explicitly for proper MPTT list columns","message":"Set `mptt_indent_field` explicitly for proper MPTT list columns\n","lang":"Python","license":"apache-2.0","repos":"edoburu\/django-categories-i18n,edoburu\/django-categories-i18n"} {"commit":"70d3bf1043af965da56339ce1c50b258f184ccb2","old_file":"lib\/bio.py","new_file":"lib\/bio.py","old_contents":"\"\"\"Utilities for working with sequences.\"\"\"\n\nimport re\nfrom Bio import SeqIO\n\nCODON_LEN = 3\n\nCOMPLEMENT = str.maketrans('ACGTUWSMKRYBDHVNXacgtuwsmkrybdhvnx',\n 'TGCAAWSKMYRVHDBNXtgcaawskmyrvhdbnx')\n\nIS_PROTEIN = re.compile(r'[EFILPQ]', re.IGNORECASE)\n\n\ndef reverse_complement(seq):\n \"\"\"Reverse complement a nucleotide sequence. We added some wildcards.\"\"\"\n return seq.translate(COMPLEMENT)[::-1]\n\n\ndef is_protein(seq):\n \"\"\"Check if the sequence a protein.\"\"\"\n return IS_PROTEIN.search(seq)\n\n\ndef fasta_file_has_protein(query_files):\n \"\"\"\n Search for protein characters in a fasta file.\n\n If the user has told us that we have a protein then return that.\n \"\"\"\n for query_file in query_files:\n with open(query_file) as in_file:\n for query in SeqIO.parse(in_file, 'fasta'):\n if is_protein(str(query.seq)):\n return True\n\n return False\n","new_contents":"\"\"\"Utilities for working with sequences.\"\"\"\n\nimport re\nfrom Bio import SeqIO\n\nCODON_LEN = 3\n\nCOMPLEMENT = str.maketrans('ACGTUWSMKRYBDHVNXacgtuwsmkrybdhvnx-',\n 'TGCAAWSKMYRVHDBNXtgcaawskmyrvhdbnx-')\n\nIS_PROTEIN = re.compile(r'[EFILPQ]', re.IGNORECASE)\n\n\ndef reverse_complement(seq):\n \"\"\"Reverse complement a nucleotide sequence. We added some wildcards.\"\"\"\n return seq.translate(COMPLEMENT)[::-1]\n\n\ndef is_protein(seq):\n \"\"\"Check if the sequence a protein.\"\"\"\n return IS_PROTEIN.search(seq)\n\n\ndef fasta_file_has_protein(query_files):\n \"\"\"\n Search for protein characters in a fasta file.\n\n If the user has told us that we have a protein then return that.\n \"\"\"\n for query_file in query_files:\n with open(query_file) as in_file:\n for query in SeqIO.parse(in_file, 'fasta'):\n if is_protein(str(query.seq)):\n return True\n\n return False\n","subject":"Add gaps \"-\" to reverse complement function","message":"Add gaps \"-\" to reverse complement function\n","lang":"Python","license":"bsd-3-clause","repos":"juliema\/aTRAM"} {"commit":"c0dc0c644fd8912d58deb416955e85259d22618e","old_file":"tests\/github_controller\/test_request_parsing.py","new_file":"tests\/github_controller\/test_request_parsing.py","old_contents":"import pytest\n\nfrom app.controllers.github_controller import GithubController\n\npytestmark = pytest.mark.asyncio\n\n\nasync def test_get_req_json(gh_sut: GithubController, mock_request):\n assert await gh_sut.get_request_json(mock_request) == 'json'\n\n\nasync def test_get_req_event_header(gh_sut: GithubController, mock_request):\n assert await gh_sut.get_request_event_header(mock_request) == 'event'\n","new_contents":"import pytest\n\nfrom app.controllers.github_controller import GithubController\n\npytestmark = pytest.mark.asyncio\n\n\nasync def test_get_req_json(gh_sut: GithubController, mock_request):\n assert await gh_sut.get_request_json(mock_request) == {'json': 'json'}\n\n\nasync def test_get_req_event_header(gh_sut: GithubController, mock_request):\n assert await gh_sut.get_request_event_header(mock_request) == 'event'\n","subject":"Fix test in request parsing","message":"Fix test in request parsing\n","lang":"Python","license":"mit","repos":"futuresimple\/triggear"} {"commit":"8982d3f5ea40b688ec7e1da18403d89ab2994a95","old_file":"comics\/comics\/yamac.py","new_file":"comics\/comics\/yamac.py","old_contents":"from comics.aggregator.crawler import CrawlerBase\nfrom comics.core.comic_data import ComicDataBase\n\n\nclass ComicData(ComicDataBase):\n name = \"you and me and cats\"\n language = \"en\"\n url = \"http:\/\/strawberry-pie.net\/SA\/\"\n start_date = \"2009-07-01\"\n rights = \"bubble\"\n active = False\n\n\nclass Crawler(CrawlerBase):\n history_capable_days = 365\n time_zone = \"US\/Pacific\"\n\n def crawl(self, pub_date):\n pass\n","new_contents":"from comics.aggregator.crawler import CrawlerBase\nfrom comics.core.comic_data import ComicDataBase\n\n\nclass ComicData(ComicDataBase):\n name = \"you and me and cats\"\n language = \"en\"\n url = \"http:\/\/strawberry-pie.net\/SA\/\"\n start_date = \"2009-07-01\"\n rights = \"bubble\"\n active = False\n\n\nclass Crawler(CrawlerBase):\n time_zone = \"US\/Pacific\"\n\n def crawl(self, pub_date):\n pass\n","subject":"Remove history capable date for \"you and me and cats\"","message":"Remove history capable date for \"you and me and cats\"\n","lang":"Python","license":"agpl-3.0","repos":"jodal\/comics,jodal\/comics,datagutten\/comics,datagutten\/comics,jodal\/comics,jodal\/comics,datagutten\/comics,datagutten\/comics"} {"commit":"26861b183085e8fe2c7c21f4e3631ddd7d30e5e8","old_file":"csibe.py","new_file":"csibe.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport subprocess\nimport unittest\n\ncsibe_path = os.path.dirname(os.path.realpath(__file__))\nbuild_directory = \"build\"\n\nif not os.path.isdir(build_directory):\n os.makedirs(build_directory)\n\nos.chdir(build_directory)\n\nsubprocess.call([\"cmake\", csibe_path])\n\n","new_contents":"#!\/usr\/bin\/env python\n\nimport argparse\nimport os\nimport subprocess\nimport sys\n\nparser = argparse.ArgumentParser()\nparser.add_argument(\"-j\", \"--jobs\", type=int, default=1, help=\"number of jobs for make\")\nargs = parser.parse_args()\n\nmake_jobs = args.jobs\n\ncsibe_path = os.path.dirname(os.path.realpath(__file__))\nbuild_directory = \"build\"\n\nif not os.path.isdir(build_directory):\n os.makedirs(build_directory)\n\nos.chdir(build_directory)\n\ncmake_return_value = subprocess.call([\"cmake\", csibe_path])\nif cmake_return_value:\n sys.exit(cmake_return_value)\n\nmake_return_value = subprocess.call([\"make\", \"-j{}\".format(make_jobs)])\nif make_return_value:\n sys.exit(make_return_value)\n\nmake_size_return_value = subprocess.call([\"make\", \"size\"])\nif make_size_return_value:\n sys.exit(make_size_return_value)\n\n","subject":"Add logic and error-handling for CMake and make invocations","message":"Add logic and error-handling for CMake and make invocations\n","lang":"Python","license":"bsd-3-clause","repos":"szeged\/csibe,bgabor666\/csibe,szeged\/csibe,szeged\/csibe,bgabor666\/csibe,bgabor666\/csibe,bgabor666\/csibe,loki04\/csibe,loki04\/csibe,loki04\/csibe,szeged\/csibe,bgabor666\/csibe,bgabor666\/csibe,bgabor666\/csibe,loki04\/csibe,loki04\/csibe,szeged\/csibe,szeged\/csibe,loki04\/csibe,loki04\/csibe,szeged\/csibe"} {"commit":"f87bab8a808e4bda3b3b7482633eaca069682b9e","old_file":"build.py","new_file":"build.py","old_contents":"# -*- coding: utf-8 -*-\nimport sys\nfrom cx_Freeze import setup, Executable\n\nbase = None\nif sys.platform == \"win32\":\n base = \"Win32GUI\"\n\nexecutables = [\n Executable('blockcheck.py', base=base)\n]\n\nsetup(name='blockcheck',\n version='0.1',\n description='BlockCheck',\n executables=executables,\n options = {'build_exe': {'init_script':'Console', 'compressed':'1'}},\n )","new_contents":"# -*- coding: utf-8 -*-\nimport sys\nfrom cx_Freeze import setup, Executable\n\nbase = None\nif sys.platform == \"win32\":\n base = \"Win32GUI\"\n\nexecutables = [\n Executable('blockcheck.py', base=base)\n]\n\nsetup(name='blockcheck',\n version='0.0.5',\n description='BlockCheck',\n executables=executables,\n options = {'build_exe': {'init_script':'Console', 'compressed':'1', 'packages':'dns'}},\n )\n","subject":"Include all the files from dns module and bump version","message":"Include all the files from dns module and bump version","lang":"Python","license":"mit","repos":"Acharvak\/blockcheck,Renji\/blockcheck,ValdikSS\/blockcheck"} {"commit":"89c005e2fd7d7f7727ba225cc20789fea992b1d4","old_file":"backend\/scripts\/mktemplate.py","new_file":"backend\/scripts\/mktemplate.py","old_contents":"#!\/usr\/bin\/env python\n\nimport json\nimport rethinkdb as r\nimport sys\nimport optparse\n\nif __name__ == \"__main__\":\n parser = optparse.OptionParser()\n parser.add_option(\"-p\", \"--port\", dest=\"port\",\n help=\"rethinkdb port\", default=30815)\n parser.add_option(\"-f\", \"--file\", dest=\"filename\",\n help=\"json file\", type=\"string\")\n (options, args) = parser.parse_args()\n if options.filename is None:\n print \"You must specify json file\"\n sys.exit(1)\n conn = r.connect('localhost', int(options.port), db='materialscommons')\n json_data = open(options.filename)\n data = json.load(json_data)\n existing = r.table('templates').get(data['id']).run(conn)\n if existing:\n r.table('templates').get(data['id']).delete().run(conn)\n r.table('templates').insert(data).run(conn)\n print 'template deleted and re-inserted into the database'\n else:\n r.table('templates').insert(data).run(conn)\n print 'template inserted into the database'\n","new_contents":"#!\/usr\/bin\/env python\n\nimport json\nimport rethinkdb as r\nimport sys\nimport optparse\n\nif __name__ == \"__main__\":\n parser = optparse.OptionParser()\n parser.add_option(\"-p\", \"--port\", dest=\"port\",\n help=\"rethinkdb port\", default=30815)\n parser.add_option(\"-f\", \"--file\", dest=\"filename\",\n help=\"json file\", type=\"string\")\n (options, args) = parser.parse_args()\n if options.filename is None:\n print \"You must specify json file\"\n sys.exit(1)\n conn = r.connect('localhost', int(options.port), db='materialscommons')\n json_data = open(options.filename)\n print \"Loading template file: %s\" % (options.filename)\n data = json.load(json_data)\n existing = r.table('templates').get(data['id']).run(conn)\n if existing:\n r.table('templates').get(data['id']).delete().run(conn)\n r.table('templates').insert(data).run(conn)\n print 'template deleted and re-inserted into the database'\n else:\n r.table('templates').insert(data).run(conn)\n print 'template inserted into the database'\n","subject":"Update script to show which file it is loading.","message":"Update script to show which file it is loading.\n","lang":"Python","license":"mit","repos":"materials-commons\/materialscommons.org,materials-commons\/materialscommons.org,materials-commons\/materialscommons.org,materials-commons\/materialscommons.org,materials-commons\/materialscommons.org"} {"commit":"6ac4764790526f435ffc6337145439d710dd455f","old_file":"virtualenv\/__init__.py","new_file":"virtualenv\/__init__.py","old_contents":"from __future__ import absolute_import, division, print_function\n\nfrom virtualenv.__about__ import (\n __author__, __copyright__, __email__, __license__, __summary__, __title__,\n __uri__, __version__\n)\nfrom virtualenv.core import create\n\n\n__all__ = [\n \"__title__\", \"__summary__\", \"__uri__\", \"__version__\", \"__author__\",\n \"__email__\", \"__license__\", \"__copyright__\",\n \"create\",\n]\n","new_contents":"from __future__ import absolute_import, division, print_function\n\nfrom virtualenv.__about__ import (\n __author__, __copyright__, __email__, __license__, __summary__, __title__,\n __uri__, __version__\n)\nfrom virtualenv.core import create\n\n\ndef create_environment(\n home_dir,\n site_packages=False, clear=False,\n unzip_setuptools=False,\n prompt=None, search_dirs=None, never_download=False,\n no_setuptools=False, no_pip=False, symlink=True\n):\n create(\n home_dir,\n system_site_packages=site_packages,\n clear=clear,\n prompt=prompt or \"\",\n extra_search_dirs=search_dirs,\n setuptools=not no_setuptools,\n pip=not no_pip\n )\n\n__all__ = [\n \"__title__\", \"__summary__\", \"__uri__\", \"__version__\", \"__author__\",\n \"__email__\", \"__license__\", \"__copyright__\",\n \"create\",\n]\n","subject":"Add a backwards compatible create_environment.","message":"Add a backwards compatible create_environment.\n","lang":"Python","license":"mit","repos":"ionelmc\/virtualenv,ionelmc\/virtualenv,ionelmc\/virtualenv"} {"commit":"7822687bb78cbe422af0d707a1ed7fc94011628d","old_file":"castor\/tasks.py","new_file":"castor\/tasks.py","old_contents":"from celery import Celery\nfrom settings import SETTINGS\nimport requests\n\n\nHOOKS = SETTINGS.get('hooks', [])\nCELERY_SETTINGS = SETTINGS.get('celery', {})\n\napp = Celery()\napp.conf.update(**CELERY_SETTINGS)\n\n\n@app.task\ndef dispatch_event(event):\n event_repr = '%s:%s' % (event['id'][:10], event['status'])\n for url in HOOKS:\n dispatch_tuple = (event_repr, url)\n print '[DISPATCH START] Dispatching event %s --> %s' % dispatch_tuple\n try:\n response = requests.post(url, data=event)\n response_tuple = (response.status_code, response.reason)\n\n if response.status_code >= 400:\n print ' [FAILURE] %s: %s' % response_tuple\n else:\n print ' [SUCCESS] %s: %s' % response_tuple\n except Exception as e:\n print ' [ERROR] Exception: %s' % e\n print '[DISPATCH END] %s --> %s' % dispatch_tuple\n","new_contents":"from celery import Celery\nfrom settings import SETTINGS\nimport requests\n\n\nHOOKS = SETTINGS.get('hooks', [])\nCELERY_SETTINGS = SETTINGS.get('celery', {})\n\napp = Celery()\napp.conf.update(**CELERY_SETTINGS)\n\n\n@app.task\ndef dispatch_event(event):\n event_repr = '%s:%s' % (event['id'][:10], event['status'])\n for url in HOOKS:\n dispatch_tuple = (event_repr, url)\n print '[DISPATCH START] Dispatching event %s --> %s' % dispatch_tuple\n try:\n response = requests.post(url, data=event)\n response_tuple = (response.status_code, response.reason)\n\n if response.status_code >= 400:\n print ' [FAILURE] %s: %s' % response_tuple\n else:\n print ' [SUCCESS] %s: %s' % response_tuple\n except Exception as e:\n print ' [ERROR] Exception: %s' % e\n print '[DISPATCH END] %s --> %s' % dispatch_tuple\n return event\n","subject":"Return event dictionary at the end of the task.","message":"Return event dictionary at the end of the task.\n","lang":"Python","license":"mit","repos":"sourcelair\/castor"} {"commit":"84d743476261d30b352e3bfc103d76e7e8350b4c","old_file":"tests\/test_urls.py","new_file":"tests\/test_urls.py","old_contents":"\"\"\"\nTesting of project level URLs.\n\"\"\"\n\nfrom __future__ import unicode_literals\n\nfrom django.core.urlresolvers import reverse\nfrom django.test import TestCase\nfrom urltools import compare\n\n\nclass TestURLs(TestCase):\n \"\"\"Verify project level URL configuration.\"\"\"\n\n def test_cas_enabled(self):\n \"\"\"Verify that CAS is wired up properly when enabled\"\"\"\n with self.settings(\n CAS_ENABLED=True,\n CAS_SERVER_URL='http:\/\/example.com\/login',\n ):\n # Because this won't actually work, we get in a redirect\n # loop, or at least, best as I can tell.\n response = self.client.get(reverse('cas_login'))\n self.assertTrue(compare(\n 'http:\/\/example.com\/login?'\n 'service=http%3A%2F%2Ftestserver%2Fcas%2Flogin%3Fnext%3D%252F',\n response['location']\n ))\n\n def test_cas_disable(self):\n \"\"\"Verify that when CAS is disabled, login is default\"\"\"\n with self.settings(\n CAS_ENABLED=False\n ):\n response = self.client.get('\/login', follow=True)\n self.assertEqual(404, response.status_code)\n","new_contents":"\"\"\"\nTesting of project level URLs.\n\"\"\"\n\nfrom __future__ import unicode_literals\n\nfrom django.core.urlresolvers import reverse\nfrom django.test import TestCase\n\nimport ssl\n\nif hasattr(ssl, '_create_unverified_context'):\n ssl._create_default_https_context = ssl._create_unverified_context # noqa pylint: disable=protected-access\n\nfrom urltools import compare # noqa\n\n\nclass TestURLs(TestCase):\n \"\"\"Verify project level URL configuration.\"\"\"\n\n def test_cas_enabled(self):\n \"\"\"Verify that CAS is wired up properly when enabled\"\"\"\n with self.settings(\n CAS_ENABLED=True,\n CAS_SERVER_URL='http:\/\/example.com\/login',\n ):\n # Because this won't actually work, we get in a redirect\n # loop, or at least, best as I can tell.\n response = self.client.get(reverse('cas_login'))\n self.assertTrue(compare(\n 'http:\/\/example.com\/login?'\n 'service=http%3A%2F%2Ftestserver%2Fcas%2Flogin%3Fnext%3D%252F',\n response['location']\n ))\n\n def test_cas_disable(self):\n \"\"\"Verify that when CAS is disabled, login is default\"\"\"\n with self.settings(\n CAS_ENABLED=False\n ):\n response = self.client.get('\/login', follow=True)\n self.assertEqual(404, response.status_code)\n","subject":"Disable SSL validation for a test which uses urltools","message":"Disable SSL validation for a test which uses urltools\n\nThis is currently a common problem with python >= 2.7.9:\nhttp:\/\/stackoverflow.com\/questions\/27835619\/ssl-certificate-verify-failed-error\n","lang":"Python","license":"agpl-3.0","repos":"mitodl\/lore,amir-qayyum-khan\/lore,amir-qayyum-khan\/lore,amir-qayyum-khan\/lore,amir-qayyum-khan\/lore,mitodl\/lore,mitodl\/lore,mitodl\/lore,amir-qayyum-khan\/lore,mitodl\/lore"} {"commit":"9fec66d5f8b1e9a0f3c0ec66dd777db578fab07e","old_file":"djangocms_installer\/__init__.py","new_file":"djangocms_installer\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n__author__ = 'Iacopo Spalletti'\n__email__ = 'i.spalletti@nephila.it'\n__version__ = '1.1.0'\n","new_contents":"# -*- coding: utf-8 -*-\n__author__ = 'Iacopo Spalletti'\n__email__ = 'i.spalletti@nephila.it'\n__version__ = '1.1.0.post1'\n","subject":"Bump develop version [ci skip]","message":"Bump develop version [ci skip]\n","lang":"Python","license":"bsd-3-clause","repos":"nephila\/djangocms-installer,nephila\/djangocms-installer"} {"commit":"2eb104ac53d40548906f4e9798a9623193da9e04","old_file":"djangocms_page_meta\/__init__.py","new_file":"djangocms_page_meta\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import, print_function, unicode_literals\n\n__version__ = '0.8.3'\n__author__ = 'Iacopo Spalletti <i.spalletti@nephila.it>'\n\ndefault_app_config = 'djangocms_page_meta.apps.PageMetaConfig'\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import, print_function, unicode_literals\n\n__version__ = '0.8.3.post1'\n__author__ = 'Iacopo Spalletti <i.spalletti@nephila.it>'\n\ndefault_app_config = 'djangocms_page_meta.apps.PageMetaConfig'\n","subject":"Bump development version [ci skip]","message":"Bump development version [ci skip]\n","lang":"Python","license":"bsd-3-clause","repos":"nephila\/djangocms-page-meta,nephila\/djangocms-page-meta"} {"commit":"8b0c08962d18536b87a948c96f7ec7daabd8b4e1","old_file":"NEO_flyby.py","new_file":"NEO_flyby.py","old_contents":"\nimport time\nimport datetime\nimport requests\nimport json\n\ndef get_NEO_flyby():\n\n neo_data = []\n\n\n unix = time.time()\n datestamp = datetime.datetime.fromtimestamp(unix).strftime(\"%Y-%b-%d\")\n\n json_data_url = requests.get(\"https:\/\/ssd-api.jpl.nasa.gov\/cad.api?body=Earth&dist-max=20LD\")\n json_data = json.loads(json_data_url.text)\n for i in range(len(json_data[\"data\"])):\n neo_date = json_data[\"data\"][i][3][:11]\n neo_time = json_data[\"data\"][i][3][11:]\n\n if neo_date == datestamp:\n neo_data.append((json_data[\"data\"][i][0],))\n\n\n\n\n \n # sorte lieber per magnitude und nimm nur das größte objekt, sonst ist der tweet zu lang\n\n\n\nget_NEO_flyby()\n\n\n# TODO: Add api indicator of numbers\n# TODO: Iterate over data and return tuple\n\n\n","new_contents":"\nimport time\nimport datetime\nimport requests\nimport json\n\n\ndef get_NEO_flyby():\n\n neo_data = []\n\n des = 0\n orbit_id = 1\n jd = 2\n cd = 3\n dist = 4\n dist_min = 5\n dist_max = 6\n v_rel = 7\n v_inf = 8\n t_signma_F = 9\n body = 10\n h = 11\n\n unix = time.time()\n datestamp = datetime.datetime.fromtimestamp(unix).strftime(\"%Y-%b-%d\")\n\n json_data_url = requests.get(\"https:\/\/ssd-api.jpl.nasa.gov\/cad.api?body=Earth&dist-max=20LD\")\n json_data = json.loads(json_data_url.text)\n for i in range(len(json_data[\"data\"])):\n neo_date = json_data[\"data\"][i][cd][:11]\n neo_time = json_data[\"data\"][i][cd][11:]\n\n if neo_date == datestamp:\n neo_data.append((json_data[\"data\"][i][des],))\n\n\n\n\n \n # sorte lieber per magnitude und nimm nur das größte objekt, sonst ist der tweet zu lang\n\n\n\nget_NEO_flyby()\n\n\n# TODO: Iterate over data and return tuple\n\n\n","subject":"Update 0.1.1 - Added api number reference","message":"Update 0.1.1\n- Added api number reference\n","lang":"Python","license":"mit","repos":"FXelix\/space_facts_bot"} {"commit":"9e5b6ea80dd1039952bb5ff821ae15555ad591be","old_file":"iterm2_tools\/images.py","new_file":"iterm2_tools\/images.py","old_contents":"import sys\nimport os\nimport base64\n\n# See https:\/\/iterm2.com\/images.html\nIMAGE_CODE= '\\033]1337;File={file};inline={inline};size={size}:{base64_img}\\a'\n\ndef iterm2_image_bytes(b, filename=None, inline=1):\n data = {\n 'file': base64.b64encode((filename or 'Unnamed file').encode('utf-8')).decode('ascii'),\n 'inline': inline,\n 'size': len(b),\n 'base64_img': base64.b64encode(b).decode('ascii'),\n }\n return (IMAGE_CODE.format(**data))\n\ndef iterm2_display_image_file(fn):\n \"\"\"\n Display an image in the terminal.\n\n A newline is not printed.\n \"\"\"\n with open(os.path.realpath(os.path.expanduser(fn)), 'br') as f:\n sys.stdout.write(iterm2_image_bytes(f.read(), filename=fn))\n","new_contents":"import sys\nimport os\nimport base64\n\n# See https:\/\/iterm2.com\/images.html\nIMAGE_CODE= '\\033]1337;File={file};inline={inline};size={size}:{base64_img}\\a'\n\ndef iterm2_image_bytes(b, filename=None, inline=1):\n data = {\n 'file': base64.b64encode((filename or 'Unnamed file').encode('utf-8')).decode('ascii'),\n 'inline': inline,\n 'size': len(b),\n 'base64_img': base64.b64encode(b).decode('ascii'),\n }\n return (IMAGE_CODE.format(**data))\n\ndef iterm2_display_image_file(fn):\n \"\"\"\n Display an image in the terminal.\n\n A newline is not printed.\n \"\"\"\n with open(os.path.realpath(os.path.expanduser(fn)), 'rb') as f:\n sys.stdout.write(iterm2_image_bytes(f.read(), filename=fn))\n","subject":"Fix Python 2 compatibility issue","message":"Fix Python 2 compatibility issue\n","lang":"Python","license":"mit","repos":"asmeurer\/iterm2-tools"} {"commit":"7570b9dd8fada221e5059a00e107ce6665b9c563","old_file":"nailgun\/nailgun\/db\/sqlalchemy\/models\/base.py","new_file":"nailgun\/nailgun\/db\/sqlalchemy\/models\/base.py","old_contents":"# -*- coding: utf-8 -*-\n\n# Copyright 2013 Mirantis, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom datetime import datetime\n\nfrom sqlalchemy import Column\nfrom sqlalchemy import DateTime\nfrom sqlalchemy import Integer\n\nfrom sqlalchemy.ext.declarative import declarative_base\n\nfrom nailgun.db.sqlalchemy.models.fields import JSON\n\nfrom nailgun.openstack.common.db.sqlalchemy import models\n\n\nBase = declarative_base(cls=models.ModelBase)\n\n\nclass CapacityLog(Base):\n __tablename__ = 'capacity_log'\n\n id = Column(Integer, primary_key=True)\n report = Column(JSON)\n datetime = Column(DateTime, default=datetime.now())\n","new_contents":"# -*- coding: utf-8 -*-\n\n# Copyright 2013 Mirantis, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\nfrom datetime import datetime\n\nfrom sqlalchemy import Column\nfrom sqlalchemy import DateTime\nfrom sqlalchemy import Integer\n\nfrom sqlalchemy.ext.declarative import declarative_base\n\nfrom nailgun.db.sqlalchemy.models.fields import JSON\n\nfrom nailgun.openstack.common.db.sqlalchemy import models\n\n\nBase = declarative_base(cls=models.ModelBase)\n\n\nclass CapacityLog(Base):\n __tablename__ = 'capacity_log'\n\n id = Column(Integer, primary_key=True)\n report = Column(JSON)\n datetime = Column(DateTime, default=lambda: datetime.now())\n","subject":"Fix select order for Capacity Log","message":"Fix select order for Capacity Log\n\nChange-Id: I0db3de15e65bb300d75741e5c86e164b1966ac89\nCloses: bug #1281986\n","lang":"Python","license":"apache-2.0","repos":"nebril\/fuel-web,SmartInfrastructures\/fuel-web-dev,stackforge\/fuel-web,nebril\/fuel-web,huntxu\/fuel-web,nebril\/fuel-web,eayunstack\/fuel-web,zhaochao\/fuel-web,huntxu\/fuel-web,koder-ua\/nailgun-fcert,zhaochao\/fuel-web,huntxu\/fuel-web,stackforge\/fuel-web,koder-ua\/nailgun-fcert,prmtl\/fuel-web,SmartInfrastructures\/fuel-web-dev,koder-ua\/nailgun-fcert,zhaochao\/fuel-web,nebril\/fuel-web,huntxu\/fuel-web,zhaochao\/fuel-web,nebril\/fuel-web,koder-ua\/nailgun-fcert,eayunstack\/fuel-web,stackforge\/fuel-web,SmartInfrastructures\/fuel-web-dev,eayunstack\/fuel-web,prmtl\/fuel-web,prmtl\/fuel-web,SmartInfrastructures\/fuel-web-dev,prmtl\/fuel-web,prmtl\/fuel-web,huntxu\/fuel-web,SmartInfrastructures\/fuel-web-dev,eayunstack\/fuel-web,eayunstack\/fuel-web,zhaochao\/fuel-web"} {"commit":"dd8d4515f5e39dcc3f23db5a5acf3478c9c16ae2","old_file":"codebox\/conf.py","new_file":"codebox\/conf.py","old_contents":"\"\"\"\ncodebox.conf\n~~~~~~~~~~~\n\n:copyright: (c) 2011 DISQUS.\n:license: Apache License 2.0, see LICENSE for more details.\n\"\"\"\n\nimport os, os.path\nimport urlparse\n\nclass Config(object):\n DEBUG = True\n TESTING = False\n SECRET_KEY = '\\x89\\x1d\\xec\\x8eJ\\xda=C`\\xf3<X\\x81\\xff\\x1e\\r{+\\x1b\\xe1\\xd1@ku'\n REDIS_DB = 0\n JANRAIN_API_KEY = '288a1ca2fedb4e1d1780c320fa4082ae69640a52'\n PODIO_CLIENT_ID = \"dcramer@gmail.com\"\n PODIO_KEY = \"f7qFIBcPTfTBLOd8ondkO9UGqU6uN1iG\"\n DOMAIN_BLACKLIST = ['gmail.com', 'hotmail.com', 'live.com', 'msn.com', 'yahoo.com', 'googlemail.com', 'facebookmail.com']\n\nif os.environ.has_key('REDISTOGO_URL'):\n # 'redis:\/\/username:password@my.host:6789' \n urlparse.uses_netloc.append('redis')\n url = urlparse.urlparse(os.environ['REDISTOGO_URL'])\n Config.REDIS_USER = url.username\n Config.REDIS_PASSWORD = url.password\n Config.REDIS_HOST = url.hostname\n Config.REDIS_PORT = url.port\n\nclass TestingConfig(Config):\n REDIS_DB = 9\n TESTING = True\n","new_contents":"\"\"\"\ncodebox.conf\n~~~~~~~~~~~\n\n:copyright: (c) 2011 DISQUS.\n:license: Apache License 2.0, see LICENSE for more details.\n\"\"\"\n\nimport os, os.path\nimport urlparse\n\nclass Config(object):\n DEBUG = True\n TESTING = False\n SECRET_KEY = os.environ.get('SECRET_KEY', '\\x89\\x1d\\xec\\x8eJ\\xda=C`\\xf3<X\\x81\\xff\\x1e\\r{+\\x1b\\xe1\\xd1@ku')\n REDIS_DB = 0\n JANRAIN_API_KEY = os.environ.get('JANRAIN_API_KEY')\n DOMAIN_BLACKLIST = ['gmail.com', 'hotmail.com', 'live.com', 'msn.com', 'yahoo.com', 'googlemail.com', 'facebookmail.com']\n\nif os.environ.has_key('REDISTOGO_URL'):\n # 'redis:\/\/username:password@my.host:6789' \n urlparse.uses_netloc.append('redis')\n url = urlparse.urlparse(os.environ['REDISTOGO_URL'])\n Config.REDIS_USER = url.username\n Config.REDIS_PASSWORD = url.password\n Config.REDIS_HOST = url.hostname\n Config.REDIS_PORT = url.port\n\nclass TestingConfig(Config):\n REDIS_DB = 9\n TESTING = True\n","subject":"Read janrain\/secret key from env","message":"Read janrain\/secret key from env\n","lang":"Python","license":"apache-2.0","repos":"disqus\/codebox,disqus\/codebox"} {"commit":"a17d3fbf19b25e1da568266b17abe575071e3f80","old_file":"server\/lib\/utils.py","new_file":"server\/lib\/utils.py","old_contents":"import yaml\nimport json\nimport os\nimport uuid\n\ndef loadFromFile(path, bytes=False):\n from config import PATH\n if not os.path.isabs(path):\n path = os.path.join(PATH, path)\n readType = 'r' if not bytes else 'rb'\n with open(path, readType) as file:\n fileContents = file.read()\n file.close()\n return fileContents\n\ndef loadYaml(fileName):\n return yaml.load(loadFromFile(fileName))\n\ndef loadJson(fileName):\n return json.loads(loadFromFile(fileName))\n\ndef writeFile(fileName, content):\n path = '\/'.join(os.path.dirname(__file__).split('\/')[0:-1])\n with open((os.path.join(path,fileName)), 'w') as file:\n file.write(content)\n file.close()\n\ndef getPath(path):\n from config import PATH\n return os.path.join(PATH, path)\n\ndef addUniqueIdToFile(filename):\n splitFilename = filename.split('.')\n splitFilename[0] = '{filename}-{id}'.format(filename=splitFilename[0], id=str(uuid.uuid4())[:6])\n return '.'.join(splitFilename)\n\ndef removeValueFromDict(k, value):\n for key in k:\n del k[key][value]\n return k\n\ndef additionalDeliveryInfo(delivery):\n if delivery == 'express':\n return 'express dodání(+30% ceny)'\n else:\n return delivery\n","new_contents":"import yaml\nimport json\nimport os\nimport uuid\n\ndef loadFromFile(path, bytes=False):\n from config import PATH\n if not os.path.isabs(path):\n path = os.path.join(PATH, path)\n readType = 'r' if not bytes else 'rb'\n with open(path, readType, encoding='utf-8') as file:\n fileContents = file.read()\n file.close()\n return fileContents\n\ndef loadYaml(fileName):\n return yaml.load(loadFromFile(fileName))\n\ndef loadJson(fileName):\n return json.loads(loadFromFile(fileName))\n\ndef writeFile(fileName, content):\n path = '\/'.join(os.path.dirname(__file__).split('\/')[0:-1])\n with open((os.path.join(path,fileName)), 'w') as file:\n file.write(content)\n file.close()\n\ndef getPath(path):\n from config import PATH\n return os.path.join(PATH, path)\n\ndef addUniqueIdToFile(filename):\n splitFilename = filename.split('.')\n splitFilename[0] = '{filename}-{id}'.format(filename=splitFilename[0], id=str(uuid.uuid4())[:6])\n return '.'.join(splitFilename)\n\ndef removeValueFromDict(k, value):\n for key in k:\n del k[key][value]\n return k\n\ndef additionalDeliveryInfo(delivery):\n if delivery == 'express':\n return 'express dodání(+30% ceny)'\n else:\n return delivery\n","subject":"Fix not being able to parse diacritics","message":"Fix not being able to parse diacritics\n\n","lang":"Python","license":"agpl-3.0","repos":"MakersLab\/custom-print"} {"commit":"09572d5f24c33b6a604cc038c4e1c13ddb977af6","old_file":"db\/db.py","new_file":"db\/db.py","old_contents":"#!\/usr\/bin\/python\nimport sys\nimport copy\nimport json\nimport getpass\nimport aesjsonfile\n\nsys.path.append(\"..\/\")\n\nimport config\n\nclass DB(object):\n def __init__(self, username, password):\n self.username = username\n self.password = password\n self.db = aesjsonfile.load(\"%s\/%s.json\"%(config.dbdir, self.username), self.password)\n\n def save(self):\n aesjsonfile.dump(\"%s\/%s.json\"%(config.dbdir, self.username), self.db, self.password)\n\n def accountstodo(self):\n return self.db[\"accounts\"]\n\n def accounts(self):\n ret = copy.deepcopy(self.db[\"accounts\"])\n for acct in ret:\n acct.pop(\"password\",None)\n return ret\n\nif __name__ == \"__main__\":\n if len(sys.argv) < 2:\n sys.exit(1)\n password = getpass.getpass()\n db = DB(sys.argv[1],password)\n print \"accountstodo\"\n print json.dumps(db.accountstodo(),indent=2)\n print \"accounts\"\n print json.dumps(db.accounts(),indent=2)\n","new_contents":"#!\/usr\/bin\/python\nimport sys\nimport copy\nimport json\nimport getpass\nimport aesjsonfile\n\nsys.path.append(\"..\/\")\n\nimport config\n\nclass DB(object):\n def __init__(self, username, password):\n self.username = username\n self.password = password\n self.db = aesjsonfile.load(\"%s\/%s.json\"%(config.dbdir, self.username), self.password)\n\n def save(self):\n aesjsonfile.dump(\"%s\/%s.json\"%(config.dbdir, self.username), self.db, self.password)\n\n def accountstodo(self):\n return self.db[\"accounts\"]\n\n def accounts(self):\n ret = copy.deepcopy(self.db[\"accounts\"])\n for acct in ret:\n acct.pop(\"password\",None)\n acct[\"subaccounts\"] = []\n for sub in self.db.get(\"balances\",{}).get(acct[\"name\"],{}):\n acct[\"subaccounts\"].append({\"name\": sub, \"amount\": self.db[\"balances\"][acct[\"name\"]][sub][0][\"amount\"],\n \"date\": self.db[\"balances\"][acct[\"name\"]][sub][0][\"lastdate\"]})\n return ret\n\nif __name__ == \"__main__\":\n if len(sys.argv) < 2:\n sys.exit(1)\n password = getpass.getpass()\n db = DB(sys.argv[1],password)\n print \"accountstodo\"\n print json.dumps(db.accountstodo(),indent=2)\n print \"accounts\"\n print json.dumps(db.accounts(),indent=2)\n","subject":"Return subaccounts with their balances on account query.","message":"Return subaccounts with their balances on account query.\n","lang":"Python","license":"agpl-3.0","repos":"vincebusam\/pyWebCash,vincebusam\/pyWebCash,vincebusam\/pyWebCash"} {"commit":"79e22a810638fbf2098f87525fa5a68d3c3b8c49","old_file":"hitcount\/management\/commands\/hitcount_cleanup.py","new_file":"hitcount\/management\/commands\/hitcount_cleanup.py","old_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom datetime import timedelta\n\nfrom django.conf import settings\nfrom django.utils import timezone\n\ntry:\n from django.core.management.base import BaseCommand\nexcept ImportError:\n from django.core.management.base import NoArgsCommand as BaseCommand\n\nfrom hitcount.models import Hit\n\n\nclass Command(BaseCommand):\n help = \"Can be run as a cronjob or directly to clean out old Hits objects from the database.\"\n\n def __init__(self, *args, **kwargs):\n super(Command, self).__init__(*args, **kwargs)\n\n def handle(self, *args, **kwargs):\n self.handle_noargs()\n\n def handle_noargs(self, **options):\n grace = getattr(settings, 'HITCOUNT_KEEP_HIT_IN_DATABASE', {'days': 30})\n period = timezone.now() - timedelta(**grace)\n qs = Hit.objects.filter(created__lt=period)\n number_removed = len(qs)\n qs.delete()\n self.stdout.write('Successfully removed %s Hits' % number_removed)\n","new_contents":"# -*- coding: utf-8 -*-\nfrom __future__ import unicode_literals\n\nfrom datetime import timedelta\n\nfrom django.conf import settings\nfrom django.utils import timezone\n\ntry:\n from django.core.management.base import BaseCommand\nexcept ImportError:\n from django.core.management.base import NoArgsCommand as BaseCommand\n\nfrom hitcount.models import Hit\n\n\nclass Command(BaseCommand):\n help = \"Can be run as a cronjob or directly to clean out old Hits objects from the database.\"\n\n def __init__(self, *args, **kwargs):\n super(Command, self).__init__(*args, **kwargs)\n\n def handle(self, *args, **kwargs):\n self.handle_noargs()\n\n def handle_noargs(self, **options):\n grace = getattr(settings, 'HITCOUNT_KEEP_HIT_IN_DATABASE', {'days': 30})\n period = timezone.now() - timedelta(**grace)\n qs = Hit.objects.filter(created__lt=period)\n number_removed = qs.count()\n qs.delete()\n self.stdout.write('Successfully removed %s Hits' % number_removed)\n","subject":"Use count() on queryset instead of len()","message":"Use count() on queryset instead of len()\n\nEnsure a fast query even for millions of rows.\n","lang":"Python","license":"mit","repos":"thornomad\/django-hitcount,thornomad\/django-hitcount,thornomad\/django-hitcount"} {"commit":"da193c2606daac82d0693cc10decdbf2d3162fa5","old_file":"lib\/markdown_deux\/conf\/settings.py","new_file":"lib\/markdown_deux\/conf\/settings.py","old_contents":"# Copyright (c) 2010 ActiveState Software Inc.\n\nfrom django.conf import settings\n\nMARKDOWN_DEUX_HELP_URL = getattr(settings, \"MARKDOWN_DEUX_HELP_URL\",\n \"http:\/\/daringfireball.net\/projects\/markdown\/syntax\")\n\nMARKDOWN_DEUX_DEFAULT_STYLE = {\n \"extras\": {\n \"code-friendly\": None,\n },\n \"safe_mode\": \"escape\",\n}\n\nMARKDOWN_DEUX_STYLES = getattr(settings, \"MARKDOWN_DEUX_STYLES\",\n {\"default\": MARKDOWN_DEUX_DEFAULT_STYLE})\n\n","new_contents":"# Copyright (c) 2010 ActiveState Software Inc.\n\nfrom django.conf import settings\n\nMARKDOWN_DEUX_HELP_URL = getattr(settings, \"MARKDOWN_DEUX_HELP_URL\",\n \"http:\/\/daringfireball.net\/projects\/markdown\/syntax\")\n\nMARKDOWN_DEUX_DEFAULT_STYLE = {\n \"extras\": {\n \"code-friendly\": None,\n },\n \"safe_mode\": \"escape\",\n}\n\nMARKDOWN_DEUX_STYLES = getattr(settings, \"MARKDOWN_DEUX_STYLES\",\n {\"default\": MARKDOWN_DEUX_DEFAULT_STYLE})\n\nDEBUG = settings.DEBUG\n","subject":"Fix debug flag import error","message":"Fix debug flag import error\n","lang":"Python","license":"mit","repos":"gogobook\/django-markdown-deux,gogobook\/django-markdown-deux,trentm\/django-markdown-deux,douzepouze\/django-markdown-tag"} {"commit":"57b1dbc45e7b78f7aa272fd5b7d4bd022850beb9","old_file":"lametro\/migrations\/0007_update_packet_links.py","new_file":"lametro\/migrations\/0007_update_packet_links.py","old_contents":"# Generated by Django 2.2.24 on 2021-10-22 19:54\n\nfrom django.db import migrations\n\n\ndef resave_packets(apps, schema_editor):\n '''\n Re-save all existing packets to update their URLs based on the\n new value of MERGE_HOST.\n '''\n for packet in ('BillPacket', 'EventPacket'):\n packet_model = apps.get_model('lametro', packet)\n for p in packet_model.objects.all():\n p.save(merge=False)\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('lametro', '0006_add_plan_program_policy'),\n ]\n\n operations = [\n migrations.RunPython(resave_packets),\n ]\n","new_contents":"# Generated by Django 2.2.24 on 2021-10-22 19:54\n\nfrom django.db import migrations\n\n\ndef resave_packets(apps, schema_editor):\n '''\n Re-save all existing packets to update their URLs based on the\n new value of MERGE_HOST.\n '''\n# for packet in ('BillPacket', 'EventPacket'):\n# packet_model = apps.get_model('lametro', packet)\n# for p in packet_model.objects.all():\n# p.save(merge=False)\n return\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('lametro', '0006_add_plan_program_policy'),\n ]\n\n operations = [\n migrations.RunPython(resave_packets),\n ]\n","subject":"Disable data migration for deployment","message":"Disable data migration for deployment\n","lang":"Python","license":"mit","repos":"datamade\/la-metro-councilmatic,datamade\/la-metro-councilmatic,datamade\/la-metro-councilmatic,datamade\/la-metro-councilmatic"} {"commit":"ce36dd825635c8487fcd9f83bd686a2dce7c318c","old_file":"hello.py","new_file":"hello.py","old_contents":"from flask import Flask\nfrom flask import request\nimport os\nfrom dogapi import dog_http_api as api\n\napp = Flask(__name__)\n\napi.api_key = os.environ.get('DD_API_KEY')\n\naction_url = \"\/\" + os.environ.get('BASE_URL') + \"\/\"\n\n@app.route(action_url, methods=['POST', 'GET'])\ndef hello():\n api.metric('mailgun.event', (request.args.post('timestamp'), 1), tags=[\"event_name:\" + request.args.post('event')])\n return \"200\"\n\nif __name__ == \"__main__\":\n port = int(os.environ.get(\"PORT\", 5000))\n app.run(host='0.0.0.0', port=port)\n","new_contents":"from flask import Flask\nfrom flask import request\nimport os\nfrom dogapi import dog_http_api as api\n\napp = Flask(__name__)\n\napi.api_key = os.environ.get('DD_API_KEY')\n\naction_url = \"\/\" + os.environ.get('BASE_URL') + \"\/\"\n\n@app.route(action_url, methods=['POST', 'GET'])\ndef hello():\n api.metric('mailgun.event', (request.form('timestamp'), 1), tags=[\"event_name:\" + request.form('event')])\n return \"200\"\n\nif __name__ == \"__main__\":\n port = int(os.environ.get(\"PORT\", 5000))\n app.run(host='0.0.0.0', port=port)\n","subject":"Use the right style of request.","message":"Use the right style of request.\n","lang":"Python","license":"apache-2.0","repos":"darron\/mailgun_datadog"} {"commit":"c711d5e2dbca4b95bebc0eed4d48a35eb3c7a998","old_file":"website\/addons\/dropbox\/settings\/local-dist.py","new_file":"website\/addons\/dropbox\/settings\/local-dist.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"Example Dropbox local settings file. Copy this file to local.py and change\nthese settings.\n\"\"\"\n# Get an app key and secret at https:\/\/www.dropbox.com\/developers\/apps\nDROPBOX_KEY = 'changeme'\nDROPBOX_SECRET = 'changeme'\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"Example Dropbox local settings file. Copy this file to local.py and change\nthese settings.\n\"\"\"\n# Get an app key and secret at https:\/\/www.dropbox.com\/developers\/apps\nDROPBOX_KEY = 'jnpncg5s2fc7cj8'\nDROPBOX_SECRET = 'sjqv1hrk7sonhu1'","subject":"Add dropbox credentials for testing.","message":"Add dropbox credentials for testing.\n","lang":"Python","license":"apache-2.0","repos":"crcresearch\/osf.io,acshi\/osf.io,felliott\/osf.io,TomHeatwole\/osf.io,RomanZWang\/osf.io,jnayak1\/osf.io,baylee-d\/osf.io,TomBaxter\/osf.io,mluke93\/osf.io,mluo613\/osf.io,pattisdr\/osf.io,samchrisinger\/osf.io,wearpants\/osf.io,mfraezz\/osf.io,kch8qx\/osf.io,Nesiehr\/osf.io,adlius\/osf.io,RomanZWang\/osf.io,abought\/osf.io,felliott\/osf.io,jnayak1\/osf.io,caseyrollins\/osf.io,doublebits\/osf.io,cslzchen\/osf.io,kwierman\/osf.io,monikagrabowska\/osf.io,rdhyee\/osf.io,laurenrevere\/osf.io,Nesiehr\/osf.io,HalcyonChimera\/osf.io,icereval\/osf.io,zamattiac\/osf.io,CenterForOpenScience\/osf.io,jnayak1\/osf.io,cwisecarver\/osf.io,SSJohns\/osf.io,icereval\/osf.io,monikagrabowska\/osf.io,wearpants\/osf.io,chrisseto\/osf.io,binoculars\/osf.io,monikagrabowska\/osf.io,mluo613\/osf.io,adlius\/osf.io,aaxelb\/osf.io,monikagrabowska\/osf.io,asanfilippo7\/osf.io,icereval\/osf.io,brianjgeiger\/osf.io,amyshi188\/osf.io,cwisecarver\/osf.io,DanielSBrown\/osf.io,crcresearch\/osf.io,kch8qx\/osf.io,SSJohns\/osf.io,abought\/osf.io,crcresearch\/osf.io,laurenrevere\/osf.io,mluo613\/osf.io,baylee-d\/osf.io,alexschiller\/osf.io,zachjanicki\/osf.io,aaxelb\/osf.io,rdhyee\/osf.io,doublebits\/osf.io,amyshi188\/osf.io,Nesiehr\/osf.io,sloria\/osf.io,hmoco\/osf.io,HalcyonChimera\/osf.io,aaxelb\/osf.io,RomanZWang\/osf.io,TomHeatwole\/osf.io,kch8qx\/osf.io,chrisseto\/osf.io,TomBaxter\/osf.io,aaxelb\/osf.io,DanielSBrown\/osf.io,mattclark\/osf.io,emetsger\/osf.io,emetsger\/osf.io,binoculars\/osf.io,zachjanicki\/osf.io,kwierman\/osf.io,kwierman\/osf.io,kwierman\/osf.io,sloria\/osf.io,mfraezz\/osf.io,kch8qx\/osf.io,acshi\/osf.io,chennan47\/osf.io,caneruguz\/osf.io,doublebits\/osf.io,mluke93\/osf.io,erinspace\/osf.io,alexschiller\/osf.io,mluo613\/osf.io,zamattiac\/osf.io,alexschiller\/osf.io,caseyrollins\/osf.io,zachjanicki\/osf.io,cwisecarver\/osf.io,samchrisinger\/osf.io,TomBaxter\/osf.io,wearpants\/osf.io,amyshi188\/osf.io,Johnetordoff\/osf.io,HalcyonChimera\/osf.io,laurenrevere\/osf.io,cslzchen\/osf.io,SSJohns\/osf.io,monikagrabowska\/osf.io,Johnetordoff\/osf.io,asanfilippo7\/osf.io,zamattiac\/osf.io,CenterForOpenScience\/osf.io,abought\/osf.io,jnayak1\/osf.io,Johnetordoff\/osf.io,mfraezz\/osf.io,samchrisinger\/osf.io,DanielSBrown\/osf.io,rdhyee\/osf.io,leb2dg\/osf.io,acshi\/osf.io,mattclark\/osf.io,chrisseto\/osf.io,leb2dg\/osf.io,brianjgeiger\/osf.io,zachjanicki\/osf.io,mluke93\/osf.io,binoculars\/osf.io,asanfilippo7\/osf.io,felliott\/osf.io,DanielSBrown\/osf.io,TomHeatwole\/osf.io,hmoco\/osf.io,kch8qx\/osf.io,caneruguz\/osf.io,saradbowman\/osf.io,felliott\/osf.io,adlius\/osf.io,doublebits\/osf.io,caneruguz\/osf.io,samchrisinger\/osf.io,HalcyonChimera\/osf.io,RomanZWang\/osf.io,emetsger\/osf.io,mluo613\/osf.io,hmoco\/osf.io,hmoco\/osf.io,RomanZWang\/osf.io,emetsger\/osf.io,rdhyee\/osf.io,mluke93\/osf.io,acshi\/osf.io,leb2dg\/osf.io,zamattiac\/osf.io,saradbowman\/osf.io,leb2dg\/osf.io,pattisdr\/osf.io,chennan47\/osf.io,acshi\/osf.io,cslzchen\/osf.io,alexschiller\/osf.io,SSJohns\/osf.io,chennan47\/osf.io,erinspace\/osf.io,mattclark\/osf.io,CenterForOpenScience\/osf.io,chrisseto\/osf.io,brianjgeiger\/osf.io,caseyrollins\/osf.io,brianjgeiger\/osf.io,alexschiller\/osf.io,Nesiehr\/osf.io,amyshi188\/osf.io,baylee-d\/osf.io,CenterForOpenScience\/osf.io,erinspace\/osf.io,abought\/osf.io,wearpants\/osf.io,asanfilippo7\/osf.io,cslzchen\/osf.io,adlius\/osf.io,TomHeatwole\/osf.io,pattisdr\/osf.io,cwisecarver\/osf.io,mfraezz\/osf.io,caneruguz\/osf.io,Johnetordoff\/osf.io,doublebits\/osf.io,sloria\/osf.io"} {"commit":"df3bbdcf08dafbf2fd6997638a575fb4e47ac61f","old_file":"factory\/tools\/cat_StarterLog.py","new_file":"factory\/tools\/cat_StarterLog.py","old_contents":"#!\/bin\/env python\n#\n# cat_StarterLog.py\n#\n# Print out the StarterLog for a glidein output file\n#\n# Usage: cat_StarterLog.py logname\n#\n\nimport os.path\nimport sys\nSTARTUP_DIR=sys.path[0]\nsys.path.append(os.path.join(STARTUP_DIR,\"lib\"))\nimport gWftLogParser\n\nUSAGE=\"Usage: cat_StarterLog.py [-monitor] <logname>\"\n\ndef main():\n if sys.argv[1]=='-monitor':\n fname=sys.argv[2]\n condor_log_id=\"((StarterLog.monitor)|(StarterLog.vm1))\"\n else:\n fname=sys.argv[1]\n condor_log_id=\"((StarterLog)|(StarterLog.vm2))\"\n \n try:\n print gWftLogParser.get_CondorLog(sys.argv[1],\"((StarterLog)|(StarterLog.vm2))\")\n except:\n sys.stderr.write(\"%s\\n\"%USAGE)\n sys.exit(1)\n\n\nif __name__ == '__main__':\n main()\n \n","new_contents":"#!\/bin\/env python\n#\n# cat_StarterLog.py\n#\n# Print out the StarterLog for a glidein output file\n#\n# Usage: cat_StarterLog.py logname\n#\n\nimport os.path\nimport sys\nSTARTUP_DIR=sys.path[0]\nsys.path.append(os.path.join(STARTUP_DIR,\"lib\"))\nimport gWftLogParser\n\nUSAGE=\"Usage: cat_StarterLog.py [-monitor] <logname>\"\n\ndef main():\n if sys.argv[1]=='-monitor':\n fname=sys.argv[2]\n condor_log_id=\"((StarterLog.monitor)|(StarterLog.vm1))\"\n else:\n fname=sys.argv[1]\n condor_log_id=\"((StarterLog)|(StarterLog.vm2))\"\n \n try:\n print gWftLogParser.get_CondorLog(fname,condor_log_id)\n except:\n sys.stderr.write(\"%s\\n\"%USAGE)\n sys.exit(1)\n\n\nif __name__ == '__main__':\n main()\n \n","subject":"Add support for monitor starterlog","message":"Add support for monitor starterlog\n","lang":"Python","license":"bsd-3-clause","repos":"holzman\/glideinwms-old,bbockelm\/glideinWMS,bbockelm\/glideinWMS,holzman\/glideinwms-old,holzman\/glideinwms-old,bbockelm\/glideinWMS,bbockelm\/glideinWMS"} {"commit":"dd6c3abcfc01b22528440e5dd62a1d3d3453b8b3","old_file":"djangopress\/templatetags\/djangopress_tags.py","new_file":"djangopress\/templatetags\/djangopress_tags.py","old_contents":"\"\"\"Templatetags for djangopress.\"\"\"\nfrom datetime import date\nfrom collections import defaultdict\n\nfrom django import template\n\nfrom djangopress.models import Post, Category\n\n\nregister = template.Library()\n\n\n@register.inclusion_tag('djangopress\/tags\/archive_list.html')\ndef archive_list():\n \"\"\"List post by date\"\"\"\n posts = Post.objects.all()\n years_dictionary = defaultdict(set)\n for post in posts:\n year = post.creation_date.year\n month = post.creation_date.month\n years_dictionary[year].add(month)\n years = {}\n for year, months in years_dictionary.items():\n year = str(year)\n years[year] = []\n for month in months:\n years[year].append(date(int(year), month, 1))\n return {'years': years}\n\n\n@register.inclusion_tag('djangopress\/tags\/category_list.html')\ndef category_list():\n \"\"\"List the categories in the blog.\"\"\"\n categories = Category.objects.all()\n return {'categories': categories}\n","new_contents":"\"\"\"Templatetags for djangopress.\"\"\"\nfrom datetime import date\nfrom collections import defaultdict\n\nfrom django import template\n\nfrom djangopress.models import Post, Category\n\n\nregister = template.Library()\n\n\n@register.inclusion_tag('djangopress\/tags\/archive_list.html')\ndef archive_list():\n \"\"\"List post by date\"\"\"\n posts = Post.objects.all()\n years_dictionary = defaultdict(set)\n for post in posts:\n year = post.creation_date.year\n month = post.creation_date.month\n years_dictionary[year].add(month)\n years = {}\n for year, months in years_dictionary.items():\n year = str(year)\n years[year] = []\n for month in months:\n years[year].append(date(int(year), month, 1))\n for year in years:\n years[year].sort(reverse=True)\n return {'years': years}\n\n\n@register.inclusion_tag('djangopress\/tags\/category_list.html')\ndef category_list():\n \"\"\"List the categories in the blog.\"\"\"\n categories = Category.objects.all()\n return {'categories': categories}\n","subject":"Sort the months in archive_templatetag","message":"Sort the months in archive_templatetag\n","lang":"Python","license":"mit","repos":"gilmrjc\/djangopress,gilmrjc\/djangopress,gilmrjc\/djangopress"} {"commit":"226cb45ba39d13c4bf2b40f3998b2631f9f461a6","old_file":"Lib\/cluster\/__init__.py","new_file":"Lib\/cluster\/__init__.py","old_contents":"#\n# cluster - Vector Quantization \/ Kmeans\n#\n\nfrom info import __doc__\n\n__all__ = ['vq']\n\nimport vq\n","new_contents":"#\n# cluster - Vector Quantization \/ Kmeans\n#\n\nfrom info import __doc__\n\n__all__ = ['vq']\n\nimport vq\nfrom numpy.testing import NumpyTest\ntest = NumpyTest().test\n","subject":"Add missing test definition in scipy.cluster","message":"Add missing test definition in scipy.cluster\n\ngit-svn-id: 003f22d385e25de9cff933a5ea4efd77cb5e7b28@2941 d6536bca-fef9-0310-8506-e4c0a848fbcf\n","lang":"Python","license":"bsd-3-clause","repos":"jasonmccampbell\/scipy-refactor,scipy\/scipy-svn,scipy\/scipy-svn,lesserwhirls\/scipy-cwt,scipy\/scipy-svn,jasonmccampbell\/scipy-refactor,lesserwhirls\/scipy-cwt,lesserwhirls\/scipy-cwt,jasonmccampbell\/scipy-refactor,scipy\/scipy-svn,lesserwhirls\/scipy-cwt,jasonmccampbell\/scipy-refactor"} {"commit":"1317d645092d94c95bcaf7a0341ac18208f9df0d","old_file":"patient\/admin.py","new_file":"patient\/admin.py","old_contents":"from django.contrib import admin\nfrom django.contrib.auth.models import Group\nfrom .models import Patient\nfrom django import forms\n\nclass CustomPatientForm(forms.ModelForm):\n class Meta:\n model = Patient\n\n def __init__(self, *args, **kwargs):\n super(CustomPatientForm, self).__init__(*args, **kwargs)\n self.fields['hub'].queryset = Group.objects.get(name=\"hubs\").user_set.all()\n self.fields['user'].queryset = Group.objects.get(name=\"patients\").user_set.all()\n\n\nclass PatientAdmin(admin.ModelAdmin):\n form = CustomPatientForm\n\nadmin.site.register(Patient, PatientAdmin)\n","new_contents":"from django.contrib import admin\nfrom django.contrib.auth.models import Group\nfrom .models import Patient\nfrom django import forms\n\nclass CustomPatientForm(forms.ModelForm):\n class Meta:\n model = Patient\n exclude = []\n\n def __init__(self, *args, **kwargs):\n super(CustomPatientForm, self).__init__(*args, **kwargs)\n self.fields['hub'].queryset = Group.objects.get(name=\"hubs\").user_set.all()\n self.fields['user'].queryset = Group.objects.get(name=\"patients\").user_set.all()\n\n\nclass PatientAdmin(admin.ModelAdmin):\n form = CustomPatientForm\n\nadmin.site.register(Patient, PatientAdmin)\n","subject":"Fix deprecated warning in CustomPatientForm","message":"Fix deprecated warning in CustomPatientForm\n","lang":"Python","license":"mit","repos":"sigurdsa\/angelika-api"} {"commit":"39461a97ef6e6b988466f41ddfee17687dd59ee1","old_file":"notifications\/match_score.py","new_file":"notifications\/match_score.py","old_contents":"from consts.notification_type import NotificationType\nfrom helpers.model_to_dict import ModelToDict\nfrom notifications.base_notification import BaseNotification\n\n\nclass MatchScoreNotification(BaseNotification):\n\n def __init__(self, match):\n self.match = match\n self.event = match.event.get()\n self._event_feed = self.event.key_name\n self._district_feed = self.event.event_district_enum\n\n @property\n def _type(self):\n return NotificationType.MATCH_SCORE\n\n def _build_dict(self):\n data = {}\n data['message_type'] = NotificationType.type_names[self._type]\n data['message_data'] = {}\n data['message_data']['event_name'] = self.event.name\n data['message_data']['match'] = ModelToDict.matchConverter(self.match)\n return data\n","new_contents":"from consts.notification_type import NotificationType\nfrom helpers.model_to_dict import ModelToDict\nfrom notifications.base_notification import BaseNotification\n\n\nclass MatchScoreNotification(BaseNotification):\n\n def __init__(self, match):\n self.match = match\n self.event = match.event.get()\n self._event_feed = self.event.key_name\n self._district_feed = self.event.event_district_enum\n\n @property\n def _type(self):\n return NotificationType.MATCH_SCORE\n\n def _build_dict(self):\n data = {}\n data['message_type'] = NotificationType.type_names[self._type]\n data['message_data'] = {}\n data['message_data']['event_name'] = self.event.name\n data['message_data']['event_key'] = self.event.key_name\n data['message_data']['match'] = ModelToDict.matchConverter(self.match)\n return data\n","subject":"Add event key to match score notification","message":"Add event key to match score notification","lang":"Python","license":"mit","repos":"bdaroz\/the-blue-alliance,tsteward\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,fangeugene\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,tsteward\/the-blue-alliance,bdaroz\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,bdaroz\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,bdaroz\/the-blue-alliance,verycumbersome\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,fangeugene\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,synth3tk\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,verycumbersome\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,verycumbersome\/the-blue-alliance,synth3tk\/the-blue-alliance,verycumbersome\/the-blue-alliance,nwalters512\/the-blue-alliance,fangeugene\/the-blue-alliance,nwalters512\/the-blue-alliance,tsteward\/the-blue-alliance,synth3tk\/the-blue-alliance,verycumbersome\/the-blue-alliance,synth3tk\/the-blue-alliance,tsteward\/the-blue-alliance,tsteward\/the-blue-alliance,synth3tk\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,nwalters512\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,fangeugene\/the-blue-alliance,synth3tk\/the-blue-alliance,bdaroz\/the-blue-alliance,fangeugene\/the-blue-alliance,bdaroz\/the-blue-alliance,fangeugene\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,tsteward\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,nwalters512\/the-blue-alliance,nwalters512\/the-blue-alliance,nwalters512\/the-blue-alliance,verycumbersome\/the-blue-alliance"} {"commit":"aca68618e5f1549faa1adfc1a1351a472ba0246d","old_file":"gettingstarted\/urls.py","new_file":"gettingstarted\/urls.py","old_contents":"from django.conf.urls import include, url\n\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\n# Examples:\n# url(r'^$', 'gettingstarted.views.home', name='home'),\n# url(r'^blog\/', include('blog.urls')),\n\nurlpatterns = [\n\n url(r'^$', include('rog.urls')),\n url(r'^admin\/', include(admin.site.urls)),\n\n]\n","new_contents":"from django.conf.urls import include, url\n\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\n# Examples:\n# url(r'^$', 'gettingstarted.views.home', name='home'),\n# url(r'^blog\/', include('blog.urls')),\n\nurlpatterns = [\n\n url(r'^', include('rog.urls')),\n url(r'^admin\/', include(admin.site.urls)),\n\n]\n","subject":"Remove the dollar from the regex to avoid problems including URLs.","message":"Remove the dollar from the regex to avoid problems including URLs.\n","lang":"Python","license":"bsd-2-clause","repos":"kenyansongithub\/django-rog,kenyansongithub\/django-rog,kenyansongithub\/django-rog"} {"commit":"6623d5679eef2c3db70edce7334582b5d524786d","old_file":"micro.py","new_file":"micro.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom sys import argv\nfrom operator import add, sub, mul, div\n\nfunctions = { \\\n\t'+': (2, add), \\\n\t'-': (2, sub), \\\n\t'*': (2, mul), \\\n\t'\/': (2, div) \\\n}\n\ndef get_code():\n\treturn argv[1]\n\ndef get_tokens(code):\n\treturn code.split(' ')\n\ndef evaluate(tokens):\n\tname = tokens[0]\n\ttokens = tokens[1:]\n\tif name not in functions:\n\t\treturn int(name), tokens\n\n\tfunction = functions[name]\n\targuments = []\n\tfor _ in xrange(function[0]):\n\t\tvalue, tokens = evaluate(tokens)\n\t\targuments.append(value)\n\n\tvalue = function[1](*arguments)\n\treturn value, tokens\n\nif __name__ == '__main__':\n\tcode = get_code()\n\ttokens = get_tokens(code)\n\tvalue, _ = evaluate(tokens)\n\tprint(value)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom sys import argv\nfrom operator import add, sub, mul, div\n\nfunctions = { \\\n\t'+': (2, add), \\\n\t'-': (2, sub), \\\n\t'*': (2, mul), \\\n\t'\/': (2, div) \\\n}\n\ndef get_code():\n\treturn argv[1]\n\ndef get_tokens(code):\n\treturn code.split(' ')\n\ndef parse_function(tokens):\n\treturn 'test', (23, None), tokens[-1:]\n\ndef evaluate(tokens):\n\tname = tokens[0]\n\ttokens = tokens[1:]\n\tif name == 'fn':\n\t\tname, function, tokens = parse_function(tokens)\n\t\tfunctions[name] = function\n\n\t\treturn 0, tokens\n\tif name not in functions:\n\t\treturn int(name), tokens\n\n\tfunction = functions[name]\n\targuments = []\n\tfor _ in xrange(function[0]):\n\t\tvalue, tokens = evaluate(tokens)\n\t\targuments.append(value)\n\n\tvalue = function[1](*arguments)\n\treturn value, tokens\n\nif __name__ == '__main__':\n\tcode = get_code()\n\ttokens = get_tokens(code)\n\tvalue, _ = evaluate(tokens)\n\tprint(value)\n\tprint(functions)\n","subject":"Add a detection of a function definition.","message":"Add a detection of a function definition.\n","lang":"Python","license":"mit","repos":"thewizardplusplus\/micro,thewizardplusplus\/micro,thewizardplusplus\/micro"} {"commit":"46afcd0e5e958e22647ef9c708918489027277e2","old_file":"modeltranslation\/tests\/settings.py","new_file":"modeltranslation\/tests\/settings.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nSettings overrided for test time\n\"\"\"\nimport os\nfrom django.conf import settings\n\n\nDIRNAME = os.path.dirname(__file__)\n\nINSTALLED_APPS = tuple(settings.INSTALLED_APPS) + (\n 'modeltranslation.tests',\n)\n# IMO this is unimportant\n#if django.VERSION[0] >= 1 and django.VERSION[1] >= 3:\n #INSTALLED_APPS += ('django.contrib.staticfiles',)\n\n#STATIC_URL = '\/static\/'\n\nMEDIA_URL = '\/media\/'\nMEDIA_ROOT = os.path.join(DIRNAME, 'media\/')\n\nLANGUAGES = (('de', 'Deutsch'),\n ('en', 'English'))\nLANGUAGE_CODE = 'de'\nMODELTRANSLATION_DEFAULT_LANGUAGE = 'de'\n\nUSE_I18N = True\n\nMODELTRANSLATION_AUTO_POPULATE = False\nMODELTRANSLATION_FALLBACK_LANGUAGES = ()\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nSettings overrided for test time\n\"\"\"\nimport os\nfrom django.conf import settings\n\n\nDIRNAME = os.path.dirname(__file__)\n\nINSTALLED_APPS = tuple(settings.INSTALLED_APPS) + (\n 'modeltranslation.tests',\n)\n# IMO this is unimportant\n#if django.VERSION[0] >= 1 and django.VERSION[1] >= 3:\n #INSTALLED_APPS += ('django.contrib.staticfiles',)\n\n#STATIC_URL = '\/static\/'\n\nMEDIA_URL = '\/media\/'\nMEDIA_ROOT = os.path.join(DIRNAME, 'media\/')\n\nLANGUAGES = (('de', 'Deutsch'),\n ('en', 'English'))\nLANGUAGE_CODE = 'de'\nMODELTRANSLATION_DEFAULT_LANGUAGE = 'de'\n\nUSE_I18N = True\nUSE_TZ = False\n\nMODELTRANSLATION_AUTO_POPULATE = False\nMODELTRANSLATION_FALLBACK_LANGUAGES = ()\n","subject":"Disable timezone support for tests, as the date \/ time fields' tests use naive datatime objects and fail if it's enabled.","message":"Disable timezone support for tests, as the date \/ time fields' tests use naive datatime objects and fail if it's enabled.\n","lang":"Python","license":"bsd-3-clause","repos":"extertioner\/django-modeltranslation,marctc\/django-modeltranslation,yoza\/django-modeltranslation,nanuxbe\/django-modeltranslation,akheron\/django-modeltranslation,vstoykov\/django-modeltranslation,SideStudios\/django-modeltranslation,yoza\/django-modeltranslation,marctc\/django-modeltranslation,deschler\/django-modeltranslation,akheron\/django-modeltranslation,nanuxbe\/django-modeltranslation,extertioner\/django-modeltranslation,deschler\/django-modeltranslation,SideStudios\/django-modeltranslation,vstoykov\/django-modeltranslation"} {"commit":"ea59df5387e15ad212b1344ee597013ffbb485d8","old_file":"foyer\/tests\/test_performance.py","new_file":"foyer\/tests\/test_performance.py","old_contents":"import mbuild as mb\nimport parmed as pmd\nimport pytest\n\nfrom foyer import Forcefield\nfrom foyer.tests.utils import get_fn\nfrom foyer.utils.io import has_mbuild\n\n\n@pytest.mark.timeout(1)\ndef test_fullerene():\n fullerene = pmd.load_file(get_fn('fullerene.pdb'), structure=True)\n forcefield = Forcefield(get_fn('fullerene.xml'))\n forcefield.apply(fullerene, assert_dihedral_params=False)\n\n\n@pytest.mark.skipif(not has_mbuild, reason=\"mbuild is not installed\")\n@pytest.mark.timeout(15)\ndef test_surface():\n surface = mb.load(get_fn('silica.mol2'))\n forcefield = Forcefield(get_fn('opls-silica.xml'))\n forcefield.apply(surface, assert_bond_params=False)\n\n\n@pytest.mark.skipif(not has_mbuild, reason=\"mbuild is not installed\")\n@pytest.mark.timeout(45)\ndef test_polymer():\n peg100 = mb.load(get_fn('peg100.mol2'))\n forcefield = Forcefield(name='oplsaa')\n forcefield.apply(peg100)\n","new_contents":"import mbuild as mb\nimport parmed as pmd\nimport pytest\n\nfrom foyer import Forcefield\nfrom foyer.tests.utils import get_fn\nfrom foyer.utils.io import has_mbuild\n\n\n@pytest.mark.timeout(1)\ndef test_fullerene():\n fullerene = pmd.load_file(get_fn('fullerene.pdb'), structure=True)\n forcefield = Forcefield(get_fn('fullerene.xml'))\n forcefield.apply(fullerene, assert_dihedral_params=False)\n\n\n@pytest.mark.skipif(not has_mbuild, reason=\"mbuild is not installed\")\n@pytest.mark.timeout(15)\ndef test_surface():\n surface = mb.load(get_fn('silica.mol2'))\n forcefield = Forcefield(get_fn('opls-silica.xml'))\n forcefield.apply(surface, assert_bond_params=False)\n\n\n@pytest.mark.skipif(not has_mbuild, reason=\"mbuild is not installed\")\n@pytest.mark.timeout(60)\ndef test_polymer():\n peg100 = mb.load(get_fn('peg100.mol2'))\n forcefield = Forcefield(name='oplsaa')\n forcefield.apply(peg100)\n","subject":"Bump timeout on polymer test to 60s","message":"Bump timeout on polymer test to 60s\n","lang":"Python","license":"mit","repos":"mosdef-hub\/foyer,iModels\/foyer,iModels\/foyer,mosdef-hub\/foyer"} {"commit":"f0b96aa0d4921e161eee9bb1a83846442f7f63b2","old_file":"likelihood.py","new_file":"likelihood.py","old_contents":"import math\n#Log-likelihood\ndef ll(ciphertext,perm,mat,k=2):\n s=0.0\n for i in range(len(ciphertext)-(k-1)):\n kmer = tuple([perm[c] for c in ciphertext[i:i+k]])\n s = s + math.log(mat[kmer])\n return s\n","new_contents":"import math\n#Log-likelihood\ndef ll(ciphertext,perm,mat):\n s=0.0\n for i in range(len(ciphertext)-(k-1)):\n kmer = tuple([perm[c] for c in ciphertext[i:i+k]])\n s = s + math.log(mat[kmer])\n return s\n","subject":"Remove default value for k - this should never be necessary for this function","message":"Remove default value for k - this should never be necessary for this function\n","lang":"Python","license":"mit","repos":"gputzel\/decode"} {"commit":"d4c67437b5c23a9c36fa1e2f79382e933c6c3297","old_file":"kolibri\/__init__.py","new_file":"kolibri\/__init__.py","old_contents":"from __future__ import absolute_import, print_function, unicode_literals\nfrom .utils.version import get_version\n\n#: This may not be the exact version as it's subject to modification with\n#: get_version() - use ``kolibri.__version__`` for the exact version string.\nVERSION = (0, 7, 0, 'final', 0)\n\n__author__ = 'Learning Equality'\n__email__ = 'info@learningequality.org'\n__version__ = str(get_version(VERSION))\n","new_contents":"from __future__ import absolute_import, print_function, unicode_literals\nfrom .utils.version import get_version\n\n#: This may not be the exact version as it's subject to modification with\n#: get_version() - use ``kolibri.__version__`` for the exact version string.\nVERSION = (0, 7, 1, 'alpha', 0)\n\n__author__ = 'Learning Equality'\n__email__ = 'info@learningequality.org'\n__version__ = str(get_version(VERSION))\n","subject":"Bump current branch to develop stage","message":"Bump current branch to develop stage\n","lang":"Python","license":"mit","repos":"christianmemije\/kolibri,lyw07\/kolibri,benjaoming\/kolibri,lyw07\/kolibri,learningequality\/kolibri,lyw07\/kolibri,learningequality\/kolibri,lyw07\/kolibri,mrpau\/kolibri,DXCanas\/kolibri,indirectlylit\/kolibri,learningequality\/kolibri,mrpau\/kolibri,indirectlylit\/kolibri,DXCanas\/kolibri,indirectlylit\/kolibri,benjaoming\/kolibri,christianmemije\/kolibri,jonboiser\/kolibri,mrpau\/kolibri,christianmemije\/kolibri,mrpau\/kolibri,learningequality\/kolibri,christianmemije\/kolibri,indirectlylit\/kolibri,DXCanas\/kolibri,benjaoming\/kolibri,jonboiser\/kolibri,jonboiser\/kolibri,benjaoming\/kolibri,DXCanas\/kolibri,jonboiser\/kolibri"} {"commit":"6b868e9bbb88dadbf27b7f1f6a4ab5fedc6c23e5","old_file":"foolbox\/tests\/test_attacks_carlini_wagner.py","new_file":"foolbox\/tests\/test_attacks_carlini_wagner.py","old_contents":"import numpy as np\n\nfrom foolbox.attacks import CarliniWagnerAttack as Attack\n\n\ndef test_targeted_attack(bn_targeted_adversarial):\n adv = bn_targeted_adversarial\n attack = Attack()\n attack(adv)\n assert adv.image is not None\n assert adv.distance.value < np.inf\n\n\ndef test_attack_impossible(bn_impossible):\n adv = bn_impossible\n attack = Attack()\n attack(adv)\n assert adv.image is None\n assert adv.distance.value == np.inf\n\n\ndef test_attack_gl(gl_bn_adversarial):\n adv = gl_bn_adversarial\n attack = Attack()\n attack(adv)\n assert adv.image is None\n assert adv.distance.value == np.inf\n","new_contents":"import numpy as np\n\nfrom foolbox.attacks import CarliniWagnerAttack as Attack\n\n\ndef test_targeted_attack(bn_targeted_adversarial):\n adv = bn_targeted_adversarial\n attack = Attack()\n attack(adv)\n assert adv.image is not None\n assert adv.distance.value < np.inf\n\n\ndef test_attack_impossible(bn_impossible):\n adv = bn_impossible\n attack = Attack()\n attack(adv)\n assert adv.image is None\n assert adv.distance.value == np.inf\n\n\ndef test_attack_gl(gl_bn_adversarial):\n adv = gl_bn_adversarial\n attack = Attack()\n attack(adv)\n assert adv.image is None\n assert adv.distance.value == np.inf","subject":"Revert \"added new line at end of file to please flake8\"","message":"Revert \"added new line at end of file to please flake8\"\n\nThis reverts commit 65fbfbc9117ba2ff06b4360faa8dddbf1ef8faa6.\n","lang":"Python","license":"mit","repos":"bethgelab\/foolbox,bethgelab\/foolbox"} {"commit":"9e27c8f803c42e65ec333ed1679ea70a5618f3c6","old_file":"dunya\/test_settings.py","new_file":"dunya\/test_settings.py","old_contents":"from settings import *\n\nif \"motif\" in DATABASES:\n del DATABASES[\"motif\"]\n\nTEST_RUNNER = \"xmlrunner.extra.djangotestrunner.XMLTestRunner\"\nTEST_OUTPUT_VERBOSE = True\nTEST_OUTPUT_DESCRIPTIONS = True\nTEST_OUTPUT_DIR = \"xmlrunner\"\n\n","new_contents":"from settings import *\n\nif \"motif\" in DATABASES:\n del DATABASES[\"motif\"]\n\nfrom xmlrunner.extra.djangotestrunner import XMLTestRunner\nfrom django.test.runner import DiscoverRunner\nfrom django.db import connections, DEFAULT_DB_ALIAS\n\n# We use the XMLTestRunner on CI\nclass DunyaTestRunner(XMLTestRunner):\n#class DunyaTestRunner(DiscoverRunner):\n def setup_databases(self):\n result = super(DunyaTestRunner, self).setup_databases()\n connection = connections[DEFAULT_DB_ALIAS]\n cursor = connection.cursor()\n cursor.execute('CREATE EXTENSION IF NOT EXISTS UNACCENT')\n return result\n\nTEST_RUNNER = \"dunya.test_settings.DunyaTestRunner\"\nTEST_OUTPUT_VERBOSE = True\nTEST_OUTPUT_DESCRIPTIONS = True\nTEST_OUTPUT_DIR = \"xmlrunner\"\n\n","subject":"Update test settings to create the unaccent ext if needed","message":"Update test settings to create the unaccent ext if needed\n","lang":"Python","license":"agpl-3.0","repos":"MTG\/dunya,MTG\/dunya,MTG\/dunya,MTG\/dunya"} {"commit":"ad415f26eec5c6a20c26123ccb6ce3e320ea9a69","old_file":"zou\/app\/blueprints\/crud\/asset_instance.py","new_file":"zou\/app\/blueprints\/crud\/asset_instance.py","old_contents":"from zou.app.models.asset_instance import AssetInstance\n\nfrom zou.app.services import assets_service, user_service\nfrom zou.app.utils import permissions\n\nfrom .base import BaseModelResource, BaseModelsResource\n\n\nclass AssetInstancesResource(BaseModelsResource):\n\n def __init__(self):\n BaseModelsResource.__init__(self, AssetInstance)\n\n\nclass AssetInstanceResource(BaseModelResource):\n\n def __init__(self):\n BaseModelResource.__init__(self, AssetInstance)\n\n def check_read_permissions(self, instance):\n if permissions.has_manager_permissions():\n return True\n else:\n asset_instance = self.get_model_or_404(instance[\"id\"])\n asset = assets_service.get_asset(asset_instance.asset_id)\n return user_service.check_has_task_related(asset[\"project_id\"])\n\n def check_update_permissions(self, asset_instance, data):\n if permissions.has_manager_permissions():\n return True\n else:\n return user_service.check_working_on_entity(\n asset_instance[\"entity_id\"]\n )\n","new_contents":"from zou.app.models.asset_instance import AssetInstance\n\nfrom zou.app.services import assets_service, user_service\nfrom zou.app.utils import permissions\n\nfrom .base import BaseModelResource, BaseModelsResource\n\n\nclass AssetInstancesResource(BaseModelsResource):\n\n def __init__(self):\n BaseModelsResource.__init__(self, AssetInstance)\n\n\nclass AssetInstanceResource(BaseModelResource):\n\n def __init__(self):\n BaseModelResource.__init__(self, AssetInstance)\n self.protected_fields.append([\"number\"])\n\n def check_read_permissions(self, instance):\n if permissions.has_manager_permissions():\n return True\n else:\n asset_instance = self.get_model_or_404(instance[\"id\"])\n asset = assets_service.get_asset(asset_instance.asset_id)\n return user_service.check_has_task_related(asset[\"project_id\"])\n\n def check_update_permissions(self, asset_instance, data):\n if permissions.has_manager_permissions():\n return True\n else:\n asset = assets_service.get_asset(asset_instance[\"asset_id\"])\n return user_service.check_has_task_related(asset[\"project_id\"])\n","subject":"Change asset instance update permissions","message":"Change asset instance update permissions\n\n* Do not allow to change instance number\n* Allow to change instance name by a CG artist\n","lang":"Python","license":"agpl-3.0","repos":"cgwire\/zou"} {"commit":"1cdcc3ada4096a8352acc7f7c3e7825a7f44e0ac","old_file":"examples\/structures.py","new_file":"examples\/structures.py","old_contents":"from numba import struct, jit, double\nimport numpy as np\n\nrecord_type = struct([('x', double), ('y', double)])\nrecord_dtype = record_type.get_dtype()\na = np.array([(1.0, 2.0), (3.0, 4.0)], dtype=record_dtype)\n\n@jit(argtypes=[record_type[:]])\ndef hypot(data):\n # return types of numpy functions are inferred\n result = np.empty_like(data, dtype=np.float64) \n # notice access to structure elements 'x' and 'y' via attribute access\n # You can also index by field name or field index:\n # data[i].x == data[i]['x'] == data[i][0]\n for i in range(data.shape[0]):\n result[i] = np.sqrt(data[i].x * data[i].x + data[i].y * data[i].y)\n\n return result\n\nprint hypot(a)\n\n# Notice inferred return type\nprint hypot.signature\n# Notice native sqrt calls and for.body direct access to memory...\nprint hypot.lfunc\n","new_contents":"from numba import struct, jit, double\nimport numpy as np\n\nrecord_type = struct([('x', double), ('y', double)])\nrecord_dtype = record_type.get_dtype()\na = np.array([(1.0, 2.0), (3.0, 4.0)], dtype=record_dtype)\n\n@jit(argtypes=[record_type[:]])\ndef hypot(data):\n # return types of numpy functions are inferred\n result = np.empty_like(data, dtype=np.float64) \n # notice access to structure elements 'x' and 'y' via attribute access\n # You can also index by field name or field index:\n # data[i].x == data[i]['x'] == data[i][0]\n for i in range(data.shape[0]):\n result[i] = np.sqrt(data[i].x * data[i].x + data[i].y * data[i].y)\n\n return result\n\nprint hypot(a)\n\n# Notice inferred return type\nprint hypot.signature\n# Notice native sqrt calls and for.body direct access to memory...\n#print hypot.lfunc\n","subject":"Comment out printing of function.","message":"Comment out printing of function.\n","lang":"Python","license":"bsd-2-clause","repos":"numba\/numba,seibert\/numba,stonebig\/numba,jriehl\/numba,sklam\/numba,GaZ3ll3\/numba,ssarangi\/numba,shiquanwang\/numba,cpcloud\/numba,stuartarchibald\/numba,pitrou\/numba,GaZ3ll3\/numba,pitrou\/numba,cpcloud\/numba,IntelLabs\/numba,pitrou\/numba,sklam\/numba,cpcloud\/numba,gdementen\/numba,seibert\/numba,jriehl\/numba,gmarkall\/numba,seibert\/numba,gmarkall\/numba,stonebig\/numba,cpcloud\/numba,gdementen\/numba,stonebig\/numba,stuartarchibald\/numba,stuartarchibald\/numba,cpcloud\/numba,pombredanne\/numba,stonebig\/numba,stefanseefeld\/numba,stefanseefeld\/numba,stuartarchibald\/numba,numba\/numba,GaZ3ll3\/numba,stefanseefeld\/numba,pitrou\/numba,GaZ3ll3\/numba,gdementen\/numba,stefanseefeld\/numba,jriehl\/numba,gmarkall\/numba,jriehl\/numba,numba\/numba,shiquanwang\/numba,pitrou\/numba,IntelLabs\/numba,ssarangi\/numba,gdementen\/numba,sklam\/numba,stonebig\/numba,ssarangi\/numba,IntelLabs\/numba,pombredanne\/numba,sklam\/numba,gmarkall\/numba,gmarkall\/numba,pombredanne\/numba,seibert\/numba,stefanseefeld\/numba,GaZ3ll3\/numba,gdementen\/numba,pombredanne\/numba,stuartarchibald\/numba,seibert\/numba,numba\/numba,ssarangi\/numba,IntelLabs\/numba,numba\/numba,IntelLabs\/numba,shiquanwang\/numba,pombredanne\/numba,ssarangi\/numba,sklam\/numba,jriehl\/numba"} {"commit":"b78aebed4771015e6292638ac1980e3acaed4db9","old_file":"heufybot\/connection.py","new_file":"heufybot\/connection.py","old_contents":"from twisted.words.protocols import irc\n\n\nclass HeufyBotConnection(irc.IRC):\n def __init__(self, protocol):\n self.protocol = protocol\n self.nick = \"PyHeufyBot\" # TODO This will be set by a configuration at some point\n self.ident = \"PyHeufyBot\" # TODO This will be set by a configuration at some point\n self.gecos = \"PyHeufyBot IRC Bot\" # TODO This will be set by a configuration at some point\n\n def connectionMade(self):\n self.cmdNICK(self.nick)\n self.cmdUSER(self.ident, self.gecos)\n\n def connectionLost(self, reason=\"\"):\n print reason\n\n def dataReceived(self, data):\n print data\n\n def sendMessage(self, command, *parameter_list, **prefix):\n print command, \" \".join(parameter_list)\n irc.IRC.sendMessage(self, command, *parameter_list, **prefix)\n\n def cmdNICK(self, nick):\n self.sendMessage(\"NICK\", nick)\n\n def cmdUSER(self, ident, gecos):\n # RFC2812 allows usermodes to be set, but this isn't implemented much in IRCds at all.\n # Pass 0 for usermodes instead.\n self.sendMessage(\"USER\", ident, \"0\", \"*\", \":{}\".format(gecos))\n","new_contents":"from twisted.words.protocols import irc\n\n\nclass HeufyBotConnection(irc.IRC):\n def __init__(self, protocol):\n self.protocol = protocol\n self.nick = \"PyHeufyBot\" # TODO This will be set by a configuration at some point\n self.ident = \"PyHeufyBot\" # TODO This will be set by a configuration at some point\n self.gecos = \"PyHeufyBot IRC Bot\" # TODO This will be set by a configuration at some point\n self.channels = {}\n self.usermodes = {}\n\n def connectionMade(self):\n self.cmdNICK(self.nick)\n self.cmdUSER(self.ident, self.gecos)\n\n def connectionLost(self, reason=\"\"):\n print reason\n\n def dataReceived(self, data):\n print data\n\n def sendMessage(self, command, *parameter_list, **prefix):\n print command, \" \".join(parameter_list)\n irc.IRC.sendMessage(self, command, *parameter_list, **prefix)\n\n def cmdNICK(self, nick):\n self.sendMessage(\"NICK\", nick)\n\n def cmdUSER(self, ident, gecos):\n # RFC2812 allows usermodes to be set, but this isn't implemented much in IRCds at all.\n # Pass 0 for usermodes instead.\n self.sendMessage(\"USER\", ident, \"0\", \"*\", \":{}\".format(gecos))\n","subject":"Add dictionaries for channels and usermodes","message":"Add dictionaries for channels and usermodes\n","lang":"Python","license":"mit","repos":"Heufneutje\/PyHeufyBot,Heufneutje\/PyHeufyBot"} {"commit":"2afe09bcbcc728e98ec8da39b68ea65f4c270fdb","old_file":"html5lib\/trie\/_base.py","new_file":"html5lib\/trie\/_base.py","old_contents":"from __future__ import absolute_import, division, unicode_literals\n\nfrom collections import Mapping\n\n\nclass Trie(Mapping):\n \"\"\"Abstract base class for tries\"\"\"\n\n def keys(self, prefix=None):\n keys = super().keys()\n\n if prefix is None:\n return set(keys)\n\n # Python 2.6: no set comprehensions\n return set([x for x in keys if x.startswith(prefix)])\n\n def has_keys_with_prefix(self, prefix):\n for key in self.keys():\n if key.startswith(prefix):\n return True\n\n return False\n\n def longest_prefix(self, prefix):\n if prefix in self:\n return prefix\n\n for i in range(1, len(prefix) + 1):\n if prefix[:-i] in self:\n return prefix[:-i]\n\n raise KeyError(prefix)\n\n def longest_prefix_item(self, prefix):\n lprefix = self.longest_prefix(prefix)\n return (lprefix, self[lprefix])\n","new_contents":"from __future__ import absolute_import, division, unicode_literals\n\nfrom collections import Mapping\n\n\nclass Trie(Mapping):\n \"\"\"Abstract base class for tries\"\"\"\n\n def keys(self, prefix=None):\n keys = super(Trie, self).keys()\n\n if prefix is None:\n return set(keys)\n\n # Python 2.6: no set comprehensions\n return set([x for x in keys if x.startswith(prefix)])\n\n def has_keys_with_prefix(self, prefix):\n for key in self.keys():\n if key.startswith(prefix):\n return True\n\n return False\n\n def longest_prefix(self, prefix):\n if prefix in self:\n return prefix\n\n for i in range(1, len(prefix) + 1):\n if prefix[:-i] in self:\n return prefix[:-i]\n\n raise KeyError(prefix)\n\n def longest_prefix_item(self, prefix):\n lprefix = self.longest_prefix(prefix)\n return (lprefix, self[lprefix])\n","subject":"Make this in practice unreachable code work on Py2","message":"Make this in practice unreachable code work on Py2\n","lang":"Python","license":"mit","repos":"html5lib\/html5lib-python,html5lib\/html5lib-python,html5lib\/html5lib-python"} {"commit":"2de06cda19c2d50c1362c9babd7c4bce735fb44a","old_file":"product_configurator_mrp\/__manifest__.py","new_file":"product_configurator_mrp\/__manifest__.py","old_contents":"{\n 'name': 'Product Configurator Manufacturing',\n 'version': '11.0.1.0.0',\n 'category': 'Manufacturing',\n 'summary': 'BOM Support for configurable products',\n 'author': 'Pledra',\n 'license': 'AGPL-3',\n 'website': 'http:\/\/www.pledra.com\/',\n 'depends': ['mrp', 'product_configurator'],\n \"data\": [\n 'security\/configurator_security.xml',\n 'security\/ir.model.access.csv',\n 'views\/product_view.xml',\n 'views\/product_config_view.xml',\n 'views\/mrp_view.xml',\n ],\n 'demo': [],\n 'test': [],\n 'installable': True,\n 'auto_install': False,\n}\n","new_contents":"{\n 'name': 'Product Configurator Manufacturing',\n 'version': '11.0.1.0.0',\n 'category': 'Manufacturing',\n 'summary': 'BOM Support for configurable products',\n 'author': 'Pledra',\n 'license': 'AGPL-3',\n 'website': 'http:\/\/www.pledra.com\/',\n 'depends': ['mrp', 'product_configurator'],\n \"data\": [\n 'security\/configurator_security.xml',\n 'security\/ir.model.access.csv',\n 'views\/product_view.xml',\n 'views\/product_config_view.xml',\n 'views\/mrp_view.xml',\n ],\n 'demo': [],\n 'test': [],\n 'installable': False,\n 'auto_install': False,\n}\n","subject":"Set product_configurator_mrp to uninstallable until fixing \/ refactoring","message":"Set product_configurator_mrp to uninstallable until fixing \/ refactoring\n","lang":"Python","license":"agpl-3.0","repos":"pledra\/odoo-product-configurator,pledra\/odoo-product-configurator,pledra\/odoo-product-configurator"} {"commit":"5c5b81312317c1750ea320666b2adc4f74d13366","old_file":"f8a_jobs\/graph_sync.py","new_file":"f8a_jobs\/graph_sync.py","old_contents":"\"\"\"Functions to retrieve pending list and invoke Graph Sync.\"\"\"\n\nimport f8a_jobs.defaults as configuration\nimport requests\nimport traceback\nimport logging\nfrom urllib.parse import urljoin\n\n\nlogger = logging.getLogger(__name__)\n\n\ndef _api_call(url, params=None):\n params = params or {}\n try:\n logger.info(\"API Call for url: %s, params: %s\" % (url, params))\n r = requests.get(url, params=params)\n r.raise_for_status()\n result = {\"data\": r.json()}\n except requests.exceptions.HTTPError:\n logger.error(traceback.format_exc())\n result = {\"error\": \"Failed to retrieve data from Data Model Importer backend\"}\n return result\n\n\ndef fetch_pending(params=None):\n params = params or {}\n \"\"\"Invoke Pending Graph Sync APIs for given parameters.\"\"\"\n url = urljoin(configuration.DATA_IMPORTER_ENDPOINT, \"\/api\/v1\/pending\")\n return _api_call(url, params)\n\n\ndef invoke_sync(params=None):\n params = params or {}\n \"\"\"Invoke Graph Sync APIs to sync for given parameters.\"\"\"\n url = urljoin(configuration.DATA_IMPORTER_ENDPOINT, \"\/api\/v1\/sync_all\")\n return _api_call(url, params)\n","new_contents":"\"\"\"Functions to retrieve pending list and invoke Graph Sync.\"\"\"\n\nimport f8a_jobs.defaults as configuration\nimport requests\nimport traceback\nimport logging\nfrom urllib.parse import urljoin\n\n\nlogger = logging.getLogger(__name__)\n\n\ndef _api_call(url, params=None):\n params = params or {}\n try:\n logger.info(\"API Call for url: %s, params: %s\" % (url, params))\n r = requests.get(url, params=params)\n r.raise_for_status()\n result = {\"data\": r.json()}\n except requests.exceptions.HTTPError:\n logger.error(traceback.format_exc())\n result = {\"error\": \"Failed to retrieve data from Data Model Importer backend\"}\n return result\n\n\ndef fetch_pending(params=None):\n \"\"\"Invoke Pending Graph Sync APIs for given parameters.\"\"\"\n params = params or {}\n url = urljoin(configuration.DATA_IMPORTER_ENDPOINT, \"\/api\/v1\/pending\")\n return _api_call(url, params)\n\n\ndef invoke_sync(params=None):\n \"\"\"Invoke Graph Sync APIs to sync for given parameters.\"\"\"\n params = params or {}\n url = urljoin(configuration.DATA_IMPORTER_ENDPOINT, \"\/api\/v1\/sync_all\")\n return _api_call(url, params)\n","subject":"Fix code for review comments","message":"Fix code for review comments\n","lang":"Python","license":"apache-2.0","repos":"fabric8-analytics\/fabric8-analytics-jobs,fabric8-analytics\/fabric8-analytics-jobs"} {"commit":"270a03dc78838137051fec49050f550c44be2359","old_file":"facebook_auth\/views.py","new_file":"facebook_auth\/views.py","old_contents":"import logging\n\nfrom django.contrib.auth import authenticate\nfrom django.contrib.auth import login\nfrom django import http\nfrom django.views import generic\nfrom facebook_auth import urls\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Handler(generic.View):\n def get(self, request):\n try:\n next_url = urls.Next().decode(request.GET['next'])\n except urls.InvalidNextUrl:\n logger.warning('Invalid facebook handler next.',\n extra={'request': request})\n return http.HttpResponseBadRequest()\n if 'code' in request.GET:\n self.login(next_url)\n response = http.HttpResponseRedirect(next_url['next'])\n response[\"P3P\"] = ('CP=\"IDC DSP COR ADM DEVi TAIi PSA PSD IVAi'\n ' IVDi CONi HIS OUR IND CNT\"')\n else:\n response = http.HttpResponseRedirect(next_url['close'])\n return response\n\n def login(self, next_url):\n user = authenticate(\n code=self.request.GET['code'],\n redirect_uri=urls.redirect_uri(next_url['next'],\n next_url['close']))\n if user:\n login(self.request, user)\n\nhandler = Handler.as_view()\n","new_contents":"import logging\n\nfrom django.contrib.auth import authenticate\nfrom django.contrib.auth import login\nfrom django import http\nfrom django.views import generic\n\nimport facepy\n\nfrom facebook_auth import urls\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass Handler(generic.View):\n def get(self, request):\n try:\n self.next_url = urls.Next().decode(request.GET['next'])\n except urls.InvalidNextUrl:\n logger.warning('Invalid facebook handler next.',\n extra={'request': request})\n return http.HttpResponseBadRequest()\n if 'code' in request.GET:\n try:\n self.login()\n except facepy.FacepyError as e:\n return self.handle_facebook_error(e)\n response = http.HttpResponseRedirect(self.next_url['next'])\n response[\"P3P\"] = ('CP=\"IDC DSP COR ADM DEVi TAIi PSA PSD IVAi'\n ' IVDi CONi HIS OUR IND CNT\"')\n else:\n response = http.HttpResponseRedirect(self.next_url['close'])\n return response\n\n def login(self):\n user = authenticate(\n code=self.request.GET['code'],\n redirect_uri=urls.redirect_uri(self.next_url['next'],\n self.next_url['close']))\n if user:\n login(self.request, user)\n\n def handle_facebook_error(self, e):\n return http.HttpResponseRedirect(self.next_url['next'])\n\nhandler = Handler.as_view()\n","subject":"Add facebook error handler in view.","message":"Add facebook error handler in view.\n\nThis assumes that there is no other backend which\ncan authenticate user with facebook credentials.\n","lang":"Python","license":"mit","repos":"jgoclawski\/django-facebook-auth,pozytywnie\/django-facebook-auth,pozytywnie\/django-facebook-auth,jgoclawski\/django-facebook-auth"} {"commit":"088d76bbe01a9a6dd0a246be8ce703d5b64c540e","old_file":"Lib\/hotshot\/__init__.py","new_file":"Lib\/hotshot\/__init__.py","old_contents":"\"\"\"High-perfomance logging profiler, mostly written in C.\"\"\"\n\nimport _hotshot\n\nfrom _hotshot import ProfilerError\n\n\nclass Profile:\n def __init__(self, logfn, lineevents=0, linetimings=1):\n self.lineevents = lineevents and 1 or 0\n self.linetimings = (linetimings and lineevents) and 1 or 0\n self._prof = p = _hotshot.profiler(\n logfn, self.lineevents, self.linetimings)\n\n def close(self):\n self._prof.close()\n\n def start(self):\n self._prof.start()\n\n def stop(self):\n self._prof.stop()\n\n # These methods offer the same interface as the profile.Profile class,\n # but delegate most of the work to the C implementation underneath.\n\n def run(self, cmd):\n import __main__\n dict = __main__.__dict__\n return self.runctx(cmd, dict, dict)\n\n def runctx(self, cmd, globals, locals):\n code = compile(cmd, \"<string>\", \"exec\")\n self._prof.runcode(code, globals, locals)\n return self\n\n def runcall(self, func, *args, **kw):\n return self._prof.runcall(func, args, kw)\n","new_contents":"\"\"\"High-perfomance logging profiler, mostly written in C.\"\"\"\n\nimport _hotshot\n\nfrom _hotshot import ProfilerError\n\n\nclass Profile:\n def __init__(self, logfn, lineevents=0, linetimings=1):\n self.lineevents = lineevents and 1 or 0\n self.linetimings = (linetimings and lineevents) and 1 or 0\n self._prof = p = _hotshot.profiler(\n logfn, self.lineevents, self.linetimings)\n\n def close(self):\n self._prof.close()\n\n def start(self):\n self._prof.start()\n\n def stop(self):\n self._prof.stop()\n\n def addinfo(self, key, value):\n self._prof.addinfo(key, value)\n\n # These methods offer the same interface as the profile.Profile class,\n # but delegate most of the work to the C implementation underneath.\n\n def run(self, cmd):\n import __main__\n dict = __main__.__dict__\n return self.runctx(cmd, dict, dict)\n\n def runctx(self, cmd, globals, locals):\n code = compile(cmd, \"<string>\", \"exec\")\n self._prof.runcode(code, globals, locals)\n return self\n\n def runcall(self, func, *args, **kw):\n return self._prof.runcall(func, args, kw)\n","subject":"Allow user code to call the addinfo() method on the profiler object.","message":"Allow user code to call the addinfo() method on the profiler object.\n","lang":"Python","license":"mit","repos":"sk-\/python2.7-type-annotator,sk-\/python2.7-type-annotator,sk-\/python2.7-type-annotator"} {"commit":"d89715196ba79da02a997688414dfa283bee5aeb","old_file":"profiles\/tests\/test_views.py","new_file":"profiles\/tests\/test_views.py","old_contents":"from django.core.urlresolvers import reverse\nfrom django.test import TestCase\nfrom django.test.client import RequestFactory\nfrom utils.factories import UserFactory\n\nfrom profiles.views import ProfileView\n\n\nclass ProfileViewTests(TestCase):\n\n def setUp(self):\n request_factory = RequestFactory()\n request = request_factory.get(reverse('profile'))\n request.user = UserFactory.create()\n self.response = ProfileView.as_view()(request)\n\n def test_profile_view_200(self):\n self.assertEqual(self.response.status_code, 200)\n\n def test_profile_view_renders(self):\n self.response.render()\n","new_contents":"from django.core.urlresolvers import reverse\nfrom django.test import TestCase\nfrom django.test.client import RequestFactory\nfrom utils.factories import UserFactory\n\nfrom profiles.views import (\n ProfileView,\n ReviewUserView,\n)\n\n\nclass ProfileViewTests(TestCase):\n\n def setUp(self):\n request_factory = RequestFactory()\n request = request_factory.get(reverse('profile'))\n request.user = UserFactory.create()\n self.response = ProfileView.as_view()(request)\n\n def test_profile_view_200(self):\n self.assertEqual(self.response.status_code, 200)\n\n def test_profile_view_renders(self):\n self.response.render()\n\n\nclass ReviewUserViewTests(TestCase):\n\n def setUp(self):\n request_factory = RequestFactory()\n self.request = request_factory.get('\/admin\/dashboard\/')\n\n def test_review_user_view_200(self):\n user = UserFactory.create()\n user.is_staff = True\n self.request.user = user\n response = ReviewUserView.as_view()(self.request)\n self.assertEqual(response.status_code, 200)\n\n def test_review_user_view_200(self):\n user = UserFactory.create()\n user.is_staff = True\n self.request.user = user\n response = ReviewUserView.as_view()(self.request)\n response.render()\n\n def test_review_user_view_not_staff(self):\n user = UserFactory.create()\n self.request.user = user\n response = ReviewUserView.as_view()(self.request)\n self.assertEqual(response.status_code, 302)\n","subject":"Add tests for user review view","message":"Add tests for user review view\n","lang":"Python","license":"mit","repos":"phildini\/logtacts,phildini\/logtacts,phildini\/logtacts,phildini\/logtacts,phildini\/logtacts"} {"commit":"f2a46687e24d82060b687922de3495111f82e558","old_file":"geist\/backends\/fake.py","new_file":"geist\/backends\/fake.py","old_contents":"import numpy as np\nfrom ..core import Location\n\n\nclass GeistFakeBackend(object):\n def __init__(self, w=800, h=600):\n self.image = np.zeros((h, w, 3))\n self.locations = [Location(0, 0, w=w, h=h, image=self.image)]\n\n def create_process(self, command):\n pass\n\n def actions_transaction(self):\n pass\n\n def capture_locations(self):\n for loc in self.locations:\n yield loc\n\n def key_down(self, name):\n pass\n\n def key_up(self, name):\n pass\n\n def button_down(self, button_num):\n pass\n\n def button_up(self, button_num):\n pass\n\n def move(self, point):\n pass\n\n def close(self):\n pass\n","new_contents":"import numpy as np\nfrom ..core import Location\n\n\nclass GeistFakeBackend(object):\n def __init__(self, image=None, w=800, h=600):\n if image is None:\n self.image = np.zeros((h, w, 3))\n self.locations = [Location(0, 0, w=w, h=h, image=self.image)]\n else:\n if isinstance(image, basestring):\n image = np.load(image)\n self.image = image\n h, w, _ = image.shape\n self.locations = [Location(0, 0, w=w, h=h, image=self.image)]\n\n def create_process(self, command):\n pass\n\n def actions_transaction(self):\n pass\n\n def capture_locations(self):\n for loc in self.locations:\n yield loc\n\n def key_down(self, name):\n pass\n\n def key_up(self, name):\n pass\n\n def button_down(self, button_num):\n pass\n\n def button_up(self, button_num):\n pass\n\n def move(self, point):\n pass\n\n def close(self):\n pass\n","subject":"Allow Fake Backend to take an image as the screen","message":"Allow Fake Backend to take an image as the screen\n","lang":"Python","license":"mit","repos":"kebarr\/Geist,thetestpeople\/Geist"} {"commit":"aded1c825385817dc39d8ff99c169e6620008abf","old_file":"blivet\/populator\/helpers\/__init__.py","new_file":"blivet\/populator\/helpers\/__init__.py","old_contents":"from .btrfs import BTRFSFormatPopulator\nfrom .boot import AppleBootFormatPopulator, EFIFormatPopulator, MacEFIFormatPopulator\nfrom .disk import DiskDevicePopulator\nfrom .disklabel import DiskLabelFormatPopulator\nfrom .dm import DMDevicePopulator\nfrom .dmraid import DMRaidFormatPopulator\nfrom .loop import LoopDevicePopulator\nfrom .luks import LUKSFormatPopulator\nfrom .lvm import LVMDevicePopulator, LVMFormatPopulator\nfrom .mdraid import MDDevicePopulator, MDFormatPopulator\nfrom .multipath import MultipathDevicePopulator\nfrom .optical import OpticalDevicePopulator\nfrom .partition import PartitionDevicePopulator\n","new_contents":"import inspect as _inspect\n\nfrom .devicepopulator import DevicePopulator\nfrom .formatpopulator import FormatPopulator\n\nfrom .btrfs import BTRFSFormatPopulator\nfrom .boot import AppleBootFormatPopulator, EFIFormatPopulator, MacEFIFormatPopulator\nfrom .disk import DiskDevicePopulator\nfrom .disklabel import DiskLabelFormatPopulator\nfrom .dm import DMDevicePopulator\nfrom .dmraid import DMRaidFormatPopulator\nfrom .loop import LoopDevicePopulator\nfrom .luks import LUKSFormatPopulator\nfrom .lvm import LVMDevicePopulator, LVMFormatPopulator\nfrom .mdraid import MDDevicePopulator, MDFormatPopulator\nfrom .multipath import MultipathDevicePopulator\nfrom .optical import OpticalDevicePopulator\nfrom .partition import PartitionDevicePopulator\n\n__all__ = [\"get_device_helper\", \"get_format_helper\"]\n\n_device_helpers = []\n_format_helpers = []\n\n\ndef _build_helper_lists():\n \"\"\"Build lists of known device and format helper classes.\"\"\"\n global _device_helpers # pylint: disable=global-variable-undefined\n global _format_helpers # pylint: disable=global-variable-undefined\n _device_helpers = []\n _format_helpers = []\n for obj in globals().values():\n if not _inspect.isclass(obj):\n continue\n elif issubclass(obj, DevicePopulator):\n _device_helpers.append(obj)\n elif issubclass(obj, FormatPopulator):\n _format_helpers.append(obj)\n\n _device_helpers.sort(key=lambda h: h.priority, reverse=True)\n _format_helpers.sort(key=lambda h: h.priority, reverse=True)\n\n_build_helper_lists()\n\n\ndef get_device_helper(data):\n \"\"\" Return the device helper class appropriate for the specified data.\n\n The helper lists are sorted according to priorities defined within each\n class. This function returns the first matching class.\n \"\"\"\n return next((h for h in _device_helpers if h.match(data)), None)\n\n\ndef get_format_helper(data, device):\n \"\"\" Return the device helper class appropriate for the specified data.\n\n The helper lists are sorted according to priorities defined within each\n class. This function returns the first matching class.\n \"\"\"\n return next((h for h in _format_helpers if h.match(data, device=device)), None)\n","subject":"Add functions to return a helper class based on device data.","message":"Add functions to return a helper class based on device data.\n\nThis is intended to be the complete API of populator.helpers.\n","lang":"Python","license":"lgpl-2.1","repos":"vojtechtrefny\/blivet,jkonecny12\/blivet,vojtechtrefny\/blivet,rhinstaller\/blivet,rhinstaller\/blivet,rvykydal\/blivet,AdamWill\/blivet,rvykydal\/blivet,vpodzime\/blivet,AdamWill\/blivet,jkonecny12\/blivet,vpodzime\/blivet"} {"commit":"d6759d0abec637753d93cd407fad5e7abc6ec86d","old_file":"astropy\/tests\/plugins\/display.py","new_file":"astropy\/tests\/plugins\/display.py","old_contents":"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\n# This plugin now lives in pytest-astropy, but we keep the code below during\n# a deprecation phase.\n\nimport warnings\nfrom astropy.utils.exceptions import AstropyDeprecationWarning\n\ntry:\n from pytest_astropy_header.display import PYTEST_HEADER_MODULES, TESTED_VERSIONS\nexcept ImportError:\n PYTEST_HEADER_MODULES = {}\n TESTED_VERSIONS = {}\n\nwarnings.warn('The astropy.tests.plugins.display plugin has been deprecated. '\n 'See the pytest-astropy documentation for information on '\n 'migrating to using pytest-astropy to customize the pytest '\n 'header.', AstropyDeprecationWarning)\n","new_contents":"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\n# This plugin now lives in pytest-astropy, but we keep the code below during\n# a deprecation phase.\n\nimport warnings\nfrom astropy.utils.exceptions import AstropyDeprecationWarning\n\ntry:\n from pytest_astropy_header.display import (PYTEST_HEADER_MODULES,\n TESTED_VERSIONS)\nexcept ImportError:\n PYTEST_HEADER_MODULES = {}\n TESTED_VERSIONS = {}\n\nwarnings.warn('The astropy.tests.plugins.display plugin has been deprecated. '\n 'See the pytest-astropy-header documentation for information on '\n 'migrating to using pytest-astropy-header to customize the '\n 'pytest header.', AstropyDeprecationWarning)\n","subject":"Fix typo in deprecation warning [ci skip]","message":"TST: Fix typo in deprecation warning [ci skip]\n","lang":"Python","license":"bsd-3-clause","repos":"stargaser\/astropy,dhomeier\/astropy,saimn\/astropy,saimn\/astropy,larrybradley\/astropy,astropy\/astropy,StuartLittlefair\/astropy,lpsinger\/astropy,dhomeier\/astropy,lpsinger\/astropy,StuartLittlefair\/astropy,larrybradley\/astropy,lpsinger\/astropy,MSeifert04\/astropy,astropy\/astropy,astropy\/astropy,MSeifert04\/astropy,larrybradley\/astropy,larrybradley\/astropy,StuartLittlefair\/astropy,StuartLittlefair\/astropy,mhvk\/astropy,StuartLittlefair\/astropy,saimn\/astropy,aleksandr-bakanov\/astropy,dhomeier\/astropy,mhvk\/astropy,astropy\/astropy,astropy\/astropy,mhvk\/astropy,aleksandr-bakanov\/astropy,MSeifert04\/astropy,mhvk\/astropy,lpsinger\/astropy,dhomeier\/astropy,stargaser\/astropy,pllim\/astropy,dhomeier\/astropy,saimn\/astropy,saimn\/astropy,pllim\/astropy,larrybradley\/astropy,pllim\/astropy,pllim\/astropy,aleksandr-bakanov\/astropy,lpsinger\/astropy,mhvk\/astropy,pllim\/astropy,aleksandr-bakanov\/astropy,stargaser\/astropy,MSeifert04\/astropy,stargaser\/astropy"} {"commit":"e3bac9c0a655ae49d6e15b16894712f4edbc994b","old_file":"campus02\/web\/views\/primarykey.py","new_file":"campus02\/web\/views\/primarykey.py","old_contents":"#!\/usr\/bin\/python\n# -*- coding: utf-8 -*-\n\nfrom rest_framework import viewsets\nfrom rest_framework.filters import DjangoObjectPermissionsFilter\n\nfrom .. import (\n models,\n permissions,\n)\nfrom ..serializers import (\n primarykey as serializers\n)\n\n\nclass MovieViewSet(viewsets.ReadOnlyModelViewSet):\n queryset = models.Movie.objects.all()\n serializer_class = serializers.MovieSerializer\n\n\nclass GenreViewSet(viewsets.ReadOnlyModelViewSet):\n queryset = models.Genre.objects.all()\n serializer_class = serializers.GenreSerializer\n\n\nclass WatchlistViewSet(viewsets.ModelViewSet):\n queryset = models.Watchlist.objects.all()\n serializer_class = serializers.WatchlistSerializer\n filter_backends = [\n DjangoObjectPermissionsFilter,\n ]\n permission_classes = [\n permissions.DjangoObjectPermissions,\n ]\n\n\nclass ResumeViewSet(viewsets.ModelViewSet):\n queryset = models.Resume.objects.all()\n serializer_class = serializers.ResumeSerializer\n filter_backends = [\n DjangoObjectPermissionsFilter,\n ]\n permission_classes = [\n permissions.DjangoObjectPermissions,\n ]\n\n\nclass RatingViewSet(viewsets.ModelViewSet):\n queryset = models.Rating.objects.all()\n serializer_class = serializers.RatingSerializer\n","new_contents":"#!\/usr\/bin\/python\n# -*- coding: utf-8 -*-\n\nfrom rest_framework import viewsets, filters\n\nfrom .. import (\n models,\n permissions,\n)\nfrom ..serializers import (\n primarykey as serializers\n)\n\n\nclass MovieViewSet(viewsets.ReadOnlyModelViewSet):\n queryset = models.Movie.objects.all()\n serializer_class = serializers.MovieSerializer\n filter_backends = (\n filters.OrderingFilter,\n )\n ordering_fields = (\n 'title',\n 'released',\n 'runtime',\n )\n\n\nclass GenreViewSet(viewsets.ReadOnlyModelViewSet):\n queryset = models.Genre.objects.all()\n serializer_class = serializers.GenreSerializer\n filter_backends = (\n filters.OrderingFilter,\n )\n ordering_fields = (\n 'name',\n )\n\n\nclass WatchlistViewSet(viewsets.ModelViewSet):\n queryset = models.Watchlist.objects.all()\n serializer_class = serializers.WatchlistSerializer\n filter_backends = (\n filters.DjangoObjectPermissionsFilter,\n )\n permission_classes = (\n permissions.DjangoObjectPermissions,\n )\n\n\nclass ResumeViewSet(viewsets.ModelViewSet):\n queryset = models.Resume.objects.all()\n serializer_class = serializers.ResumeSerializer\n filter_backends = (\n filters.DjangoObjectPermissionsFilter,\n )\n permission_classes = (\n permissions.DjangoObjectPermissions,\n )\n\n\nclass RatingViewSet(viewsets.ModelViewSet):\n queryset = models.Rating.objects.all()\n serializer_class = serializers.RatingSerializer\n","subject":"Add ordering filters for primary key based api.","message":"Add ordering filters for primary key based api.\n","lang":"Python","license":"mit","repos":"fladi\/django-campus02,fladi\/django-campus02"} {"commit":"06645a637c0d34270f88f9a6b96133da5c415dd7","old_file":"froide\/publicbody\/admin.py","new_file":"froide\/publicbody\/admin.py","old_contents":"from django.contrib import admin\nfrom froide.publicbody.models import PublicBody, FoiLaw\n\nclass PublicBodyAdmin(admin.ModelAdmin):\n prepopulated_fields = {\"slug\": (\"geography\", \"name\",)}\n list_display = ('name', 'classification', 'geography')\n list_filter = ('classification',)\n search_fields = ['name', \"description\"]\n exclude = ('confirmed',)\n\nclass FoiLawAdmin(admin.ModelAdmin):\n prepopulated_fields = {\"slug\": (\"jurisdiction\", \"name\",)}\n\nadmin.site.register(PublicBody, PublicBodyAdmin)\nadmin.site.register(FoiLaw, FoiLawAdmin)\n","new_contents":"from django.contrib import admin\nfrom froide.publicbody.models import PublicBody, FoiLaw\n\nclass PublicBodyAdmin(admin.ModelAdmin):\n prepopulated_fields = {\"slug\": (\"geography\", \"name\",)}\n list_display = ('name', 'classification', 'topic', 'geography')\n list_filter = ('classification', 'topic',)\n search_fields = ['name', \"description\"]\n exclude = ('confirmed',)\n\nclass FoiLawAdmin(admin.ModelAdmin):\n prepopulated_fields = {\"slug\": (\"jurisdiction\", \"name\",)}\n\nadmin.site.register(PublicBody, PublicBodyAdmin)\nadmin.site.register(FoiLaw, FoiLawAdmin)\n","subject":"Add topic to PublicBodyAdmin list_filter and list_display","message":"Add topic to PublicBodyAdmin list_filter and list_display","lang":"Python","license":"mit","repos":"catcosmo\/froide,ryankanno\/froide,stefanw\/froide,catcosmo\/froide,fin\/froide,CodeforHawaii\/froide,fin\/froide,okfse\/froide,LilithWittmann\/froide,okfse\/froide,stefanw\/froide,ryankanno\/froide,LilithWittmann\/froide,LilithWittmann\/froide,CodeforHawaii\/froide,fin\/froide,catcosmo\/froide,catcosmo\/froide,okfse\/froide,ryankanno\/froide,okfse\/froide,catcosmo\/froide,CodeforHawaii\/froide,ryankanno\/froide,LilithWittmann\/froide,fin\/froide,LilithWittmann\/froide,stefanw\/froide,stefanw\/froide,stefanw\/froide,ryankanno\/froide,CodeforHawaii\/froide,CodeforHawaii\/froide,okfse\/froide"} {"commit":"621d285c05ce3a6257edcffec03c8a96507b6179","old_file":"name\/feeds.py","new_file":"name\/feeds.py","old_contents":"from django.contrib.syndication.views import Feed\nfrom django.core.urlresolvers import reverse_lazy\nfrom django.utils.feedgenerator import Atom1Feed\n\nfrom . import app_settings\nfrom .models import Name, Location\n\n\nclass NameAtomFeedType(Atom1Feed):\n \"\"\"Create an Atom feed that sets the Content-Type response\n header to application\/xml.\n \"\"\"\n mime_type = 'application\/xml'\n\n\nclass NameAtomFeed(Feed):\n feed_type = NameAtomFeedType\n link = reverse_lazy(\"name_feed\")\n title = \"Name App\"\n subtitle = \"New Name Records\"\n author_name = app_settings.NAME_FEED_AUTHOR_NAME\n author_email = app_settings.NAME_FEED_AUTHOR_EMAIL\n author_link = app_settings.NAME_FEED_AUTHOR_LINK\n\n def items(self):\n # last 5 added items\n return Name.objects.order_by('-date_created')[:20]\n\n def item_title(self, item):\n return item.name\n\n def item_location(self, item):\n \"\"\"\n Returns an extra keyword arguments dictionary that is used\n with the `add_item` call of the feed generator. Add the\n 'content' field of the 'Entry' item, to be used by the custom\n feed generator.\n \"\"\"\n location_set = []\n for l in Location.objects.filter(belong_to_name=item):\n location_set.append(\n 'georss:point', \"%s %s\" % (l.latitude, l.longitude)\n )\n return location_set\n\n def item_description(self, item):\n return \"Name Type: %s\" % item.get_name_type_label()\n\n def item_link(self, item):\n return item.get_absolute_url()\n","new_contents":"from django.contrib.syndication.views import Feed\nfrom django.core.urlresolvers import reverse_lazy\nfrom django.utils.feedgenerator import Atom1Feed\n\nfrom . import app_settings\nfrom .models import Name\n\n\nclass NameAtomFeedType(Atom1Feed):\n \"\"\"Create an Atom feed that sets the Content-Type response\n header to application\/xml.\n \"\"\"\n mime_type = 'application\/xml'\n\n\nclass NameAtomFeed(Feed):\n feed_type = NameAtomFeedType\n link = reverse_lazy(\"name_feed\")\n title = \"Name App\"\n subtitle = \"New Name Records\"\n author_name = app_settings.NAME_FEED_AUTHOR_NAME\n author_email = app_settings.NAME_FEED_AUTHOR_EMAIL\n author_link = app_settings.NAME_FEED_AUTHOR_LINK\n\n def items(self):\n # last 5 added items\n return Name.objects.order_by('-date_created')[:20]\n\n def item_title(self, obj):\n return obj.name\n\n def item_description(self, obj):\n return 'Name Type: {0}'.format(obj.get_name_type_label())\n\n def item_link(self, obj):\n return obj.get_absolute_url()\n","subject":"Change the formating of item_description. Remove item_location because it was not used.","message":"Change the formating of item_description. Remove item_location because it was not used.\n","lang":"Python","license":"bsd-3-clause","repos":"unt-libraries\/django-name,damonkelley\/django-name,unt-libraries\/django-name,unt-libraries\/django-name,damonkelley\/django-name,damonkelley\/django-name"} {"commit":"002dd6fa4af36bd722b3f194c93f1e2e628ad561","old_file":"inboxen\/app\/model\/email.py","new_file":"inboxen\/app\/model\/email.py","old_contents":"from inboxen.models import Alias, Attachment, Email, Header\nfrom config.settings import datetime_format, recieved_header_name\nfrom datetime import datetime\n\ndef make_email(message, alias, domain):\n inbox = Alias.objects.filter(alias=alias, domain__domain=domain)[0]\n user = inbox.user\n body = message.base.body\n recieved_date = datetime.strptime(message[recieved_header_name], datetime_format)\n del message[recieved_header_name]\n\n email = Email(inbox=inbox, user=user, body=body, recieved_date=recieved_date)\n email.save()\n\n for name in message.keys():\n email.headers.create(name=name, data=message[name])\n\n for part in message.walk():\n if not part.body:\n part.body = u''\n email.attachments.create(\n content_type=part.content_encoding['Content-Type'][0],\n content_disposition=part.content_encoding['Content-Disposition'][0],\n data=part.body\n )\n\n email.save()\n","new_contents":"from inboxen.models import Alias, Attachment, Email, Header\nfrom config.settings import datetime_format, recieved_header_name\nfrom datetime import datetime\n\ndef make_email(message, alias, domain):\n inbox = Alias.objects.filter(alias=alias, domain__domain=domain)[0]\n user = inbox.user\n body = message.base.body\n recieved_date = datetime.strptime(message[recieved_header_name], datetime_format)\n del message[recieved_header_name]\n\n email = Email(inbox=inbox, user=user, body=body, recieved_date=recieved_date)\n email.save()\n\n head_list = []\n for name in message.keys():\n header = Header(name=name, data=message[name])\n header.save()\n head_list.append(header)\n # add all the headers at once should save us some queries\n email.headers.add(*head_list)\n\n attach_list = []\n for part in message.walk():\n if not part.body:\n part.body = u''\n attachment = Attachment(\n content_type=part.content_encoding['Content-Type'][0],\n content_disposition=part.content_encoding['Content-Disposition'][0],\n data=part.body\n )\n attachment.save()\n attach_list.append(attachment)\n # as with headers above\n email.attachments.add(*attach_list)\n","subject":"Reduce number of queries to DB","message":"Reduce number of queries to DB\n","lang":"Python","license":"agpl-3.0","repos":"Inboxen\/router,Inboxen\/Inboxen,Inboxen\/Inboxen,Inboxen\/Inboxen,Inboxen\/Inboxen"} {"commit":"1ecd6cacac15bff631b958ee6773b2ad8659df50","old_file":"opps\/images\/widgets.py","new_file":"opps\/images\/widgets.py","old_contents":"from django import forms\nfrom django.conf import settings\nfrom django.template.loader import render_to_string\n\n\nclass MultipleUpload(forms.FileInput):\n\n def render(self, name, value, attrs=None):\n _value = \"\"\n if value:\n _value = \"{0}{1}\".format(settings.MEDIA_URL, value)\n return render_to_string(\"admin\/opps\/images\/multiupload.html\",\n {\"name\": name, \"value\": _value,\n \"STATIC_URL\": settings.STATIC_URL})\n","new_contents":"from django import forms\nfrom django.conf import settings\nfrom django.template.loader import render_to_string\n\n\nclass MultipleUpload(forms.FileInput):\n\n def render(self, name, value, attrs=None):\n _value = \"\"\n if value:\n _value = \"{0}{1}\".format(settings.MEDIA_URL, value)\n return render_to_string(\"admin\/opps\/images\/multiupload.html\",\n {\"name\": name, \"value\": _value,\n \"STATIC_URL\": settings.STATIC_URL})\n\nclass CropExample(forms.TextInput):\n\n def render(self, name, value, attrs=None):\n return render_to_string(\n \"admin\/opps\/images\/cropexample.html\",\n {\"name\": name, \"value\": value,\n \"THUMBOR_SERVER\": settings.THUMBOR_SERVER,\n \"THUMBOR_MEDIA_URL\": settings.THUMBOR_MEDIA_URL})\n","subject":"Create widget CropExample on images","message":"Create widget CropExample on images\n","lang":"Python","license":"mit","repos":"jeanmask\/opps,williamroot\/opps,opps\/opps,jeanmask\/opps,opps\/opps,YACOWS\/opps,jeanmask\/opps,opps\/opps,williamroot\/opps,jeanmask\/opps,williamroot\/opps,YACOWS\/opps,opps\/opps,YACOWS\/opps,williamroot\/opps,YACOWS\/opps"} {"commit":"1a03fbbb612d8faa5a6733fe7d4920f3ca158a69","old_file":"acme\/utils\/observers.py","new_file":"acme\/utils\/observers.py","old_contents":"# Copyright 2018 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Metrics observers.\"\"\"\n\nimport abc\nfrom typing import Dict, Union\n\nimport dm_env\nimport numpy as np\n\n\nNumber = Union[int, float]\n\n\nclass EnvLoopObserver(abc.ABC):\n \"\"\"An interface for collecting metrics\/counters in EnvironmentLoop.\"\"\"\n\n def observe_first(self, env: dm_env.Environment, timestep: dm_env.TimeStep\n ) -> None:\n \"\"\"Observes the initial state.\"\"\"\n\n @abc.abstractmethod\n def observe(self, env: dm_env.Environment, timestep: dm_env.TimeStep,\n action: np.ndarray) -> None:\n \"\"\"Records one environment step.\"\"\"\n\n @abc.abstractmethod\n def get_metrics(self) -> Dict[str, Number]:\n \"\"\"Returns metrics collected for the current episode.\"\"\"\n","new_contents":"# Copyright 2018 DeepMind Technologies Limited. All rights reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"Metrics observers.\"\"\"\n\nimport abc\nfrom typing import Dict, Union\n\nimport dm_env\nimport numpy as np\n\n\nNumber = Union[int, float]\n\n\nclass EnvLoopObserver(abc.ABC):\n \"\"\"An interface for collecting metrics\/counters in EnvironmentLoop.\"\"\"\n\n @abc.abstractmethod\n def observe_first(self, env: dm_env.Environment, timestep: dm_env.TimeStep\n ) -> None:\n \"\"\"Observes the initial state.\"\"\"\n\n @abc.abstractmethod\n def observe(self, env: dm_env.Environment, timestep: dm_env.TimeStep,\n action: np.ndarray) -> None:\n \"\"\"Records one environment step.\"\"\"\n\n @abc.abstractmethod\n def get_metrics(self) -> Dict[str, Number]:\n \"\"\"Returns metrics collected for the current episode.\"\"\"\n","subject":"Add a missing @abc.abstractmethod declaration","message":"Add a missing @abc.abstractmethod declaration\n\nPiperOrigin-RevId: 424036906\nChange-Id: I3979c35ec30caa4b684d43376b2a36d21f7e79df\n","lang":"Python","license":"apache-2.0","repos":"deepmind\/acme,deepmind\/acme"} {"commit":"018d062f2ed9ca9acd3d555d439dd81b89c88a6f","old_file":"parse_push\/managers.py","new_file":"parse_push\/managers.py","old_contents":"\nfrom django.db import models\n\n\nclass DeviceManager(models.Manager):\n\n def latest(self):\n \"\"\" Returns latest Device instance \"\"\"\n return self.latest('created_at')\n\n","new_contents":"\nfrom django.db import models\n\n\nclass DeviceManager(models.Manager):\n\n def get_latest(self):\n \"\"\" Returns latest Device instance \"\"\"\n return self.get_queryset().latest('created_at')\n\n","subject":"Use .get_latest() instead so that we do not override built-in .latest()","message":"Use .get_latest() instead so that we do not override built-in .latest()\n","lang":"Python","license":"bsd-3-clause","repos":"willandskill\/django-parse-push"} {"commit":"69b82f664ba0c294498f8c7d139a1e0ecee12e2d","old_file":"imgcat\/__init__.py","new_file":"imgcat\/__init__.py","old_contents":"\"\"\"\nThe imgcat module\n\"\"\"\n\n__version__ = '0.5.0'\n\n\nfrom .imgcat import (\n imgcat,\n main\n)\n\ntry:\n # support module:\/\/imgcat backend\n from .mpl_backend import (\n new_figure_manager, show\n )\nexcept ImportError:\n # matplotlib is not available, do nothing\n pass\n\n\n# IPython magic support: %load_ext imgcat\ndef load_ipython_extension(ipython):\n from .ipython_magic import ImgcatMagics\n ipython.register_magics(ImgcatMagics)\n","new_contents":"\"\"\"\nThe imgcat module\n\"\"\"\n\n__version__ = '0.6.0.dev0'\n\n\nfrom .imgcat import (\n imgcat,\n main\n)\n\ntry:\n # support module:\/\/imgcat backend\n from .mpl_backend import (\n new_figure_manager, show\n )\nexcept ImportError:\n # matplotlib is not available, do nothing\n pass\n\n\n# IPython magic support: %load_ext imgcat\ndef load_ipython_extension(ipython):\n from .ipython_magic import ImgcatMagics\n ipython.register_magics(ImgcatMagics)\n","subject":"Bump to next development version: 0.6.0.dev0","message":"Bump to next development version: 0.6.0.dev0\n","lang":"Python","license":"mit","repos":"wookayin\/python-imgcat"} {"commit":"ab2f4aaf9546787a269a3d0ec5b3b83c86a43bde","old_file":"Languages.py","new_file":"Languages.py","old_contents":"#!\/usr\/bin\/env python3\nimport requests\nimport re\n\ndef findAllLanguages():\n \"Find a list of Crowdin language codes to which KA is translated to\"\n response = requests.get(\"https:\/\/crowdin.com\/project\/khanacademy\")\n txt = response.text\n for match in re.findall(r\"https?:\/\/[a-z0-9]*\\.cloudfront\\.net\/images\/flags\/([^\\.]+)\\.png\", txt):\n yield match\n\nif __name__ == \"__main__\":\n print(list(findAllLanguages()))","new_contents":"#!\/usr\/bin\/env python3\nimport requests\nimport re\n\ndef findAllLanguages():\n \"Find a list of Crowdin language codes to which KA is translated to\"\n response = requests.get(\"https:\/\/crowdin.com\/project\/khanacademy\")\n txt = response.text\n langs = set()\n for match in re.findall(r\"https?:\/\/[a-z0-9]*\\.cloudfront\\.net\/images\/flags\/([^\\.]+)\\.png\", txt):\n langs.add(match)\n return langs\n\nif __name__ == \"__main__\":\n print(findAllLanguages())","subject":"Return language set instead of language generator","message":"Return language set instead of language generator\n","lang":"Python","license":"apache-2.0","repos":"ulikoehler\/KATranslationCheck,ulikoehler\/KATranslationCheck,KA-Advocates\/KATranslationCheck,KA-Advocates\/KATranslationCheck,ulikoehler\/KATranslationCheck,KA-Advocates\/KATranslationCheck,ulikoehler\/KATranslationCheck,ulikoehler\/KATranslationCheck,KA-Advocates\/KATranslationCheck"} {"commit":"8d313884a52b06e2fdf9a3c0d152b9e711ff02c2","old_file":"kkbox\/trac\/secretticket.py","new_file":"kkbox\/trac\/secretticket.py","old_contents":"from trac.core import Component, implements\nfrom trac.perm import IPermissionRequestor\n\nclass KKBOXSecretTicketsPolicy(Component):\n implements(IPermissionRequestor)\n\n def get_permission_actions(self):\n return ['SECRET_VIEW']\n","new_contents":"from trac.ticket.model import Ticket\nfrom trac.core import Component, implements, TracError\nfrom trac.perm import IPermissionPolicy\n\nclass KKBOXSecretTicketsPolicy(Component):\n implements(IPermissionPolicy)\n\n def __init__(self):\n config = self.env.config\n self.sensitive_keyword = config.get('kkbox', 'sensitive_keyword').strip()\n\n def check_permission(self, action, user, resource, perm):\n while resource:\n if 'ticket' == resource.realm:\n break\n resource = resource.parent\n\n if resource and 'ticket' == resource.realm and resource.id:\n return self.check_ticket_access(perm, resource)\n\n def check_ticket_access(self, perm, res):\n if not self.sensitive_keyword:\n return None\n\n try:\n ticket = Ticket(self.env, res.id)\n keywords = [k.strip() for k in ticket['keywords'].split(',')]\n if self.sensitive_keyword in keywords:\n cc_list = [cc.strip() for cc in ticket['cc'].split(',')]\n\n if perm.username == ticket['reporter'] or \\\n perm.username == ticket['owner'] or \\\n perm.username in cc_list:\n return None\n else:\n return False\n except TracError as e:\n self.log.error(e.message)\n return None\n","subject":"Mark ticket as sensitive by keyword","message":"Mark ticket as sensitive by keyword\n\nSet sensitive_keyword in trac.ini as following example, These ticket has\n\"secret\" keyword are viewable by reporter, owner and cc.\n\n[kkbox]\nsensitive_keyword = secret\n","lang":"Python","license":"bsd-3-clause","repos":"KKBOX\/trac-keyword-secret-ticket-plugin"} {"commit":"01516489dbf9ee78128d653b3ebc46730d466425","old_file":"apps\/api\/serializers.py","new_file":"apps\/api\/serializers.py","old_contents":"# -*- coding: utf-8 -*-\nfrom rest_framework import serializers\n\nfrom apps.broadcasts.models import Broadcast, Host, Raid, Series\nfrom apps.games.models import Game, Platform\nfrom apps.subscribers.models import Ticket\n\n\nclass HostSerializer(serializers.ModelSerializer):\n class Meta:\n fields = ('id', 'timestamp', 'username', 'broadcast')\n model = Host\n\n\nclass RaidSerializer(serializers.ModelSerializer):\n class Meta:\n fields = ('id', 'timestamp', 'username', 'broadcast', 'game')\n model = Raid\n\n\nclass SeriesSerializer(serializers.ModelSerializer):\n class Meta:\n model = Series\n\n\nclass GameSerializer(serializers.ModelSerializer):\n class Meta:\n model = Game\n\n\nclass PlatformSerializer(serializers.ModelSerializer):\n class Meta:\n model = Platform\n\n\nclass TicketSerializer(serializers.ModelSerializer):\n class Meta:\n model = Ticket\n\n\nclass BroadcastSerializer(serializers.ModelSerializer):\n hosts = serializers.PrimaryKeyRelatedField(many=True, read_only=True)\n raids = serializers.PrimaryKeyRelatedField(many=True, read_only=True)\n\n class Meta:\n model = Broadcast\n","new_contents":"# -*- coding: utf-8 -*-\nfrom rest_framework import serializers\n\nfrom apps.broadcasts.models import Broadcast, Host, Raid, Series\nfrom apps.games.models import Game, Platform\nfrom apps.subscribers.models import Ticket\n\n\nclass HostSerializer(serializers.ModelSerializer):\n class Meta:\n fields = ('id', 'timestamp', 'username', 'broadcast')\n model = Host\n\n\nclass RaidSerializer(serializers.ModelSerializer):\n class Meta:\n fields = ('id', 'timestamp', 'username', 'broadcast', 'game')\n model = Raid\n\n\nclass SeriesSerializer(serializers.ModelSerializer):\n class Meta:\n model = Series\n\n\nclass GameSerializer(serializers.ModelSerializer):\n appearances = serializers.IntegerField(source='appears_on.count', read_only=True)\n\n class Meta:\n model = Game\n\n\nclass PlatformSerializer(serializers.ModelSerializer):\n class Meta:\n model = Platform\n\n\nclass TicketSerializer(serializers.ModelSerializer):\n class Meta:\n model = Ticket\n\n\nclass BroadcastSerializer(serializers.ModelSerializer):\n hosts = serializers.PrimaryKeyRelatedField(many=True, read_only=True)\n raids = serializers.PrimaryKeyRelatedField(many=True, read_only=True)\n\n class Meta:\n model = Broadcast\n","subject":"Add appearance count to the API.","message":"Add appearance count to the API.\n","lang":"Python","license":"apache-2.0","repos":"bryanveloso\/avalonstar-tv,bryanveloso\/avalonstar-tv,bryanveloso\/avalonstar-tv"} {"commit":"8ee38953a9f8bdbd95ace4ea45e3673cc260bb4b","old_file":"scripts\/cronRefreshEdxQualtrics.py","new_file":"scripts\/cronRefreshEdxQualtrics.py","old_contents":"import getopt\nimport sys\nimport os\n\n### Script for scheduling regular EdxQualtrics updates\n### Usage for cron should be \"cronRefreshEdxQualtrics.py -m -s -r\"\n\n# Append directory for dependencies to PYTHONPATH\n# sys.path.append(\"\/home\/dataman\/Code\/qualtrics_etl\/src\/qualtrics_etl\/\")\nsource_dir = [os.path.join(os.path.dirname(os.path.abspath(__file__)), \"..\/json_to_relation\/\")]\nsource_dir.extend(sys.path)\nsys.path = source_dir\n\nfrom surveyextractor import QualtricsExtractor\n\nqe = QualtricsExtractor()\nopts, args = getopt.getopt(sys.argv[1:], 'amsr', ['--reset', '--loadmeta', '--loadsurveys', '--loadresponses'])\nfor opt, arg in opts:\n if opt in ('-a', '--reset'):\n qe.resetMetadata()\n qe.resetSurveys()\n qe.resetResponses()\n elif opt in ('-m', '--loadmeta'):\n qe.loadSurveyMetadata()\n elif opt in ('-s', '--loadsurvey'):\n qe.resetSurveys()\n qe.loadSurveyData()\n elif opt in ('-r', '--loadresponses'):\n qe.loadResponseData()\n","new_contents":"#!\/usr\/bin\/env python\n\nimport getopt\nimport sys\nimport os\n\n### Script for scheduling regular EdxQualtrics updates\n### Usage for cron should be \"cronRefreshEdxQualtrics.py -m -s -r\"\n\n# Append directory for dependencies to PYTHONPATH\n# sys.path.append(\"\/home\/dataman\/Code\/qualtrics_etl\/src\/qualtrics_etl\/\")\nsource_dir = [os.path.join(os.path.dirname(os.path.abspath(__file__)), \"..\/json_to_relation\/\")]\nsource_dir.extend(sys.path)\nsys.path = source_dir\n\nfrom surveyextractor import QualtricsExtractor\n\nqe = QualtricsExtractor()\nopts, args = getopt.getopt(sys.argv[1:], 'amsr', ['--reset', '--loadmeta', '--loadsurveys', '--loadresponses'])\nfor opt, arg in opts:\n if opt in ('-a', '--reset'):\n qe.resetMetadata()\n qe.resetSurveys()\n qe.resetResponses()\n elif opt in ('-m', '--loadmeta'):\n qe.loadSurveyMetadata()\n elif opt in ('-s', '--loadsurvey'):\n qe.resetSurveys()\n qe.loadSurveyData()\n elif opt in ('-r', '--loadresponses'):\n qe.loadResponseData()\n","subject":"Add python environment to cron qualtrics script","message":"Add python environment to cron qualtrics script\n","lang":"Python","license":"bsd-3-clause","repos":"paepcke\/json_to_relation,paepcke\/json_to_relation,paepcke\/json_to_relation,paepcke\/json_to_relation"} {"commit":"6ac1c09422e82d97e3a9e9bc8d52c8814c33bc27","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom setuptools import setup, find_packages, Command\nimport os\n\npackages = find_packages()\n\n\nclass PyTest(Command):\n user_options = []\n\n def initialize_options(self):\n pass\n\n def finalize_options(self):\n pass\n\n def run(self):\n import sys\n import subprocess\n errno = subprocess.call([sys.executable, 'runtests.py'])\n raise SystemExit(errno)\n\n\ndef get_locals(filename):\n l = {}\n execfile(filename, {}, l)\n return l\n\nmetadata = get_locals(os.path.join('bids_writer', '_metadata.py'))\n\nsetup(\n name=\"bids-json-writer\",\n version=metadata['version'],\n author=metadata['author'],\n author_email=metadata['author_email'],\n license=metadata['license'],\n url=metadata['url'],\n packages=find_packages(),\n cmdclass={'test': PyTest},\n entry_points={\n 'console_scripts': [\n ]}\n )\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom setuptools import setup, find_packages, Command\nimport os\n\npackages = find_packages()\n\n\nclass PyTest(Command):\n user_options = []\n\n def initialize_options(self):\n pass\n\n def finalize_options(self):\n pass\n\n def run(self):\n import sys\n import subprocess\n errno = subprocess.call([sys.executable, 'runtests.py'])\n raise SystemExit(errno)\n\n\ndef get_locals(filename):\n l = {}\n execfile(filename, {}, l)\n return l\n\nmetadata = get_locals(os.path.join('markdown_to_json', '_metadata.py'))\n\nsetup(\n name=\"markdown-to-json\",\n version=metadata['version'],\n author=metadata['author'],\n author_email=metadata['author_email'],\n license=metadata['license'],\n url=metadata['url'],\n packages=find_packages(),\n cmdclass={'test': PyTest},\n entry_points={\n 'console_scripts': [\n 'md_to_json = markdown_to_json.scripts.md_to_json:main'\n ]}\n )\n","subject":"Rename package, add script to manifest","message":"Rename package, add script to manifest","lang":"Python","license":"mit","repos":"njvack\/markdown-to-json"} {"commit":"91ad52d6d47ce12966e5fb23913a8c5b600b2c13","old_file":"setup.py","new_file":"setup.py","old_contents":"from buckle.version import VERSION\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name='buckle',\n version=VERSION,\n description='Buckle: It ties your toolbelt together',\n author='Nextdoor',\n author_email='eng@nextdoor.com',\n packages=find_packages(exclude=['ez_setup']),\n scripts=['bin\/buckle',\n 'bin\/buckle-init',\n 'bin\/buckle-help',\n 'bin\/buckle-_help-helper',\n 'bin\/buckle-readme',\n 'bin\/buckle-version',\n ],\n test_suite=\"tests\",\n install_requires=[\n 'future>=0.15.2',\n ],\n tests_require=[\n 'pytest',\n ],\n extras_require={\n ':python_version <= \"3.2\"': [\n 'subprocess32',\n ],\n },\n url='https:\/\/github.com\/Nextdoor\/buckle',\n include_package_data=True\n)\n","new_contents":"from buckle.version import VERSION\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name='buckle',\n version=VERSION,\n description='Buckle: It ties your toolbelt together',\n author='Nextdoor',\n author_email='eng@nextdoor.com',\n packages=find_packages(exclude=['ez_setup']),\n scripts=['bin\/buckle',\n 'bin\/buckle-init',\n 'bin\/buckle-help',\n 'bin\/buckle-_help-helper',\n 'bin\/buckle-readme',\n 'bin\/buckle-version',\n ],\n test_suite=\"tests\",\n install_requires=[\n 'future>=0.15.2',\n ],\n tests_require=[\n 'pytest',\n ],\n extras_require={\n ':python_version < \"3.3\"': [\n 'subprocess32',\n ],\n },\n url='https:\/\/github.com\/Nextdoor\/buckle',\n include_package_data=True\n)\n","subject":"Use < instead of <= in environment markers.","message":"Use < instead of <= in environment markers.\n","lang":"Python","license":"bsd-2-clause","repos":"Nextdoor\/buckle,Nextdoor\/buckle"} {"commit":"ff0468f51b6a5cebd00f2cea8d2abd5f74e925d6","old_file":"ometa\/tube.py","new_file":"ometa\/tube.py","old_contents":"from ometa.interp import TrampolinedGrammarInterpreter, _feed_me\n\nclass TrampolinedParser:\n \"\"\"\n A parser that incrementally parses incoming data.\n \"\"\"\n def __init__(self, grammar, receiver, bindings):\n \"\"\"\n Initializes the parser.\n\n @param grammar: The grammar used to parse the incoming data.\n @param receiver: Responsible for logic operation on the parsed data.\n Typically, the logic operation will be invoked inside the grammar,\n e.g., rule = expr1 expr2 (-> receiver.doSomeStuff())\n @param bindings: The namespace that can be accessed inside the grammar.\n \"\"\"\n self.grammar = grammar\n self.bindings = dict(bindings)\n self.bindings['receiver'] = self.receiver = receiver\n self._setupInterp()\n\n\n def _setupInterp(self):\n \"\"\"\n Resets the parser. The parser will begin parsing with the rule named\n 'initial'.\n \"\"\"\n self._interp = TrampolinedGrammarInterpreter(\n grammar=self.grammar, ruleName='initial', callback=None,\n globals=self.bindings)\n\n\n def receive(self, data):\n \"\"\"\n Receive the incoming data and begin parsing. The parser will parse the\n data incrementally according to the 'initial' rule in the grammar.\n\n @param data: The raw data received.\n \"\"\"\n while data:\n try:\n status = self._interp.receive(data)\n except Exception as e:\n # maybe we should raise it?\n raise e\n else:\n if status is _feed_me:\n return\n data = ''.join(self._interp.input.data[self._interp.input.position:])\n self._setupInterp()","new_contents":"from ometa.interp import TrampolinedGrammarInterpreter, _feed_me\n\nclass TrampolinedParser:\n \"\"\"\n A parser that incrementally parses incoming data.\n \"\"\"\n def __init__(self, grammar, receiver, bindings):\n \"\"\"\n Initializes the parser.\n\n @param grammar: The grammar used to parse the incoming data.\n @param receiver: Responsible for logic operation on the parsed data.\n Typically, the logic operation will be invoked inside the grammar,\n e.g., rule = expr1 expr2 (-> receiver.doSomeStuff())\n @param bindings: The namespace that can be accessed inside the grammar.\n \"\"\"\n self.grammar = grammar\n self.bindings = dict(bindings)\n self.bindings['receiver'] = self.receiver = receiver\n self._setupInterp()\n\n\n def _setupInterp(self):\n \"\"\"\n Resets the parser. The parser will begin parsing with the rule named\n 'initial'.\n \"\"\"\n self._interp = TrampolinedGrammarInterpreter(\n grammar=self.grammar, ruleName=self.receiver.currentRule,\n callback=None, globals=self.bindings)\n\n\n def receive(self, data):\n \"\"\"\n Receive the incoming data and begin parsing. The parser will parse the\n data incrementally according to the 'initial' rule in the grammar.\n\n @param data: The raw data received.\n \"\"\"\n while data:\n status = self._interp.receive(data)\n if status is _feed_me:\n return\n data = ''.join(self._interp.input.data[self._interp.input.position:])\n self._setupInterp()\n","subject":"Update TrampolinedParser a little for my purposes.","message":"Update TrampolinedParser a little for my purposes.\n","lang":"Python","license":"mit","repos":"rbtcollins\/parsley,python-parsley\/parsley,python-parsley\/parsley,rbtcollins\/parsley"} {"commit":"3955d10f5dd905610c9621046069ae8dacbb1c1e","old_file":"setup.py","new_file":"setup.py","old_contents":"try:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\n\nconfig = {\n 'description': 'A simple python LOC count tool',\n 'author': 'Tihomir Saulic',\n 'url': 'http:\/\/github.com\/tsaulic\/pycount',\n 'download_url': 'http:\/\/github.com\/tsaulic\/pycount',\n 'author_email': 'tihomir[DOT]saulic[AT]gmail[DOT]com',\n 'version': '0.6.1',\n 'install_requires': ['binaryornot'],\n 'packages': ['pycount'],\n 'scripts': ['bin\/pycount'],\n 'name': 'pycount'\n}\n\nsetup(**config)\n","new_contents":"try:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\n\nconfig = {\n 'description': 'A simple python LOC count tool',\n 'author': 'Tihomir Saulic',\n 'url': 'http:\/\/github.com\/tsaulic\/pycount',\n 'download_url': 'http:\/\/github.com\/tsaulic\/pycount',\n 'author_email': 'tihomir[DOT]saulic[AT]gmail[DOT]com',\n 'version': '0.6.2',\n 'install_requires': ['binaryornot', 'pygal'],\n 'packages': ['pycount'],\n 'scripts': ['bin\/pycount'],\n 'name': 'pycount'\n}\n\nsetup(**config)\n","subject":"Add dependency and bump version.","message":"Add dependency and bump version.\n","lang":"Python","license":"mit","repos":"tsaulic\/pycount"} {"commit":"c1def9580859eb97368aa49a2e26aab483785b35","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\nimport biobox_cli\n\nsetup(\n name = 'biobox_cli',\n version = biobox_cli.__version__,\n description = 'Run biobox Docker containers on the command line',\n author = 'bioboxes',\n author_email = 'mail@bioboxes.org',\n url = 'http:\/\/bioboxes.org',\n packages = ['biobox_cli'],\n scripts = ['bin\/biobox'],\n install_requires = open('requirements.txt').read().splitlines(),\n\n classifiers = [\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Scientific\/Engineering :: Bio-Informatics',\n 'Intended Audience :: Science\/Research',\n 'Operating System :: POSIX'\n ],\n)\n","new_contents":"from setuptools import setup, find_packages\nimport biobox_cli\n\nsetup(\n name = 'biobox-cli',\n version = biobox_cli.__version__,\n description = 'Run biobox Docker containers on the command line',\n author = 'bioboxes',\n author_email = 'mail@bioboxes.org',\n url = 'http:\/\/bioboxes.org',\n packages = ['biobox_cli'],\n scripts = ['bin\/biobox'],\n install_requires = open('requirements.txt').read().splitlines(),\n\n classifiers = [\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Scientific\/Engineering :: Bio-Informatics',\n 'Intended Audience :: Science\/Research',\n 'Operating System :: POSIX'\n ],\n)\n","subject":"Use hyphen in package name","message":"Use hyphen in package name\n\nSigned-off-by: Michael Barton <1d6e1cf70ec6f9ab28d3ea4b27a49a77654d370e@michaelbarton.me.uk>\n","lang":"Python","license":"mit","repos":"pbelmann\/command-line-interface,michaelbarton\/command-line-interface,fungs\/bbx-cli,michaelbarton\/command-line-interface,bioboxes\/command-line-interface,pbelmann\/command-line-interface,bioboxes\/command-line-interface,fungs\/bbx-cli"} {"commit":"7b4d9e3c1a2da7c7546d4366910eb2a5582e6abc","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport sys\n\nimport crabpy\n\nfrom setuptools import setup, find_packages\n\npackages = [\n 'crabpy',\n]\n\nrequires = [\n 'suds-jurko>=0.5.0',\n 'dogpile.cache'\n]\n\nsetup(\n name='crabpy',\n version='0.2.1',\n description='Interact with AGIV webservices.',\n long_description=open('README.rst').read() + '\\n\\n' +\n open('CHANGES.rst').read(),\n author='Onroerend Erfgoed',\n author_email='ict@onroerenderfgoed.be',\n url='http:\/\/github.com\/onroerenderfgoed\/crabpy',\n packages=find_packages(),\n package_data={'': ['LICENSE']},\n package_dir={'crabpy': 'crabpy'},\n include_package_data=True,\n install_requires=requires,\n license='MIT',\n zip_safe=False,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n ],\n test_suite='nose.collector'\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport sys\n\nimport crabpy\n\nfrom setuptools import setup, find_packages\n\npackages = [\n 'crabpy',\n]\n\nrequires = [\n 'suds-jurko>=0.6.0',\n 'dogpile.cache'\n]\n\nsetup(\n name='crabpy',\n version='0.2.1',\n description='Interact with AGIV webservices.',\n long_description=open('README.rst').read() + '\\n\\n' +\n open('CHANGES.rst').read(),\n author='Onroerend Erfgoed',\n author_email='ict@onroerenderfgoed.be',\n url='http:\/\/github.com\/onroerenderfgoed\/crabpy',\n packages=find_packages(),\n package_data={'': ['LICENSE']},\n package_dir={'crabpy': 'crabpy'},\n include_package_data=True,\n install_requires=requires,\n license='MIT',\n zip_safe=False,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n ],\n test_suite='nose.collector'\n)\n","subject":"Bump dependency to suds-jurko 0.6.","message":"Bump dependency to suds-jurko 0.6.\n","lang":"Python","license":"mit","repos":"OnroerendErfgoed\/crabpy"} {"commit":"dee908d28734f5dba0a98e19edc39bc35c9bb062","old_file":"setup.py","new_file":"setup.py","old_contents":"#! \/usr\/bin\/env python\nfrom ez_setup import use_setuptools\nuse_setuptools()\nfrom setuptools import setup, find_packages\n\n\nsetup(name='permamodel',\n version='0.1.0',\n author='Elchin Jafarov and Scott Stewart',\n author_email='james.stewart@colorado.edu',\n description='Permamodel',\n long_description=open('README.md').read(),\n packages=find_packages(),\n #install_requires=('numpy', 'nose', 'gdal', 'pyproj'),\n install_requires=('numpy', 'nose',),\n package_data={'': ['examples\/*.cfg', 'examples\/*.dat']}\n)\n","new_contents":"#! \/usr\/bin\/env python\nfrom ez_setup import use_setuptools\nuse_setuptools()\nfrom setuptools import setup, find_packages\n\n\nsetup(name='permamodel',\n version='0.1.0',\n author='Elchin Jafarov and Scott Stewart',\n author_email='james.stewart@colorado.edu',\n description='Permamodel',\n long_description=open('README.md').read(),\n packages=find_packages(),\n #install_requires=('numpy', 'nose', 'gdal', 'pyproj'),\n install_requires=('affine', 'netCDF4', 'scipy', 'numpy', 'nose',),\n package_data={'': ['examples\/*.cfg', 'examples\/*.dat']}\n)\n","subject":"Include packages needed on install","message":"Include packages needed on install\n","lang":"Python","license":"mit","repos":"permamodel\/permamodel,permamodel\/permamodel"} {"commit":"d2b4e85fd0b3c44a460bc843eb480dd82f216f6e","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\nfrom codecs import open\nfrom os import path\n\nhere = path.abspath(path.dirname(__file__))\n\n# Get the long description from the README file\nwith open(path.join(here, 'flask_swagger_ui\/README.md'), encoding='utf-8') as f:\n long_description = f.read()\n\nsetup(\n name='flask-swagger-ui',\n version='3.0.12',\n description='Swagger UI blueprint for Flask',\n long_description=long_description,\n\n url='https:\/\/github.com\/sveint\/flask-swagger-ui',\n\n author='Svein Tore Koksrud Seljebotn',\n author_email='sveint@gmail.com',\n license='MIT',\n\n classifiers=[\n 'Development Status :: 4 - Beta',\n\n 'Intended Audience :: Developers',\n\n 'License :: OSI Approved :: MIT License',\n\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n ],\n\n keywords='flask swagger',\n packages=['flask_swagger_ui'],\n\n package_data={\n 'flask_swagger_ui': [\n 'README.md',\n 'templates\/*.html',\n 'dist\/VERSION',\n 'dist\/LICENSE',\n 'dist\/README.md',\n 'dist\/*.html',\n 'dist\/*.js',\n 'dist\/*\/*.js',\n 'dist\/*\/*.css',\n 'dist\/*\/*.gif',\n 'dist\/*\/*.png',\n 'dist\/*\/*.ico',\n 'dist\/*\/*.ttf',\n ],\n }\n)\n","new_contents":"from setuptools import setup\nfrom codecs import open\nfrom os import path\n\nhere = path.abspath(path.dirname(__file__))\n\n# Get the long description from the README file\nwith open(path.join(here, 'flask_swagger_ui\/README.md'), encoding='utf-8') as f:\n long_description = f.read()\n\nsetup(\n name='flask-swagger-ui',\n version='3.0.12a',\n description='Swagger UI blueprint for Flask',\n long_description=long_description,\n\n url='https:\/\/github.com\/sveint\/flask-swagger-ui',\n\n author='Svein Tore Koksrud Seljebotn',\n author_email='sveint@gmail.com',\n license='MIT',\n\n classifiers=[\n 'Development Status :: 4 - Beta',\n\n 'Intended Audience :: Developers',\n\n 'License :: OSI Approved :: MIT License',\n\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n ],\n\n keywords='flask swagger',\n packages=['flask_swagger_ui'],\n\n package_data={\n 'flask_swagger_ui': [\n 'README.md',\n 'templates\/*.html',\n 'dist\/VERSION',\n 'dist\/LICENSE',\n 'dist\/README.md',\n 'dist\/*.html',\n 'dist\/*.js',\n 'dist\/*.css',\n 'dist\/*.png'\n ],\n }\n)\n","subject":"Fix file inclusion and make new release.","message":"Fix file inclusion and make new release.\n","lang":"Python","license":"mit","repos":"sveint\/flask-swagger-ui,sveint\/flask-swagger-ui,sveint\/flask-swagger-ui"} {"commit":"ade992a1cb078d0c3bd67453307b5ebf7e3c21e5","old_file":"setup.py","new_file":"setup.py","old_contents":"from distutils.core import setup\r\nfrom os import path\r\ntry:\r\n import setuptools # noqa, setuptools namespace\r\nexcept Exception:\r\n pass\r\n\r\nCLASSIFIERS = [\r\n 'Development Status :: 4 - Beta',\r\n 'Intended Audience :: Science\/Research',\r\n 'Programming Language :: Python',\r\n 'Topic :: Scientific\/Engineering',\r\n 'Topic :: System :: Hardware',\r\n 'License :: OSI Approved :: GNU General Public License (GPL)'\r\n]\r\n\r\nhere = path.dirname(path.abspath(__file__))\r\nwith open(path.join(here, 'README.rst')) as f:\r\n long_description = f.read().strip()\r\n\r\nlong_description += '''\r\n\r\nSource code: http:\/\/github.com\/LABSN\/tdtpy\r\n\r\nDocumentation: http:\/\/tdtpy.readthedocs.org\r\n\r\n'''\r\n\r\nsetup(\r\n name='TDTPy',\r\n version='0.8.dev0',\r\n author='Brad Buran',\r\n author_email='bburan@alum.mit.edu',\r\n packages=['tdt', 'tdt.actxobjects', 'tdt.device'],\r\n url='http:\/\/bradburan.com\/programs\/tdtpy',\r\n license='GPLv3',\r\n description='Module for communicating with TDT\\'s System 3 hardware',\r\n long_description=long_description,\r\n requires=['win32com', 'six'],\r\n package_data={'tdt': ['components\/*.rcx']},\r\n classifiers=CLASSIFIERS,\r\n)\r\n","new_contents":"from distutils.core import setup\r\nimport tdt\r\nfrom os import path\r\ntry:\r\n import setuptools # noqa, setuptools namespace\r\nexcept Exception:\r\n pass\r\n\r\nCLASSIFIERS = [\r\n 'Development Status :: 4 - Beta',\r\n 'Intended Audience :: Science\/Research',\r\n 'Programming Language :: Python',\r\n 'Topic :: Scientific\/Engineering',\r\n 'Topic :: System :: Hardware',\r\n 'License :: OSI Approved'\r\n]\r\n\r\nhere = path.dirname(path.abspath(__file__))\r\nwith open(path.join(here, 'README.rst')) as f:\r\n long_description = f.read().strip()\r\n\r\nlong_description += '''\r\n\r\nSource code: http:\/\/github.com\/LABSN\/tdtpy\r\n\r\nDocumentation: http:\/\/tdtpy.readthedocs.org\r\n\r\n'''\r\n\r\nsetup(\r\n name='TDTPy',\r\n version='0.8.dev0',\r\n author='Brad Buran',\r\n author_email='bburan@alum.mit.edu',\r\n packages=['tdt', 'tdt.actxobjects', 'tdt.device'],\r\n url='http:\/\/tdtpy.readthedocs.org',\r\n license='BSD (3-clause)',\r\n description='Module for communicating with TDT\\'s System 3 hardware',\r\n long_description=long_description,\r\n requires=['win32com', 'six'],\r\n package_data={'tdt': ['components\/*.rcx']},\r\n classifiers=CLASSIFIERS,\r\n)","subject":"Update version, license, and website","message":"Update version, license, and website\n","lang":"Python","license":"bsd-3-clause","repos":"LABSN\/tdtpy"} {"commit":"3263a691db55ed72c4f98096748ad930c7ecdd68","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os.path\nfrom distutils.core import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), \"README.rst\")).read()\n\nCLASSIFIERS = [\n\t\"Development Status :: 5 - Production\/Stable\",\n\t\"Intended Audience :: Developers\",\n\t\"License :: OSI Approved :: MIT License\",\n\t\"Programming Language :: Python\",\n\t\"Topic :: Software Development :: Libraries :: Python Modules\",\n]\n\nimport ptch\nVERSION = ptch.__version__\n\nsetup(\n\tname = \"python-ptch\",\n\tpy_modules = [\"ptch\"],\n\tauthor = \"Jerome Leclanche\",\n\tauthor_email = \"jerome.leclanche+python-ptch@gmail.com\",\n\tclassifiers = CLASSIFIERS,\n\tdescription = \"Blizzard BSDIFF-based PTCH file format support\",\n\tdownload_url = \"http:\/\/github.com\/Adys\/python-ptch\/tarball\/master\",\n\tlong_description = README,\n\turl = \"http:\/\/github.com\/Adys\/python-ptch\",\n\tversion = VERSION,\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os.path\nfrom distutils.core import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), \"README.rst\")).read()\n\nCLASSIFIERS = [\n\t\"Development Status :: 5 - Production\/Stable\",\n\t\"Intended Audience :: Developers\",\n\t\"License :: OSI Approved :: MIT License\",\n\t\"Programming Language :: Python\",\n\t\"Topic :: Software Development :: Libraries :: Python Modules\",\n]\n\nimport ptch\nVERSION = ptch.__version__\n\nsetup(\n\tname = \"python-ptch\",\n\tpy_modules = [\"ptch\"],\n\tauthor = \"Jerome Leclanche\",\n\tauthor_email = \"jerome@leclan.ch\",\n\tclassifiers = CLASSIFIERS,\n\tdescription = \"Blizzard BSDIFF-based PTCH file format support\",\n\tdownload_url = \"https:\/\/github.com\/jleclanche\/python-ptch\/tarball\/master\",\n\tlong_description = README,\n\turl = \"https:\/\/github.com\/jleclanche\/python-ptch\",\n\tversion = VERSION,\n)\n","subject":"Update repository addresses and emails","message":"Update repository addresses and emails\n","lang":"Python","license":"mit","repos":"jleclanche\/python-ptch"} {"commit":"48701a9f582d65f8086b2bdefe02315d6aca1e77","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom setuptools import setup, find_packages\n\n\ndef read(filename):\n return open(os.path.join(os.path.dirname(__file__), filename)).read()\n\n\nsetup(\n name='gears-stylus',\n version='0.1.1',\n url='https:\/\/github.com\/gears\/gears-stylus',\n license='ISC',\n author='Mike Yumatov',\n author_email='mike@yumatov.org',\n description='Stylus compiler for Gears',\n long_description=read('README.rst'),\n packages=find_packages(),\n include_package_data=True,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: ISC License (ISCL)',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.5',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n ],\n)\n","new_contents":"import os\nfrom setuptools import setup, find_packages\n\n\ndef read(filename):\n return open(os.path.join(os.path.dirname(__file__), filename)).read()\n\n\nsetup(\n name='gears-stylus',\n version='0.1.1',\n url='https:\/\/github.com\/gears\/gears-stylus',\n license='ISC',\n author='Mike Yumatov',\n author_email='mike@yumatov.org',\n description='Stylus compiler for Gears',\n long_description=read('README.rst'),\n packages=find_packages(),\n include_package_data=True,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: ISC License (ISCL)',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n ],\n)\n","subject":"Drop Python 2.5 support, add support for Python 3.2","message":"Drop Python 2.5 support, add support for Python 3.2\n","lang":"Python","license":"isc","repos":"gears\/gears-stylus,gears\/gears-stylus"} {"commit":"9abbb0b79da1466d2719496b479e43a74e798b97","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\nsetup(\n name=\"librobinson\",\n version=\"0.1\",\n packages=find_packages(),\n scripts=['robinson'],\n\n # Project uses reStructuredText, so ensure that the docutils get\n # installed or upgraded on the target machine\n #install_requires=['docutils>=0.3'],\n\n package_data={\n # If any package contains *.txt or *.md files, include them:\n '': ['*.txt', '*.md'],\n # And include any *.msg files found in the 'hello' package, too:\n #'hello': ['*.msg'],\n },\n\n # metadata for upload to PyPI\n author=\"Ulrik Sandborg-Petersen\",\n author_email=\"ulrikp@scripturesys.com\",\n description=\"A library to parse and convert the New Testament Greek files of Dr. Maurice A. Robinson\",\n license=\"MIT\",\n keywords=\"Maurice A. Robinson, New Testament Greek, parse, convert\",\n url=\"http:\/\/github.com\/byztxt\/librobinson\"\n\n # could also include long_description, download_url, classifiers, etc.\n)\n","new_contents":"from setuptools import setup, find_packages\nsetup(\n name=\"librobinson\",\n version=\"0.2.0\",\n packages=find_packages(),\n scripts=[\n 'robinson\/booknames.py',\n 'robinson\/book.py',\n 'robinson\/chapter.py',\n 'robinson\/convert.py',\n 'robinson\/__init__.py',\n 'robinson\/kind.py',\n 'robinson\/reader.py',\n 'robinson\/readwhat.py',\n 'robinson\/robinson.py',\n 'robinson\/robinsontags.py',\n 'robinson\/variant.py',\n 'robinson\/verse.py',\n 'robinson\/word.py',\n ],\n\n # Project uses reStructuredText, so ensure that the docutils get\n # installed or upgraded on the target machine\n #install_requires=['docutils>=0.3'],\n\n package_data={\n # If any package contains *.txt or *.md files, include them:\n '': ['*.txt', '*.md'],\n # And include any *.msg files found in the 'hello' package, too:\n #'hello': ['*.msg'],\n },\n\n # metadata for upload to PyPI\n author=\"Ulrik Sandborg-Petersen\",\n author_email=\"ulrikp@scripturesys.com\",\n description=\"A library to parse and convert the New Testament Greek files of Dr. Maurice A. Robinson\",\n license=\"MIT\",\n keywords=\"Maurice A. Robinson, New Testament Greek, parse, convert\",\n url=\"http:\/\/github.com\/byztxt\/librobinson\"\n\n # could also include long_description, download_url, classifiers, etc.\n)\n","subject":"Add all python files explicitly, and bump to version 0.2.0","message":"Add all python files explicitly, and bump to version 0.2.0\n","lang":"Python","license":"mit","repos":"byztxt\/librobinson"} {"commit":"b90553ddc7a27d2b594fcc88130d999c70ae6f5b","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom setuptools import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name = 'RecordExpress',\n version = '0.0',\n packages = ['collection_record'],\n include_package_data = True,\n dependency_links = ['https:\/\/github.com\/cdlib\/RecordExpress.git'],\n license = 'BSD License - see LICENSE file', \n description = 'A lightweight EAD creator',\n long_description = README,\n author = 'Mark Redar',\n author_email = 'mark.redar@ucop.edu',\n classifiers = [\n 'Environment :: Web Environment',\n 'Development Status :: 3 - Alpha',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n install_requires = [\n 'django>=1.4',\n 'django-dublincore>=0.1',\n 'django-sortable',\n 'BeautifulSoup',\n 'webtest',\n 'django-webtest'\n ],\n)\n\n","new_contents":"import os\nfrom setuptools import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name = 'RecordExpress',\n version = '0.0',\n packages = ['collection_record'],\n include_package_data = True,\n dependency_links = ['https:\/\/github.com\/cdlib\/RecordExpress.git',\n 'https:\/\/github.com\/drewyeaton\/django-sortable\/archive\/master.zip#egg=django-sortable', #pypi package currently broken - 2013\/09\n ], \n license = 'BSD License - see LICENSE file', \n description = 'A lightweight EAD creator',\n long_description = README,\n author = 'Mark Redar',\n author_email = 'mark.redar@ucop.edu',\n classifiers = [\n 'Environment :: Web Environment',\n 'Development Status :: 3 - Alpha',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n install_requires = [\n 'django>=1.4',\n 'django-dublincore>=0.1',\n 'django-sortable',\n 'BeautifulSoup',\n 'webtest',\n 'django-webtest'\n ],\n)\n\n","subject":"Make django-sortable install, pypi package is broken.","message":"Make django-sortable install, pypi package is broken.\n","lang":"Python","license":"bsd-3-clause","repos":"cdlib\/RecordExpress,cdlib\/RecordExpress,cdlib\/RecordExpress,cdlib\/RecordExpress"} {"commit":"2fa8b2f4a63579633272b1cc8d972baf27c661f2","old_file":"pmg\/models\/__init__.py","new_file":"pmg\/models\/__init__.py","old_contents":"from .users import *\nfrom .resources import *\nfrom .emails import *\nfrom .pages import *\nfrom .soundcloud_track import *\n","new_contents":"from .users import *\nfrom .resources import *\nfrom .emails import *\nfrom .pages import *\nfrom .soundcloud_track import SoundcloudTrack\n","subject":"Fix error on admin user_report view","message":"Fix error on admin user_report view\n","lang":"Python","license":"apache-2.0","repos":"Code4SA\/pmg-cms-2,Code4SA\/pmg-cms-2,Code4SA\/pmg-cms-2"} {"commit":"6a9193fdc43361ca12b7f22d18954d17c2049ba1","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\r\nsetup(\r\n name = 'CommonModules',\r\n packages = find_packages(where = '.'), # this must be the same as the name above\r\n version = '0.1.11',\r\n description = 'Common Python modules\/functionalities used in practice.',\r\n author = 'Wang Hewen',\r\n author_email = 'wanghewen2@sina.com',\r\n url = 'https:\/\/github.com\/wanghewen\/CommonModules', # use the URL to the github repo\r\n keywords = ['library'], # arbitrary keywords\r\n license='MIT',\r\n install_requires=[],\r\n\textras_require = {\r\n 'Advance DataStructureOperations': ['scipy', 'numpy'],\r\n\t\t'Advance DataStructure IO': ['networkx', 'numpy', 'scipy']\r\n }\r\n)","new_contents":"from setuptools import setup, find_packages\r\nsetup(\r\n name = 'CommonModules',\r\n packages = find_packages(where = '.'), # this must be the same as the name above\r\n version = '0.1.13',\r\n description = 'Common Python modules\/functionalities used in practice.',\r\n author = 'Wang Hewen',\r\n author_email = 'wanghewen2@sina.com',\r\n url = 'https:\/\/github.com\/wanghewen\/CommonModules', # use the URL to the github repo\r\n keywords = ['library'], # arbitrary keywords\r\n license='MIT',\r\n install_requires=[],\r\n\textras_require = {\r\n 'Advance DataStructureOperations': ['scipy', 'numpy'],\r\n\t\t'Advance DataStructure IO': ['networkx', 'numpy', 'scipy']\r\n }\r\n)","subject":"Fix version number. Will check the file in another machine and fix the conflict","message":"Fix version number. Will check the file in another machine and fix the conflict\n","lang":"Python","license":"mit","repos":"wanghewen\/CommonModules"} {"commit":"ad0f91e9d120e4c6b34aabf13fa3c68f0d7f5611","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nfrom setuptools import setup, find_packages\nsetup(\n name = \"cobe\",\n version = \"0.5\",\n author = \"Peter Teichman\",\n author_email = \"peter@teichman.org\",\n packages = [\"cobe\"],\n test_suite = \"tests.cobe_suite\",\n install_requires = [\"cmdparse>=0.9\"],\n classifiers = [\n \"Development Status :: 4 - Beta\",\n \"Environment :: Console\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: End Users\/Desktop\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Topic :: Scientific\/Engineering :: Artificial Intelligence\"\n ],\n entry_points = {\n \"console_scripts\" : [\n \"cobe-control = cobe.control:main\"\n ]\n }\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nfrom setuptools import setup, find_packages\nsetup(\n name = \"cobe\",\n version = \"0.5\",\n author = \"Peter Teichman\",\n author_email = \"peter@teichman.org\",\n packages = [\"cobe\"],\n test_suite = \"tests.cobe_suite\",\n install_requires = [\"cmdparse>=0.9\"],\n classifiers = [\n \"Development Status :: 4 - Beta\",\n \"Environment :: Console\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: End Users\/Desktop\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Topic :: Scientific\/Engineering :: Artificial Intelligence\"\n ],\n entry_points = {\n \"console_scripts\" : [\n \"cobe = cobe.control:main\"\n ]\n }\n)\n","subject":"Rename the control script to \"cobe\"","message":"Rename the control script to \"cobe\"\n","lang":"Python","license":"mit","repos":"pteichman\/cobe,wodim\/cobe-ng,tiagochiavericosta\/cobe,LeMagnesium\/cobe,DarkMio\/cobe,meska\/cobe,pteichman\/cobe,meska\/cobe,DarkMio\/cobe,wodim\/cobe-ng,tiagochiavericosta\/cobe,LeMagnesium\/cobe"} {"commit":"518cafbd053843b0aee48ac75eccb6a05ff237f5","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\nimport proxyprefix\n\n\nsetup(\n name='proxyprefix',\n version=proxyprefix.__version__,\n description=proxyprefix.__doc__,\n url='https:\/\/github.com\/yola\/proxyprefix',\n packages=['proxyprefix'],\n)\n","new_contents":"from setuptools import setup\n\nimport proxyprefix\n\n\nsetup(\n name='proxyprefix',\n version=proxyprefix.__version__,\n description='Prefix SCRIPT_NAME with X-Forwarded-Prefix header',\n long_description=proxyprefix.__doc__,\n url='https:\/\/github.com\/yola\/proxyprefix',\n packages=['proxyprefix'],\n)\n","subject":"Use module doc as long_description not description","message":"Use module doc as long_description not description\n\nIt's too long for description.\nSee https:\/\/github.com\/yola\/proxyprefix\/pull\/3#issuecomment-75107125\n","lang":"Python","license":"mit","repos":"yola\/proxyprefix"} {"commit":"678e872de192b09c1bafc7a26dc67d7737a14e20","old_file":"altair\/examples\/us_population_over_time.py","new_file":"altair\/examples\/us_population_over_time.py","old_contents":"\"\"\"\nUS Population Over Time\n=======================\nThis chart visualizes the age distribution of the US population over time.\nIt uses a slider widget that is bound to the year to visualize the age\ndistribution over time.\n\"\"\"\n# category: case studies\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.population.url\n\npink_blue = alt.Scale(domain=('Male', 'Female'),\n range=[\"steelblue\", \"salmon\"])\n\nslider = alt.binding_range(min=1900, max=2000, step=10)\nselect_year = alt.selection_single(name=\"year\", fields=['year'],\n bind=slider, init={'year': 2000})\n\nalt.Chart(source).mark_bar().encode(\n x=alt.X('sex:N', title=None),\n y=alt.Y('people:Q', scale=alt.Scale(domain=(0, 12000000))),\n color=alt.Color('sex:N', scale=pink_blue),\n column='age:O'\n).properties(\n width=20\n).add_selection(\n select_year\n).transform_calculate(\n \"sex\", alt.expr.if_(alt.datum.sex == 1, \"Male\", \"Female\")\n).transform_filter(\n select_year\n).configure_facet(\n spacing=8\n)\n","new_contents":"\"\"\"\nUS Population by Age and Sex\n============================\nThis chart visualizes the age distribution of the US population over time.\nIt uses a slider widget that is bound to the year to visualize the age\ndistribution over time.\n\"\"\"\n# category: case studies\nimport altair as alt\nfrom vega_datasets import data\n\nsource = data.population.url\n\nselect_year = alt.selection_single(\n name=\"Year\",\n fields=[\"year\"],\n bind=alt.binding_range(min=1900, max=2000, step=10, name=\"Year\"),\n init={\"year\": 2000},\n)\n\nalt.Chart(source).mark_bar().encode(\n x=alt.X(\"sex:N\", axis=alt.Axis(labels=False, title=None, ticks=False)),\n y=alt.Y(\"people:Q\", scale=alt.Scale(domain=(0, 12000000)), title=\"Population\"),\n color=alt.Color(\n \"sex:N\",\n scale=alt.Scale(domain=(\"Male\", \"Female\"), range=[\"steelblue\", \"salmon\"]),\n title=\"Sex\",\n ),\n column=alt.Column(\"age:O\", title=\"Age\"),\n).properties(width=20, title=\"U.S. Population by Age and Sex\").add_selection(\n select_year\n).transform_calculate(\n \"sex\", alt.expr.if_(alt.datum.sex == 1, \"Male\", \"Female\")\n).transform_filter(\n select_year\n).configure_facet(\n spacing=8\n)\n","subject":"Tidy up U.S. Population by Age and Sex","message":"Tidy up U.S. Population by Age and Sex","lang":"Python","license":"bsd-3-clause","repos":"altair-viz\/altair"} {"commit":"1ff696abdee762303dfc79e23e0cdabc6e411270","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n#from distutils.core import setup\nfrom setuptools import setup, find_packages\n\nsetup(\n name=\"imaplib2\",\n version=\"2.28.3\",\n description=\"A threaded Python IMAP4 client.\",\n author=\"Piers Lauder\",\n url=\"http:\/\/github.com\/bcoe\/imaplib2\",\n classifiers = [\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\"\n ],\n packages = find_packages()\n)\n","new_contents":"#!\/usr\/bin\/env python\n# from distutils.core import setup\nfrom setuptools import setup, find_packages\n\ntry:\n from distutils.command.build_py import build_py_2to3 as build_py\nexcept ImportError:\n from distutils.command.build_py import build_py\n\nsetup(\n name=\"imaplib2\",\n version=\"2.28.4\",\n description=\"A threaded Python IMAP4 client.\",\n author=\"Piers Lauder\",\n url=\"http:\/\/github.com\/bcoe\/imaplib2\",\n classifiers=[\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\"\n ],\n packages=find_packages(),\n cmdclass={'build_py': build_py}\n)\n","subject":"Add support for Python 3 by doing 2to3 conversion when installing the package with distutils. This way we don't have to maintain two separate repositories to support Python 2.x and Python 3.x.","message":"Add support for Python 3 by doing 2to3 conversion when installing the package with distutils. This way we don't have to maintain two separate repositories to support Python 2.x and Python 3.x.\n","lang":"Python","license":"mit","repos":"mbmccoy\/smtp_to_tcp"} {"commit":"71b68c990977e78abf8dbaf6562719f39492657f","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\n\n# First update the version in loompy\/_version.py, then:\n\n# cd loompy (the root loompy folder, not the one inside!)\n# rm -r dist (otherwise twine will upload the oldest build!)\n# python setup.py sdist\n# twine upload dist\/*\n\n# NOTE: Don't forget to update the release version at loompy.github.io (index.html)!\n\n\n\n# pylint: disable=exec-used\n__version__ = '0.0.0'\nexec(open('loompy\/_version.py').read())\n\nsetup(\n\tname=\"loompy\",\n\tversion=__version__,\n\tpackages=find_packages(),\n\tinstall_requires=['h5py', 'numpy', 'scipy', \"typing\", \"setuptools\"],\n\t# metadata for upload to PyPI\n\tauthor=\"Linnarsson Lab\",\n\tauthor_email=\"sten.linnarsson@ki.se\",\n\tdescription=\"Work with .loom files for single-cell RNA-seq data\",\n\tlicense=\"BSD\",\n\tkeywords=\"loom omics transcriptomics bioinformatics\",\n\turl=\"https:\/\/github.com\/linnarsson-lab\/loompy\",\n\tdownload_url=f\"https:\/\/github.com\/linnarsson-lab\/loompy\/archive\/{__version__}.tar.gz\",\n)\n","new_contents":"from setuptools import setup, find_packages\n\n# First update the version in loompy\/_version.py, then:\n\n# cd loompy (the root loompy folder, not the one inside!)\n# rm -r dist (otherwise twine will upload the oldest build!)\n# python setup.py sdist\n# twine upload dist\/*\n\n# NOTE: Don't forget to update the release version at loompy.github.io (index.html)!\n\n\n\n# pylint: disable=exec-used\n__version__ = '0.0.0'\nexec(open('loompy\/_version.py').read())\n\nsetup(\n\tname=\"loompy\",\n\tversion=__version__,\n\tpackages=find_packages(),\n\tinstall_requires=['h5py', 'numpy', 'scipy', \"typing\", \"setuptools\"],\n\tpython_requires='>=3.6',\n\t# metadata for upload to PyPI\n\tauthor=\"Linnarsson Lab\",\n\tauthor_email=\"sten.linnarsson@ki.se\",\n\tdescription=\"Work with .loom files for single-cell RNA-seq data\",\n\tlicense=\"BSD\",\n\tkeywords=\"loom omics transcriptomics bioinformatics\",\n\turl=\"https:\/\/github.com\/linnarsson-lab\/loompy\",\n\tdownload_url=f\"https:\/\/github.com\/linnarsson-lab\/loompy\/archive\/{__version__}.tar.gz\",\n)\n","subject":"Make clear we need Python 3.6","message":"Make clear we need Python 3.6\n","lang":"Python","license":"bsd-2-clause","repos":"linnarsson-lab\/loompy,linnarsson-lab\/loompy"} {"commit":"1d686d4e5cd4ff610dda2b8be9fc747d6314a4b4","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nimport os\nfrom setuptools import setup, find_packages\n\nhere = os.path.dirname(__file__)\n\nversion_file = os.path.join(here, 'src\/iptools\/__init__.py')\nd = {}\nexecfile(version_file, d)\nversion = d['__version__']\n\nsetup(\n name = 'iptools',\n version = version,\n description = 'Python utilites for manipulating IP addresses',\n long_description = \"Utilities for manipulating IP addresses including a class that can be used to include CIDR network blocks in Django's INTERNAL_IPS setting.\",\n url = 'http:\/\/python-iptools.googlecode.com',\n author = 'Bryan Davis',\n author_email = 'casadebender+iptools@gmail.com',\n license = 'BSD',\n platforms = ['any',],\n package_dir = {'': 'src'},\n packages = find_packages('src'),\n include_package_data = True,\n test_suite='iptools.test_iptools',\n classifiers = [\n 'Development Status :: 4 - Beta', \n 'Environment :: Web Environment',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 3',\n 'Topic :: Utilities',\n 'Topic :: Internet', \n ],\n zip_safe=False,\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nimport os\nfrom setuptools import setup, find_packages\n\nhere = os.path.dirname(__file__)\n\nversion_file = os.path.join(here, 'src\/iptools\/__init__.py')\nd = {}\nexecfile(version_file, d)\nversion = d['__version__']\n\nsetup(\n name = 'iptools',\n version = version,\n description = 'Python utilites for manipulating IP addresses',\n long_description = \"Utilities for manipulating IP addresses including a class that can be used to include CIDR network blocks in Django's INTERNAL_IPS setting.\",\n url = 'http:\/\/python-iptools.googlecode.com',\n download_url = 'http:\/\/pypi.python.org\/packages\/source\/i\/iptools\/',\n author = 'Bryan Davis',\n author_email = 'casadebender+iptools@gmail.com',\n license = 'BSD',\n platforms = ['any',],\n package_dir = {'': 'src'},\n packages = find_packages('src'),\n include_package_data = True,\n test_suite='iptools.test_iptools',\n classifiers = [\n 'Development Status :: 4 - Beta', \n 'Environment :: Web Environment',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 3',\n 'Topic :: Utilities',\n 'Topic :: Internet', \n ],\n zip_safe=False,\n)\n","subject":"Set download_url to pypi directory.","message":"Set download_url to pypi directory.\n\ngit-svn-id: c8188841f5432f3fe42d04dee4f87e556eb5cf84@23 99efc558-b41a-11dd-8714-116ca565c52f\n","lang":"Python","license":"bsd-2-clause","repos":"bd808\/python-iptools"} {"commit":"0d4fe588023869044755644dfa162c488a7fdea8","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\n\n\nsetup(\n name='raimon49.guestbook',\n version='1.0.0',\n packages=find_packages(),\n include_package_data=True,\n install_requires=[\n 'Flask',\n ],\n entry_points=\"\"\"\n [console_scripts]\n guestbook = guestbook:main\n \"\"\"\n)\n","new_contents":"import os\nfrom setuptools import setup, find_packages\n\n\ndef read_file(filename):\n basepath = os.path.dirname(os.path.dirname(__file__))\n filepath = os.path.join(basepath, filename)\n if os.path.exists(filepath):\n return open(filepath.read())\n else:\n return ''\n\n\nsetup(\n name='raimon49.guestbook',\n version='1.0.0',\n description='A guestbook web application.',\n long_description=read_file('README.rst'),\n author='raimon49',\n author_email='raimon49@hotmail.com',\n url='https:\/\/github.com\/raimon49\/pypro2-guestbook-webapp',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Framework :: Flask',\n 'License :: OSI Approved :: BSD License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n ]\n packages=find_packages(),\n include_package_data=True,\n keywords=['web', 'guestbook'],\n License='BSD License',\n install_requires=[\n 'Flask',\n ],\n entry_points=\"\"\"\n [console_scripts]\n guestbook = guestbook:main\n \"\"\"\n)\n","subject":"Update meta data for distributed PyPI","message":"Update meta data for distributed PyPI\n","lang":"Python","license":"bsd-3-clause","repos":"raimon49\/pypro2-guestbook-webapp,raimon49\/pypro2-guestbook-webapp"} {"commit":"42418b638a581b0243182e8a4e24662c7e7cc003","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/python\n\nimport setuptools\nfrom setuptools import find_packages\n\nsetuptools.setup(\n name = 'js.handlebars',\n version = '1.0.rc.1',\n license = 'BSD',\n description = 'Fanstatic package for Handlebars.js',\n long_description = open('README.txt').read(),\n author = 'Matt Good',\n author_email = 'matt@matt-good.net',\n url = 'http:\/\/github.com\/mgood\/js.handlebars\/',\n platforms = 'any',\n packages=find_packages(),\n namespace_packages=['js'],\n zip_safe = False,\n install_requires=[\n 'fanstatic',\n ],\n entry_points={\n 'fanstatic.libraries': [\n 'handlebars = js.handlebars:library',\n ],\n },\n)\n","new_contents":"#!\/usr\/bin\/python\n\nimport setuptools\nfrom setuptools import find_packages\n\nsetuptools.setup(\n name = 'js.handlebars',\n version = '1.0.rc.1-1',\n license = 'BSD',\n description = 'Fanstatic package for Handlebars.js',\n long_description = open('README.txt').read(),\n author = 'Matt Good',\n author_email = 'matt@matt-good.net',\n url = 'http:\/\/github.com\/mgood\/js.handlebars\/',\n platforms = 'any',\n packages=find_packages(),\n namespace_packages=['js'],\n include_package_data=True,\n zip_safe = False,\n install_requires=[\n 'fanstatic',\n ],\n entry_points={\n 'fanstatic.libraries': [\n 'handlebars = js.handlebars:library',\n ],\n },\n)\n","subject":"Fix including JS resources as package data","message":"Fix including JS resources as package data\n","lang":"Python","license":"bsd-2-clause","repos":"mgood\/js.handlebars,mgood\/js.handlebars"} {"commit":"a7f24ba803c13bf7b263aed4d974ad604d53df2f","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup, find_packages\n\n__author__ = \"Nitrax <nitrax@lokisec.fr>\"\n__copyright__ = \"Copyright 2017, Legobot\"\n\ndescription = 'Lego providing networking tools'\nname = 'legos.nettools'\nsetup(\n name=name,\n version='0.1.0',\n namespace_packages=name.split('.')[:-1],\n license='MIT',\n description=description,\n author='Nitrax',\n url='https:\/\/github.com\/Legobot\/' + name,\n install_requires=['legobot>=1.1.4,<=2.0.0',\n 'python-whois',\n 'urllib3',\n 'bandit==1.3.0',\n 'flake8==3.2.1',\n 'pytest==3.0.5'\n ],\n classifiers=[\n 'License :: MIT',\n\n 'Programming Language :: Python :: 3'\n ],\n packages=find_packages()\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup, find_packages\n\n__author__ = \"Nitrax <nitrax@lokisec.fr>\"\n__copyright__ = \"Copyright 2017, Legobot\"\n\ndescription = 'Lego providing networking tools'\nname = 'legos.nettools'\nsetup(\n name=name,\n version='0.1.0',\n namespace_packages=name.split('.')[:-1],\n license='MIT',\n description=description,\n author='Nitrax',\n url='https:\/\/github.com\/Legobot\/' + name,\n install_requires=['legobot>=1.1.4,<=2.0.0',\n 'python-whois',\n 'urllib3',\n 'bandit==1.3.0',\n 'flake8==3.2.1',\n 'pytest==3.0.5'\n ],\n classifiers=[\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3'\n ],\n packages=find_packages()\n)\n","subject":"Fix trove classifier for pypi","message":"Fix trove classifier for pypi\n","lang":"Python","license":"mit","repos":"Legobot\/legos.nettools"} {"commit":"f7784f2023e8f351c539586c56d2f9ec3a9086e1","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nfrom setuptools import setup\n\n\ninstall_requires = [\n 'argparse',\n 'jsonschema',\n 'mock',\n 'M2Crypto',\n 'pycrypto',\n 'python-augeas',\n 'python2-pythondialog',\n 'requests',\n]\n\ndocs_extras = [\n 'Sphinx',\n]\n\ntesting_extras = [\n 'coverage',\n 'nose',\n 'nosexcover',\n 'pylint',\n 'tox',\n]\n\nsetup(\n name=\"letsencrypt\",\n version=\"0.1\",\n description=\"Let's Encrypt\",\n author=\"Let's Encrypt Project\",\n license=\"\",\n url=\"https:\/\/letsencrypt.org\",\n packages=[\n 'letsencrypt',\n 'letsencrypt.client',\n 'letsencrypt.scripts',\n ],\n install_requires=install_requires,\n tests_require=install_requires,\n test_suite='letsencrypt',\n extras_require={\n 'docs': docs_extras,\n 'testing': testing_extras,\n },\n entry_points={\n 'console_scripts': [\n 'letsencrypt = letsencrypt.scripts.main:main',\n ],\n },\n zip_safe=False,\n include_package_data=True,\n)\n","new_contents":"#!\/usr\/bin\/env python\nfrom setuptools import setup\n\n\ninstall_requires = [\n 'argparse',\n 'jsonschema',\n 'M2Crypto',\n 'mock',\n 'pycrypto',\n 'python-augeas',\n 'python2-pythondialog',\n 'requests',\n]\n\ndocs_extras = [\n 'Sphinx',\n]\n\ntesting_extras = [\n 'coverage',\n 'nose',\n 'nosexcover',\n 'pylint',\n 'tox',\n]\n\nsetup(\n name=\"letsencrypt\",\n version=\"0.1\",\n description=\"Let's Encrypt\",\n author=\"Let's Encrypt Project\",\n license=\"\",\n url=\"https:\/\/letsencrypt.org\",\n packages=[\n 'letsencrypt',\n 'letsencrypt.client',\n 'letsencrypt.scripts',\n ],\n install_requires=install_requires,\n tests_require=install_requires,\n test_suite='letsencrypt',\n extras_require={\n 'docs': docs_extras,\n 'testing': testing_extras,\n },\n entry_points={\n 'console_scripts': [\n 'letsencrypt = letsencrypt.scripts.main:main',\n ],\n },\n zip_safe=False,\n include_package_data=True,\n)\n","subject":"Fix lexicographic order in install_requires","message":"Fix lexicographic order in install_requires\n","lang":"Python","license":"apache-2.0","repos":"Jadaw1n\/letsencrypt,TheBoegl\/letsencrypt,Sveder\/letsencrypt,tdfischer\/lets-encrypt-preview,letsencrypt\/letsencrypt,hsduk\/lets-encrypt-preview,mrb\/letsencrypt,bsmr-misc-forks\/letsencrypt,mrb\/letsencrypt,beermix\/letsencrypt,PeterMosmans\/letsencrypt,lbeltrame\/letsencrypt,wteiken\/letsencrypt,bsmr-misc-forks\/letsencrypt,TheBoegl\/letsencrypt,martindale\/letsencrypt,goofwear\/letsencrypt,letsencrypt\/letsencrypt,riseofthetigers\/letsencrypt,sjerdo\/letsencrypt,rlustin\/letsencrypt,diracdeltas\/lets-encrypt-preview,jtl999\/certbot,tdfischer\/lets-encrypt-preview,luorenjin\/letsencrypt,lmcro\/letsencrypt,deserted\/letsencrypt,rutsky\/letsencrypt,dietsche\/letsencrypt,BillKeenan\/lets-encrypt-preview,VladimirTyrin\/letsencrypt,stewnorriss\/letsencrypt,sapics\/letsencrypt,BillKeenan\/lets-encrypt-preview,beermix\/letsencrypt,lmcro\/letsencrypt,Sveder\/letsencrypt,skynet\/letsencrypt,sjerdo\/letsencrypt,sapics\/letsencrypt,hsduk\/lets-encrypt-preview,Hasimir\/letsencrypt,deserted\/letsencrypt,stewnorriss\/letsencrypt,brentdax\/letsencrypt,tyagi-prashant\/letsencrypt,mitnk\/letsencrypt,VladimirTyrin\/letsencrypt,ghyde\/letsencrypt,solidgoldbomb\/letsencrypt,jmaurice\/letsencrypt,stweil\/letsencrypt,armersong\/letsencrypt,modulexcite\/letsencrypt,vcavallo\/letsencrypt,goofwear\/letsencrypt,jmaurice\/letsencrypt,hlieberman\/letsencrypt,kevinlondon\/letsencrypt,g1franc\/lets-encrypt-preview,Jadaw1n\/letsencrypt,ahojjati\/letsencrypt,jmhodges\/letsencrypt,bestwpw\/letsencrypt,luorenjin\/letsencrypt,jtl999\/certbot,twstrike\/le_for_patching,fmarier\/letsencrypt,tyagi-prashant\/letsencrypt,piru\/letsencrypt,modulexcite\/letsencrypt,jsha\/letsencrypt,solidgoldbomb\/letsencrypt,kevinlondon\/letsencrypt,jmhodges\/letsencrypt,rugk\/letsencrypt,xgin\/letsencrypt,brentdax\/letsencrypt,digideskio\/lets-encrypt-preview,ruo91\/letsencrypt,thanatos\/lets-encrypt-preview,Jonadabe\/letsencrypt,jsha\/letsencrypt,stweil\/letsencrypt,Bachmann1234\/letsencrypt,riseofthetigers\/letsencrypt,kuba\/letsencrypt,Jonadabe\/letsencrypt,xgin\/letsencrypt,Hasimir\/letsencrypt,BKreisel\/letsencrypt,lbeltrame\/letsencrypt,DavidGarciaCat\/letsencrypt,ruo91\/letsencrypt,Bachmann1234\/letsencrypt,piru\/letsencrypt,bestwpw\/letsencrypt,g1franc\/lets-encrypt-preview,twstrike\/le_for_patching,PeterMosmans\/letsencrypt,skynet\/letsencrypt,hlieberman\/letsencrypt,digideskio\/lets-encrypt-preview,martindale\/letsencrypt,mitnk\/letsencrypt,wteiken\/letsencrypt,diracdeltas\/lets-encrypt-preview,vcavallo\/letsencrypt,dietsche\/letsencrypt,ahojjati\/letsencrypt,thanatos\/lets-encrypt-preview,fmarier\/letsencrypt,ghyde\/letsencrypt,kuba\/letsencrypt,DavidGarciaCat\/letsencrypt,rutsky\/letsencrypt,armersong\/letsencrypt,rugk\/letsencrypt,BKreisel\/letsencrypt,rlustin\/letsencrypt"} {"commit":"84ded02cba3caee164e848c1200e46b08011f93f","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\n\nversion = '1.0.17'\n\nrequires = [\n 'neo4j-driver<1.2.0',\n 'six>=1.10.0',\n]\n\ntesting_requires = [\n 'nose',\n 'coverage',\n 'nosexcover',\n]\n\nsetup(\n name='norduniclient',\n version=version,\n url='https:\/\/github.com\/NORDUnet\/python-norduniclient',\n license='Apache License, Version 2.0',\n author='Johan Lundberg',\n author_email='lundberg@nordu.net',\n description='Neo4j (>=3.2.2) database client using bolt for NORDUnet network inventory',\n packages=find_packages(),\n zip_safe=False,\n install_requires=requires,\n tests_require=testing_requires,\n test_suite='nose.collector',\n extras_require={\n 'testing': testing_requires\n }\n)\n","new_contents":"from setuptools import setup, find_packages\n\nversion = '1.0.17'\n\nrequires = [\n 'neo4j-driver<1.5,0',\n 'six>=1.10.0',\n]\n\ntesting_requires = [\n 'nose',\n 'coverage',\n 'nosexcover',\n]\n\nsetup(\n name='norduniclient',\n version=version,\n url='https:\/\/github.com\/NORDUnet\/python-norduniclient',\n license='Apache License, Version 2.0',\n author='Johan Lundberg',\n author_email='lundberg@nordu.net',\n description='Neo4j (>=3.2.2) database client using bolt for NORDUnet network inventory',\n packages=find_packages(),\n zip_safe=False,\n install_requires=requires,\n tests_require=testing_requires,\n test_suite='nose.collector',\n extras_require={\n 'testing': testing_requires\n }\n)\n","subject":"Update requirement neo4j-driver to <1.5.0","message":"Update requirement neo4j-driver to <1.5.0\n","lang":"Python","license":"apache-2.0","repos":"NORDUnet\/python-norduniclient,NORDUnet\/python-norduniclient"} {"commit":"beedabe50213cff1d8dc723822b35850437b8b25","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nfrom setuptools import setup\n\ntry:\n with open('README') as f:\n long_description = f.read()\nexcept:\n long_description = ''\n\nsetup(\n name='crontab',\n version='0.20',\n description='Parse and use crontab schedules in Python',\n author='Josiah Carlson',\n author_email='josiah.carlson@gmail.com',\n url='https:\/\/github.com\/josiahcarlson\/parse-crontab',\n packages=['crontab'],\n test_suite=\"tests\",\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'License :: OSI Approved :: GNU Library or Lesser General Public License (LGPL)',\n 'License :: OSI Approved :: GNU Lesser General Public License v2 (LGPLv2)',\n 'License :: OSI Approved :: GNU Lesser General Public License v3 (LGPLv3)',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n ],\n license='GNU LGPL v2.1',\n long_description=long_description,\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nfrom setuptools import setup\n\ntry:\n with open('README') as f:\n long_description = f.read()\nexcept:\n long_description = ''\n\nsetup(\n name='crontab',\n version='0.20.1',\n description='Parse and use crontab schedules in Python',\n author='Josiah Carlson',\n author_email='josiah.carlson@gmail.com',\n url='https:\/\/github.com\/josiahcarlson\/parse-crontab',\n packages=['crontab'],\n test_suite=\"tests\",\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'License :: OSI Approved :: GNU Library or Lesser General Public License (LGPL)',\n 'License :: OSI Approved :: GNU Lesser General Public License v2 (LGPLv2)',\n 'License :: OSI Approved :: GNU Lesser General Public License v3 (LGPLv3)',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n ],\n license='GNU LGPL v2.1',\n long_description=long_description,\n)\n","subject":"Bump version for push to pypi","message":"Bump version for push to pypi\n","lang":"Python","license":"lgpl-2.1","repos":"josiahcarlson\/parse-crontab"} {"commit":"8476597698e6f21404e784b43f21f01d93c5b576","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nimport nirvana\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nwith open('README.rst') as readme:\n long_description = readme.read()\n\nsetup(\n name='nirvana',\n version=nirvana.__version__,\n description=('Library for interacting with the Nirvana task manager '\n '(nirvanahq.com)'),\n long_description=long_description,\n author='Nick Wilson',\n author_email='nick@njwilson.net',\n url='http:\/\/github.com\/njwilson\/nirvana-python',\n license='MIT',\n packages=['nirvana'],\n classifiers=[\n 'Development Status :: 1 - Planning',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nimport nirvana\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nwith open('README.rst') as readme:\n long_description = readme.read()\n\nsetup(\n name='nirvana',\n version=nirvana.__version__,\n description=('Library for interacting with the Nirvana task manager '\n '(nirvanahq.com)'),\n long_description=long_description,\n author='Nick Wilson',\n author_email='nick@njwilson.net',\n url='http:\/\/nirvana-python.readthedocs.org',\n license='MIT',\n packages=['nirvana'],\n classifiers=[\n 'Development Status :: 1 - Planning',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n)\n","subject":"Update project URL to readthedocs","message":"Update project URL to readthedocs\n","lang":"Python","license":"mit","repos":"njwilson\/nirvana-python,njwilson\/nirvana-python"} {"commit":"902635648187847fe1f167981c55476e6b5db907","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\nsetup(\n name=\"teamscale-client\",\n version=\"3.0.0\",\n author=\"Thomas Kinnen - CQSE GmbH\",\n author_email=\"kinnen@cqse.eu\",\n description=(\"A simple service client to interact with Teamscale's REST API.\"),\n license=\"Apache\",\n keywords=\"rest api teamscale\",\n url=\"https:\/\/github.com\/cqse\/teamscale-client-python\",\n packages=['teamscale_client'],\n long_description=\"A simple service client to interact with Teamscale's REST API.\",\n classifiers=[\n \"Topic :: Utilities\",\n ],\n install_requires=[\n 'simplejson',\n 'requests>=2.0',\n 'jsonpickle'\n ],\n\n tests_require=[\n 'pytest',\n 'responses'\n ],\n setup_requires=[\"pytest-runner\"]\n)\n","new_contents":"from setuptools import setup\n\nsetup(\n name=\"teamscale-client\",\n version=\"3.1.0\",\n author=\"Thomas Kinnen - CQSE GmbH\",\n author_email=\"kinnen@cqse.eu\",\n description=(\"A simple service client to interact with Teamscale's REST API.\"),\n license=\"Apache\",\n keywords=\"rest api teamscale\",\n url=\"https:\/\/github.com\/cqse\/teamscale-client-python\",\n packages=['teamscale_client'],\n long_description=\"A simple service client to interact with Teamscale's REST API.\",\n classifiers=[\n \"Topic :: Utilities\",\n ],\n install_requires=[\n 'simplejson',\n 'requests>=2.0',\n 'jsonpickle'\n ],\n\n tests_require=[\n 'pytest',\n 'responses'\n ],\n setup_requires=[\"pytest-runner\"]\n)\n","subject":"Update version for 3.1.0 release","message":"Update version for 3.1.0 release\n","lang":"Python","license":"apache-2.0","repos":"cqse\/teamscale-client-python"} {"commit":"dd063b68311209c51018cad7e9c91d2c6b4eef3c","old_file":"setup.py","new_file":"setup.py","old_contents":"# coding:utf-8\n\nfrom setuptools import setup, find_packages\nfrom qingstor.qsctl import __version__\n\nsetup(\n name='qsctl',\n version=__version__,\n description='Advanced command line tool for QingStor.',\n long_description=open('README.rst', 'rb').read().decode('utf-8'),\n keywords='yunify qingcloud qingstor qsctl object_storage',\n author='QingStor Dev Team',\n author_email='qs-devel@yunify.com',\n url='https:\/\/www.qingstor.com',\n scripts=['bin\/qsctl', 'bin\/qsctl.cmd'],\n packages=find_packages('.'),\n package_dir={'qsctl': 'qingstor'},\n namespace_packages=['qingstor'],\n include_package_data=True,\n install_requires=[\n 'argparse >= 1.1',\n 'PyYAML >= 3.1',\n 'qingstor-sdk >= 2.1.0',\n 'docutils >= 0.10',\n 'tqdm >= 4.0.0'\n ])\n","new_contents":"# coding:utf-8\n\nfrom sys import version_info\nfrom setuptools import setup, find_packages\nfrom qingstor.qsctl import __version__\n\ninstall_requires = [\n 'argparse >= 1.1',\n 'PyYAML >= 3.1',\n 'qingstor-sdk >= 2.1.0',\n 'docutils >= 0.10',\n 'tqdm >= 4.0.0'\n]\n\nif version_info[:3] < (2, 7, 9):\n install_requires.append(\"requests[security]\")\n\nsetup(\n name='qsctl',\n version=__version__,\n description='Advanced command line tool for QingStor.',\n long_description=open('README.rst', 'rb').read().decode('utf-8'),\n keywords='yunify qingcloud qingstor qsctl object_storage',\n author='QingStor Dev Team',\n author_email='qs-devel@yunify.com',\n url='https:\/\/www.qingstor.com',\n scripts=['bin\/qsctl', 'bin\/qsctl.cmd'],\n packages=find_packages('.'),\n package_dir={'qsctl': 'qingstor'},\n namespace_packages=['qingstor'],\n include_package_data=True,\n install_requires=install_requires\n)\n","subject":"Fix SSL Warnings with old python versions","message":"Fix SSL Warnings with old python versions\n\nSigned-off-by: Xuanwo <9d9ffaee821234cdfed458cf06eb6f407f8dbe47@yunify.com>\n","lang":"Python","license":"apache-2.0","repos":"yunify\/qsctl,Fiile\/qsctl"} {"commit":"8692eef51cf9b77aa0d6d09eec4bc4f36850d902","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\n\nsetup(\n name='django-force-logout',\n\n url=\"https:\/\/chris-lamb.co.uk\/projects\/django-force-logout\",\n version='2.0.0',\n description=\"Framework to be able to forcibly log users out of Django projects\",\n\n author=\"Chris Lamb\",\n author_email='chris@chris-lamb.co.uk',\n license=\"BSD\",\n\n packages=find_packages(),\n)\n","new_contents":"from setuptools import setup, find_packages\n\nsetup(\n name='django-force-logout',\n\n url=\"https:\/\/chris-lamb.co.uk\/projects\/django-force-logout\",\n version='2.0.0',\n description=\"Framework to be able to forcibly log users out of Django projects\",\n\n author=\"Chris Lamb\",\n author_email='chris@chris-lamb.co.uk',\n license=\"BSD\",\n\n packages=find_packages(),\n\n install_requires=(\n 'Django>=1.8',\n ),\n)\n","subject":"Update Django requirement to latest LTS","message":"Update Django requirement to latest LTS\n","lang":"Python","license":"bsd-3-clause","repos":"lamby\/django-force-logout"} {"commit":"15e66e00be22d7177fcba292720f55f548839469","old_file":"packages\/dependencies\/intel_quicksync_mfx.py","new_file":"packages\/dependencies\/intel_quicksync_mfx.py","old_contents":"{\n\t'repo_type' : 'git',\n\t'url' : 'https:\/\/github.com\/lu-zero\/mfx_dispatch.git',\n\t'conf_system' : 'cmake',\n\t'source_subfolder' : '_build',\n\t'configure_options' : '.. {cmake_prefix_options} -DCMAKE_INSTALL_PREFIX={target_prefix} -DBUILD_SHARED_LIBS=0 -DCMAKE_BUILD_TYPE=Release',\n\t'_info' : { 'version' : None, 'fancy_name' : 'intel_quicksync_mfx' },\n}","new_contents":"{\n\t'repo_type' : 'git',\n\t'do_not_bootstrap' : True,\n\t'run_post_patch' : [\n\t\t'autoreconf -fiv',\n\t],\n\t'patches' :\t[\n\t\t( 'mfx\/mfx-0001-mingwcompat-disable-va.patch', '-p1' ),\n\t],\n\t'url' : 'https:\/\/github.com\/lu-zero\/mfx_dispatch.git',\n\t'configure_options' : '{autoconf_prefix_options} --without-libva_drm --without-libva_x11',\n\t'_info' : { 'version' : None, 'fancy_name' : 'intel_quicksync_mfx' },\n}","subject":"Revert \"packages\/quicksync-mfx: switch to cmake\"","message":"Revert \"packages\/quicksync-mfx: switch to cmake\"\n\nThis reverts commit b3db211b42f26480fe817d26d7515ec8bd6e5c9e.\n","lang":"Python","license":"mpl-2.0","repos":"DeadSix27\/python_cross_compile_script"} {"commit":"5ffdfd7eb103d6974c3fb782eecaf457f53c972f","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom distutils.core import setup\n\nversion = '0.9.3'\n\n\ndef read_file(name):\n return open(os.path.join(os.path.dirname(__file__),\n name)).read()\n\nreadme = read_file('README.rst')\nchanges = read_file('CHANGES')\n\nsetup(\n name='django-maintenancemode',\n version=version,\n description='Django-maintenancemode allows you to temporary shutdown your site for maintenance work',\n long_description='\\n\\n'.join([readme, changes]),\n author='Remco Wendt',\n author_email='remco@maykinmedia.nl',\n license=\"BSD\",\n platforms=[\"any\"],\n url='https:\/\/github.com\/shanx\/django-maintenancemode',\n packages=[\n 'maintenancemode',\n 'maintenancemode.conf',\n 'maintenancemode.conf.settings',\n 'maintenancemode.conf.urls',\n 'maintenancemode.tests',\n 'maintenancemode.views',\n ],\n package_data={\n 'maintenancemode': [\n 'tests\/templates\/503.html',\n ],\n },\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Utilities',\n ],\n)\n","new_contents":"import os\nfrom distutils.core import setup\n\nversion = '0.9.3'\n\nhere = os.path.abspath(os.path.dirname(__file__))\n\ndef read_file(name):\n return open(os.path.join(here, name)).read()\n\nreadme = read_file('README.rst')\nchanges = read_file('CHANGES')\n\nsetup(\n name='django-maintenancemode',\n version=version,\n description='Django-maintenancemode allows you to temporary shutdown your site for maintenance work',\n long_description='\\n\\n'.join([readme, changes]),\n author='Remco Wendt',\n author_email='remco@maykinmedia.nl',\n license=\"BSD\",\n platforms=[\"any\"],\n url='https:\/\/github.com\/shanx\/django-maintenancemode',\n packages=[\n 'maintenancemode',\n 'maintenancemode.conf',\n 'maintenancemode.conf.settings',\n 'maintenancemode.conf.urls',\n 'maintenancemode.tests',\n 'maintenancemode.views',\n ],\n package_data={\n 'maintenancemode': [\n 'tests\/templates\/503.html',\n ],\n },\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Utilities',\n ],\n)\n","subject":"Use the absolute path for the long description to work around CI issues.","message":"Use the absolute path for the long description to work around CI issues.","lang":"Python","license":"bsd-3-clause","repos":"aarsan\/django-maintenancemode,shanx\/django-maintenancemode,aarsan\/django-maintenancemode,21strun\/django-maintenancemode,shanx\/django-maintenancemode,21strun\/django-maintenancemode"} {"commit":"1224552892d1d459864d5ab2dada328a20cc66e7","old_file":"jobs\/spiders\/tvinna.py","new_file":"jobs\/spiders\/tvinna.py","old_contents":"import dateutil.parser\nimport scrapy.spiders\n\nfrom jobs.items import JobsItem\n\n\nclass TvinnaSpider(scrapy.spiders.XMLFeedSpider):\n name = \"tvinna\"\n start_urls = ['http:\/\/www.tvinna.is\/feed\/?post_type=job_listing']\n itertag = 'item'\n namespaces = [\n ('atom', 'http:\/\/www.w3.org\/2005\/Atom'),\n ('content', 'http:\/\/purl.org\/rss\/1.0\/modules\/content\/'),\n ('dc', 'http:\/\/purl.org\/dc\/elements\/1.1\/'),\n ('slash', 'http:\/\/purl.org\/rss\/1.0\/modules\/slash\/'),\n ('sy', 'http:\/\/purl.org\/rss\/1.0\/modules\/syndication\/'),\n ('wfw', 'http:\/\/wellformedweb.org\/CommentAPI\/'),\n ]\n\n def parse_node(self, response, node):\n item = JobsItem()\n item['spider'] = self.name\n item['title'] = node.xpath('title\/text()').extract_first()\n item['company'] = node.xpath('dc:creator\/text()').extract_first()\n item['url'] = node.xpath('link\/text()').extract_first()\n time_posted = node.xpath('pubDate\/text()').extract_first()\n item['posted'] = dateutil.parser.parse(time_posted).isoformat()\n return item\n","new_contents":"import dateutil.parser\nimport scrapy\nimport scrapy.spiders\n\nfrom jobs.items import JobsItem\n\n\nclass TvinnaSpider(scrapy.spiders.XMLFeedSpider):\n name = \"tvinna\"\n start_urls = ['http:\/\/www.tvinna.is\/feed\/?post_type=job_listing']\n itertag = 'item'\n namespaces = [\n ('atom', 'http:\/\/www.w3.org\/2005\/Atom'),\n ('content', 'http:\/\/purl.org\/rss\/1.0\/modules\/content\/'),\n ('dc', 'http:\/\/purl.org\/dc\/elements\/1.1\/'),\n ('slash', 'http:\/\/purl.org\/rss\/1.0\/modules\/slash\/'),\n ('sy', 'http:\/\/purl.org\/rss\/1.0\/modules\/syndication\/'),\n ('wfw', 'http:\/\/wellformedweb.org\/CommentAPI\/'),\n ]\n\n def parse_node(self, response, node):\n item = JobsItem()\n item['spider'] = self.name\n item['title'] = node.xpath('title\/text()').extract_first()\n item['url'] = url = node.xpath('link\/text()').extract_first()\n time_posted = node.xpath('pubDate\/text()').extract_first()\n item['posted'] = dateutil.parser.parse(time_posted).isoformat()\n\n request = scrapy.Request(url, callback=self.parse_specific_job)\n request.meta['item'] = item\n yield request\n\n def parse_specific_job(self, response):\n item = response.meta['item']\n item['company'] = response.css('.company a::text').extract_first()\n yield item\n","subject":"Fix the extraction of the company name.","message":"Fix the extraction of the company name.\n\nThere's an apparent bug in the Tvinna rss feed, such that the username of the person creating the listing is used in place of a company name in the `<cd:creator>` field. As a work around, we need to visit the job listing page, and extract it from that instead. It requires more requests, but yields more accurate results.\n","lang":"Python","license":"apache-2.0","repos":"multiplechoice\/workplace"} {"commit":"d2189fef32470b3dce05b4765a40b501174bd0d4","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\nfrom setuptools import find_packages\n\n\nsetup(name='Keras',\n version='0.1.2',\n description='Theano-based Deep Learning library',\n author='Francois Chollet',\n author_email='francois.chollet@gmail.com',\n url='https:\/\/github.com\/fchollet\/keras',\n download_url='https:\/\/github.com\/fchollet\/keras\/tarball\/0.1.2',\n license='MIT',\n install_requires=['theano', 'pyyaml'],\n extras_require = {\n 'h5py': ['h5py'],\n },\n packages=find_packages())\n","new_contents":"from setuptools import setup\nfrom setuptools import find_packages\n\n\nsetup(name='Keras',\n version='0.2.0',\n description='Theano-based Deep Learning library',\n author='Francois Chollet',\n author_email='francois.chollet@gmail.com',\n url='https:\/\/github.com\/fchollet\/keras',\n download_url='https:\/\/github.com\/fchollet\/keras\/tarball\/0.2.0',\n license='MIT',\n install_requires=['theano', 'pyyaml'],\n extras_require={\n 'h5py': ['h5py'],\n },\n packages=find_packages())\n","subject":"Update version number: now 0.2.0","message":"Update version number: now 0.2.0\n","lang":"Python","license":"mit","repos":"dolaameng\/keras,kuza55\/keras,imcomking\/Convolutional-GRU-keras-extension-,johmathe\/keras,relh\/keras,keras-team\/keras,daviddiazvico\/keras,keras-team\/keras,DeepGnosis\/keras,nebw\/keras,kemaswill\/keras,jimgoo\/keras"} {"commit":"166015e9b4cad5b9d00df31e0d242c335c93ab79","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\n# Standard library modules.\nimport os\n\n# Third party modules.\nfrom setuptools import setup, find_packages\n\n# Local modules.\nimport versioneer\n\n# Globals and constants variables.\nBASEDIR = os.path.abspath(os.path.dirname(__file__))\n\n# Get the long description from the relevant file\nwith open(os.path.join(BASEDIR, 'README.rst'), 'r') as f:\n long_description = f.read()\n\nsetup(name='matplotlib-scalebar',\n version=versioneer.get_version(),\n description='Artist for matplotlib to display a scale bar',\n long_description=long_description,\n\n author='Philippe Pinard',\n author_email='philippe.pinard@gmail.com',\n maintainer='Philippe Pinard',\n maintainer_email='philippe.pinard@gmail.com',\n\n url='https:\/\/github.com\/ppinard\/matplotlib-scalebar',\n license='BSD',\n keywords='matplotlib scale micron bar',\n\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Science\/Research',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python :: 3',\n 'Topic :: Scientific\/Engineering :: Visualization'\n ],\n\n packages=find_packages(),\n package_data={},\n\n install_requires=['matplotlib'],\n\n zip_safe=True,\n\n test_suite='nose.collector',\n\n cmdclass=versioneer.get_cmdclass(),\n\n )\n","new_contents":"#!\/usr\/bin\/env python\n\n# Standard library modules.\nfrom pathlib import Path\n\n# Third party modules.\nfrom setuptools import setup, find_packages\n\n# Local modules.\nimport versioneer\n\n# Globals and constants variables.\nBASEDIR = Path(__file__).parent.resolve()\n\n# Get the long description from the relevant file\nwith open(BASEDIR.joinpath(\"README.rst\"), \"r\") as f:\n long_description = f.read()\n\nsetup(\n name=\"matplotlib-scalebar\",\n version=versioneer.get_version(),\n description=\"Artist for matplotlib to display a scale bar\",\n long_description=long_description,\n author=\"Philippe Pinard\",\n author_email=\"philippe.pinard@gmail.com\",\n maintainer=\"Philippe Pinard\",\n maintainer_email=\"philippe.pinard@gmail.com\",\n url=\"https:\/\/github.com\/ppinard\/matplotlib-scalebar\",\n license=\"BSD\",\n keywords=\"matplotlib scale micron bar\",\n classifiers=[\n \"Development Status :: 4 - Beta\",\n \"Intended Audience :: Science\/Research\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python :: 3\",\n \"Topic :: Scientific\/Engineering :: Visualization\",\n ],\n packages=find_packages(),\n package_data={},\n install_requires=[\"matplotlib\"],\n zip_safe=True,\n test_suite=\"nose.collector\",\n cmdclass=versioneer.get_cmdclass(),\n)\n","subject":"Use pathlib instead of os.path","message":"Use pathlib instead of os.path","lang":"Python","license":"bsd-2-clause","repos":"ppinard\/matplotlib-scalebar"} {"commit":"def2ab4a860a48222fa26ede36c9a47622aa5209","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n#\n# Setup script for Django Evolution\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nfrom setuptools import setup, find_packages\nfrom setuptools.command.test import test\n\nfrom django_evolution import get_package_version, VERSION\n\n\ndef run_tests(*args):\n import os\n os.system('tests\/runtests.py')\n\ntest.run_tests = run_tests\n\n\nPACKAGE_NAME = 'django_evolution'\n\ndownload_url = (\n 'http:\/\/downloads.reviewboard.org\/releases\/django-evolution\/%s.%s\/' %\n (VERSION[0], VERSION[1]))\n\n\n# Build the package\nsetup(\n name=PACKAGE_NAME,\n version=get_package_version(),\n description='A database schema evolution tool for the Django web framework.',\n url='http:\/\/code.google.com\/p\/django-evolution\/',\n author='Ben Khoo',\n author_email='khoobks@westnet.com.au',\n maintainer='Christian Hammond',\n maintainer_email='christian@beanbaginc.com',\n download_url=download_url,\n packages=find_packages(exclude=['tests']),\n install_requires=[\n 'Django>=1.4.10,<1.7.0',\n ],\n include_package_data=True,\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ]\n)\n","new_contents":"#!\/usr\/bin\/env python\n#\n# Setup script for Django Evolution\n\nfrom ez_setup import use_setuptools\nuse_setuptools()\n\nfrom setuptools import setup, find_packages\nfrom setuptools.command.test import test\n\nfrom django_evolution import get_package_version, VERSION\n\n\ndef run_tests(*args):\n import os\n os.system('tests\/runtests.py')\n\ntest.run_tests = run_tests\n\n\nPACKAGE_NAME = 'django_evolution'\n\ndownload_url = (\n 'http:\/\/downloads.reviewboard.org\/releases\/django-evolution\/%s.%s\/' %\n (VERSION[0], VERSION[1]))\n\n\n# Build the package\nsetup(\n name=PACKAGE_NAME,\n version=get_package_version(),\n description='A database schema evolution tool for the Django web framework.',\n url='http:\/\/code.google.com\/p\/django-evolution\/',\n author='Ben Khoo',\n author_email='khoobks@westnet.com.au',\n maintainer='Christian Hammond',\n maintainer_email='christian@beanbaginc.com',\n download_url=download_url,\n packages=find_packages(exclude=['tests']),\n install_requires=[\n 'Django>=1.4.10',\n ],\n include_package_data=True,\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ]\n)\n","subject":"Allow Django Evolution to install along with Django >= 1.7.","message":"Allow Django Evolution to install along with Django >= 1.7.\n\nAs we're working toward some degree of compatibility with newer versions\nof Django, we need to ease up on the version restriction. Now's a good\ntime to do so. Django Evolution no longer has an upper bounds on the\nversion range.\n","lang":"Python","license":"bsd-3-clause","repos":"beanbaginc\/django-evolution"} {"commit":"aa4498eea07bd3a0c09a11782f881312020d725d","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\n\nfrom setuptools import setup\n\ndef read(*paths):\n \"\"\"Build a file path from *paths* and return the contents.\"\"\"\n with open(os.path.join(*paths), 'r') as f:\n return f.read()\n\nsetup(\n name='chrome-webstore-deploy',\n version='0.0.1',\n description='Automate deployment of Chrome extensions\/apps to Chrome Web Store.',\n long_description=(read('README.rst') + '\\n\\n'),\n url='http:\/\/github.com\/jonnor\/chrome-webstore-deploy\/',\n license='MIT',\n author='Jon Nordby',\n author_email='jononor@gmail.com',\n# py_modules=['foo'],\n scripts=['bin\/chrome-webstore.py'],\n install_requires=[\n \"oauth2client >= 1.2\",\n \"httplib2 >= 0.9\",\n ],\n include_package_data=True,\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n)\n","new_contents":"import os\n\nfrom setuptools import setup\n\ndef read(*paths):\n \"\"\"Build a file path from *paths* and return the contents.\"\"\"\n with open(os.path.join(*paths), 'r') as f:\n return f.read()\n\nsetup(\n name='chrome-webstore-deploy',\n version='0.0.2',\n description='Automate deployment of Chrome extensions\/apps to Chrome Web Store.',\n long_description=(read('README.rst') + '\\n\\n'),\n url='http:\/\/github.com\/jonnor\/chrome-webstore-deploy\/',\n license='MIT',\n author='Jon Nordby',\n author_email='jononor@gmail.com',\n# py_modules=['foo'],\n scripts=['bin\/chrome-web-store.py'],\n install_requires=[\n \"oauth2client >= 1.2\",\n \"httplib2 >= 0.9\",\n ],\n include_package_data=True,\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n)\n","subject":"Fix typo in script packaging","message":"Fix typo in script packaging\n","lang":"Python","license":"mit","repos":"jonnor\/chrome-webstore-deploy"} {"commit":"838d83df29b905110f8bd317e08eaaa64e97f402","old_file":"setup.py","new_file":"setup.py","old_contents":"# -*- coding: utf-8 -*-\nimport os\nfrom setuptools import setup\n\ndef read(fname):\n try:\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n except:\n return ''\n\nsetup(\n name='todoist-python',\n version='0.2.18',\n packages=['todoist', 'todoist.managers'],\n author='Doist Team',\n author_email='info@todoist.com',\n license='BSD',\n description='todoist-python - The official Todoist Python API library',\n long_description = read('README.md'),\n install_requires=[\n 'requests',\n ],\n # see here for complete list of classifiers\n # http:\/\/pypi.python.org\/pypi?%3Aaction=list_classifiers\n classifiers=(\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Programming Language :: Python',\n ),\n)\n","new_contents":"# -*- coding: utf-8 -*-\nimport os\nfrom setuptools import setup\n\ndef read(fname):\n try:\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n except:\n return ''\n\nsetup(\n name='todoist-python',\n version='0.2.19',\n packages=['todoist', 'todoist.managers'],\n author='Doist Team',\n author_email='info@todoist.com',\n license='BSD',\n description='todoist-python - The official Todoist Python API library',\n long_description = read('README.md'),\n install_requires=[\n 'requests',\n ],\n # see here for complete list of classifiers\n # http:\/\/pypi.python.org\/pypi?%3Aaction=list_classifiers\n classifiers=(\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Programming Language :: Python',\n ),\n)\n","subject":"Update the PyPI version to 0.2.19.","message":"Update the PyPI version to 0.2.19.\n","lang":"Python","license":"mit","repos":"Doist\/todoist-python"} {"commit":"d56f3d77b8e9883df0b9c4199f74ca36b39f44ef","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import find_packages\nfrom setuptools import setup\n\n\ndef get_long_description():\n with open('README.md') as readme_file:\n return readme_file.read()\n\n\nsetup(\n name=\"RouterOS-api\",\n version='0.16.1.dev0',\n description='Python API to RouterBoard devices produced by MikroTik.',\n long_description=get_long_description(),\n long_description_content_type='text\/markdown',\n author='Social WiFi',\n author_email='it@socialwifi.com',\n url='https:\/\/github.com\/socialwifi\/RouterOS-api',\n packages=find_packages(),\n test_suite=\"tests\",\n license=\"MIT\",\n install_requires=['six'],\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n ],\n)\n","new_contents":"import sys\n\nfrom setuptools import find_packages\nfrom setuptools import setup\n\nrequirements = ['six']\n\nif sys.version_info[0] < 3 or (sys.version_info[0] == 3 and sys.version_info[1] < 3):\n requirements.append('ipaddress')\n\ndef get_long_description():\n with open('README.md') as readme_file:\n return readme_file.read()\n\n\nsetup(\n name=\"RouterOS-api\",\n version='0.16.1.dev0',\n description='Python API to RouterBoard devices produced by MikroTik.',\n long_description=get_long_description(),\n long_description_content_type='text\/markdown',\n author='Social WiFi',\n author_email='it@socialwifi.com',\n url='https:\/\/github.com\/socialwifi\/RouterOS-api',\n packages=find_packages(),\n test_suite=\"tests\",\n license=\"MIT\",\n install_requires=requirements,\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n ],\n)\n","subject":"Add ipaddress requirement to python older than 3.3","message":"Add ipaddress requirement to python older than 3.3\n","lang":"Python","license":"mit","repos":"socialwifi\/RouterOS-api,pozytywnie\/RouterOS-api"} {"commit":"8c8e09709037079598c0d690efb0fabf67961989","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom setuptools import setup, find_packages\n\nREADME = open(os.path.join(os.path.dirname(__file__), \"README.rst\")).read()\n\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name=\"recipe_scrapers\",\n url=\"https:\/\/github.com\/hhursev\/recipe-scrapers\/\",\n version=\"8.0.0\",\n author=\"Hristo Harsev\",\n author_email=\"r+pypi@hharsev.com\",\n description=\"Python package, scraping recipes from all over the internet\",\n keywords=\"python recipes scraper harvest recipe-scraper recipe-scrapers\",\n long_description=README,\n install_requires=[\n \"beautifulsoup4>=4.6.0\",\n \"extruct>=0.8.0\",\n \"language-tags>=1.0.0\",\n \"requests>=2.19.1\",\n \"tldextract==2.2.2\",\n ],\n packages=find_packages(),\n package_data={\"\": [\"LICENSE\"]},\n include_package_data=True,\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: MIT License\",\n \"Intended Audience :: Developers\",\n \"Operating System :: OS Independent\",\n 'Topic :: Internet :: WWW\/HTTP',\n ],\n python_requires='>=3.5'\n)\n","new_contents":"import os\nfrom setuptools import setup, find_packages\n\nREADME = open(os.path.join(os.path.dirname(__file__), \"README.rst\")).read()\n\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name=\"recipe_scrapers\",\n url=\"https:\/\/github.com\/hhursev\/recipe-scrapers\/\",\n version=\"8.0.1\",\n author=\"Hristo Harsev\",\n author_email=\"r+pypi@hharsev.com\",\n description=\"Python package, scraping recipes from all over the internet\",\n keywords=\"python recipes scraper harvest recipe-scraper recipe-scrapers\",\n long_description=README,\n install_requires=[\n \"beautifulsoup4>=4.6.0\",\n \"extruct>=0.8.0\",\n \"language-tags>=1.0.0\",\n \"requests>=2.19.1\",\n \"tldextract==2.2.2\",\n ],\n packages=find_packages(),\n package_data={\"\": [\"LICENSE\"]},\n include_package_data=True,\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: MIT License\",\n \"Intended Audience :: Developers\",\n \"Operating System :: OS Independent\",\n 'Topic :: Internet :: WWW\/HTTP',\n ],\n python_requires='>=3.5'\n)\n","subject":"Bump to 8.0.1 to verify auto pypi deployment is working","message":"Bump to 8.0.1 to verify auto pypi deployment is working\n","lang":"Python","license":"mit","repos":"hhursev\/recipe-scraper"} {"commit":"5e3253217a971a996aa182f199c8f1413aa7bc40","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import find_packages, setup\n\n\nVERSION = '1.0.0'\n\nsetup(\n name='django-fakery',\n version=VERSION,\n url='https:\/\/github.com\/fcurella\/django-factory\/',\n author='Flavio Curella',\n author_email='flavio.curella@gmail.com',\n description='A model instances generator for Django',\n license='MIT',\n packages=find_packages(exclude=['*.tests']),\n platforms=[\"any\"],\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Framework :: Django',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n ],\n install_requires=[\n \"fake-factory==0.5.3\",\n \"Django>=1.7\",\n ],\n test_suite='django_fakery.tests.runtests.runtests',\n)\n","new_contents":"from setuptools import find_packages, setup\n\n\nVERSION = '1.0.0'\n\nsetup(\n name='django-fakery',\n version=VERSION,\n url='https:\/\/github.com\/fcurella\/django-factory\/',\n author='Flavio Curella',\n author_email='flavio.curella@gmail.com',\n description='A model instances generator for Django',\n license='MIT',\n packages=find_packages(exclude=['*.tests']),\n platforms=[\"any\"],\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Framework :: Django',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n ],\n install_requires=[\n \"fake-factory==0.5.3\",\n \"Django>=1.7\",\n ],\n test_suite='django_fakery.tests.runtests.runtests',\n)\n","subject":"Remove trove classifier for python 3.5","message":"Remove trove classifier for python 3.5\n","lang":"Python","license":"mit","repos":"fcurella\/django-fakery"} {"commit":"7e0c61aa54dd26760aba0d78926b599d5b8f6d5f","old_file":"tests\/__init__.py","new_file":"tests\/__init__.py","old_contents":"","new_contents":"# This file needs to exist in order for pytest-cov to work.\n# See this: https:\/\/bitbucket.org\/memedough\/pytest-cov\/issues\/4\/no-coverage-unless-test-directory-has-an\n","subject":"Add explanation of what was not working before.","message":"Add explanation of what was not working before.\n","lang":"Python","license":"mit","repos":"praveenv253\/sht,praveenv253\/sht"} {"commit":"1c4e929e0a915a5a2610862ee2ef1c57495392c5","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\n#pandoc -f rst -t markdown README.mkd -o README\n\nimport os\nfrom setuptools import setup\n\ndef read(fname):\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n\nsetup(\n name='mass',\n version='0.1.2',\n description='Merge and Simplify Scripts: an automated tool for managing, combining and minifying javascript assets for web projects.',\n long_description=read('README'),\n author='jack boberg alex padgett',\n author_email='info@codedbyhand.com',\n url='https:\/\/github.com\/coded-by-hand\/mass',\n license='BSD License',\n platforms=['Mac OSX'],\n packages=['mass'],\n install_requires=['distribute','jsmin','macfsevents'],\n zip_safe = False,\n entry_points = {\n 'console_scripts': [\n \"mass = mass.monitor:main\"\n ],\n }\n)\n","new_contents":"#!\/usr\/bin\/env python\n\n#pandoc -t rst -f markdown README.mkd -o README\n\nimport os\nfrom setuptools import setup\n\ndef read(fname):\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n\nsetup(\n name='mass',\n version='0.1.3',\n description='Merge and Simplify Scripts: an automated tool for managing, combining and minifying javascript assets for web projects.',\n long_description=read('README'),\n author='jack boberg alex padgett',\n author_email='info@codedbyhand.com',\n url='https:\/\/github.com\/coded-by-hand\/mass',\n license='BSD License',\n platforms=['Mac OSX'],\n packages=['mass'],\n install_requires=['distribute','jsmin','macfsevents'],\n zip_safe = False,\n entry_points = {\n 'console_scripts': [\n \"mass = mass.monitor:main\"\n ],\n }\n)\n","subject":"Fix pandoc statement Update version number","message":"Fix pandoc statement\nUpdate version number\n","lang":"Python","license":"bsd-2-clause","repos":"coded-by-hand\/mass,coded-by-hand\/mass"} {"commit":"ef471f80412d49456725565349bd0e5a09e6e721","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nimport os\nimport sys\nimport codecs\n\nfrom setuptools import setup\n\ntry:\n # Python 3\n from os import dirname\nexcept ImportError:\n # Python 2\n from os.path import dirname\n\nhere = os.path.abspath(dirname(__file__))\n\nwith codecs.open(os.path.join(here, 'README.rst'), encoding='utf-8') as f:\n long_description = '\\n' + f.read()\n\n\nif sys.argv[-1] == \"publish\":\n os.system(\"python setup.py sdist bdist_wheel upload\")\n sys.exit()\n\nrequired = [\n 'humanize',\n 'pytz',\n 'dateparser',\n 'ruamel.yaml',\n 'tzlocal',\n 'pendulum'\n]\n\nsetup(\n name='maya',\n version='0.3.1',\n description='Datetimes for Humans.',\n long_description=long_description,\n author='Kenneth Reitz',\n author_email='me@kennethreitz.org',\n url='https:\/\/github.com\/kennethreitz\/maya',\n packages=['maya'],\n install_requires=required,\n license='MIT',\n classifiers=(\n\n ),\n)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nimport os\nimport sys\nimport codecs\n\nfrom setuptools import setup\n\ntry:\n # Python 3\n from os import dirname\nexcept ImportError:\n # Python 2\n from os.path import dirname\n\nhere = os.path.abspath(dirname(__file__))\n\nwith codecs.open(os.path.join(here, 'README.rst'), encoding='utf-8') as f:\n long_description = '\\n' + f.read()\n\n\nif sys.argv[-1] == \"publish\":\n os.system(\"python setup.py sdist bdist_wheel upload\")\n sys.exit()\n\nrequired = [\n 'humanize',\n 'pytz',\n 'dateparser',\n 'ruamel.yaml',\n 'tzlocal',\n 'pendulum'\n]\n\nsetup(\n name='maya',\n version='0.3.1',\n description='Datetimes for Humans.',\n long_description=long_description,\n author='Kenneth Reitz',\n author_email='me@kennethreitz.org',\n url='https:\/\/github.com\/kennethreitz\/maya',\n packages=['maya'],\n install_requires=required,\n license='MIT',\n classifiers=(\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Natural Language :: English',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: Implementation',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Topic :: Software Development :: Libraries :: Python Modules'\n ),\n)\n","subject":"Add some reasonable trove classifiers","message":"Add some reasonable trove classifiers\n","lang":"Python","license":"mit","repos":"kennethreitz\/maya,timofurrer\/maya,emattiza\/maya"} {"commit":"2800e2cf0a7a998a5081929e6750265f30b09130","old_file":"tests\/test_bql.py","new_file":"tests\/test_bql.py","old_contents":"# -*- coding: utf-8 -*-\n\n# Copyright (c) 2010-2014, MIT Probabilistic Computing Project\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport StringIO\n\nimport bayeslite.bql as bql\nimport bayeslite.parse as parse\n\nimport test_smoke\n\ndef bql2sql(string):\n with test_smoke.t1() as bdb:\n phrases = parse.parse_bql_string(string)\n out = StringIO.StringIO()\n bql.compile_bql(bdb, phrases, out)\n return out.getvalue()\n\ndef test_select_trivial():\n assert bql2sql('select 0;') == 'select 0;'\n","new_contents":"# -*- coding: utf-8 -*-\n\n# Copyright (c) 2010-2014, MIT Probabilistic Computing Project\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nimport StringIO\n\nimport bayeslite.bql as bql\nimport bayeslite.parse as parse\n\nimport test_smoke\n\ndef bql2sql(string):\n with test_smoke.t1() as bdb:\n phrases = parse.parse_bql_string(string)\n out = StringIO.StringIO()\n bql.compile_bql(bdb, phrases, out)\n return out.getvalue()\n\ndef test_select_trivial():\n assert bql2sql('select 0;') == 'select 0;'\n assert bql2sql('select 0 as z;') == 'select 0 as \"z\";'\n assert bql2sql('select * from t;') == 'select * from \"t\";'\n assert bql2sql('select t.* from t;') == 'select \"t\".* from \"t\";'\n assert bql2sql('select c from t;') == 'select \"c\" from \"t\";'\n assert bql2sql('select c as d from t;') == 'select \"c\" as \"d\" from \"t\";'\n","subject":"Add some more trivial bql2sql tests.","message":"Add some more trivial bql2sql tests.\n","lang":"Python","license":"apache-2.0","repos":"probcomp\/bayeslite,probcomp\/bayeslite"} {"commit":"2d6e0710dbc781f54295e18299be6f4c1bb0ec43","old_file":"jsonmerge\/jsonvalue.py","new_file":"jsonmerge\/jsonvalue.py","old_contents":"# vim:ts=4 sw=4 expandtab softtabstop=4\n\nclass JSONValue(object):\n def __init__(self, val=None, ref='#', undef=False):\n assert not isinstance(val, JSONValue)\n self.val = val\n self.ref = ref\n self.undef = undef\n\n def is_undef(self):\n return self.undef\n\n def _subval(self, key, **kwargs):\n return JSONValue(ref=self.ref+'\/'+str(key), **kwargs)\n\n def __getitem__(self, key):\n return self._subval(key, val=self.val[key])\n\n def get(self, key, *args):\n r = self.val.get(key, *args)\n if r is None:\n return self._subval(key, undef=True)\n else:\n return self._subval(key, val=r)\n\n def __repr__(self):\n if self.is_undef():\n return 'JSONValue(undef=True)'\n else:\n return 'JSONValue(%r,%r)' % (self.val, self.ref)\n\n def iteritems(self):\n for k, v in self.val.iteritems():\n yield (k, self._subval(k, val=v))\n\n def items(self):\n return list(self.iteritems())\n\n def __iter__(self):\n assert isinstance(self.val, list)\n\n for i, v in enumerate(self.val):\n yield self._subval(i, val=v)\n","new_contents":"# vim:ts=4 sw=4 expandtab softtabstop=4\n\nclass JSONValue(object):\n def __init__(self, val=None, ref='#', undef=False):\n assert not isinstance(val, JSONValue)\n self.val = val\n self.ref = ref\n self.undef = undef\n\n def is_undef(self):\n return self.undef\n\n def _subval(self, key, **kwargs):\n return JSONValue(ref=self.ref+'\/'+str(key), **kwargs)\n\n def __getitem__(self, key):\n return self._subval(key, val=self.val[key])\n\n def get(self, key, *args):\n r = self.val.get(key, *args)\n if r is None:\n return self._subval(key, undef=True)\n else:\n return self._subval(key, val=r)\n\n def __repr__(self):\n if self.is_undef():\n return 'JSONValue(undef=True)'\n else:\n return 'JSONValue(%r,%r)' % (self.val, self.ref)\n\n def items(self):\n for k, v in self.val.items():\n yield (k, self._subval(k, val=v))\n\n def __iter__(self):\n assert isinstance(self.val, list)\n\n for i, v in enumerate(self.val):\n yield self._subval(i, val=v)\n","subject":"Fix Python3 support: remove iteritems","message":"Fix Python3 support: remove iteritems\n","lang":"Python","license":"mit","repos":"avian2\/jsonmerge"} {"commit":"707e0be6f7e750e580aecc9bced2cc19b9ccf906","old_file":"lib\/windspharm\/__init__.py","new_file":"lib\/windspharm\/__init__.py","old_contents":"\"\"\"Spherical harmonic vector wind analysis.\"\"\"\n# Copyright (c) 2012-2014 Andrew Dawson\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and\/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\nfrom __future__ import absolute_import\n\nfrom . import standard\nfrom . import tools\n\n\n# List to define the behaviour of imports of the form:\n# from windspharm import *\n__all__ = []\n\n# Package version number.\n__version__ = '1.3.2'\n\ntry:\n from . import cdms\n __all__.append('cdms')\n metadata = cdms\nexcept ImportError:\n pass\n\ntry:\n from . import iris\n __all__.append('iris')\nexcept ImportError:\n pass\n","new_contents":"\"\"\"Spherical harmonic vector wind analysis.\"\"\"\n# Copyright (c) 2012-2014 Andrew Dawson\n#\n# Permission is hereby granted, free of charge, to any person obtaining a copy\n# of this software and associated documentation files (the \"Software\"), to deal\n# in the Software without restriction, including without limitation the rights\n# to use, copy, modify, merge, publish, distribute, sublicense, and\/or sell\n# copies of the Software, and to permit persons to whom the Software is\n# furnished to do so, subject to the following conditions:\n#\n# The above copyright notice and this permission notice shall be included in\n# all copies or substantial portions of the Software.\n#\n# THE SOFTWARE IS PROVIDED \"AS IS\", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR\n# IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,\n# FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE\n# AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER\n# LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,\n# OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN\n# THE SOFTWARE.\nfrom __future__ import absolute_import\n\nfrom . import standard\nfrom . import tools\n\n\n# List to define the behaviour of imports of the form:\n# from windspharm import *\n__all__ = []\n\n# Package version number.\n__version__ = '1.3.x'\n\ntry:\n from . import cdms\n __all__.append('cdms')\n metadata = cdms\nexcept ImportError:\n pass\n\ntry:\n from . import iris\n __all__.append('iris')\nexcept ImportError:\n pass\n","subject":"Revert version number on release branch.","message":"Revert version number on release branch.\n","lang":"Python","license":"mit","repos":"nicolasfauchereau\/windspharm,ajdawson\/windspharm"} {"commit":"188f0ac84e041259585172f5cffc21828ac534e3","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\nsetup(\n name='daffodil',\n version='0.3.8',\n author='James Robert',\n description='A Super-simple DSL for filtering datasets',\n license='MIT',\n keywords='data filtering',\n url='https:\/\/github.com\/mediapredict\/daffodil',\n packages=['daffodil'],\n install_requires=[\n \"parsimonious\",\n ],\n long_description='A Super-simple DSL for filtering datasets',\n classifiers=[\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Intended Audience :: Developers',\n 'Operating System :: OS Independent',\n 'Topic :: Utilities'\n ]\n)\n","new_contents":"from setuptools import setup\n\nsetup(\n name='daffodil',\n version='0.3.9',\n author='James Robert',\n description='A Super-simple DSL for filtering datasets',\n license='MIT',\n keywords='data filtering',\n url='https:\/\/github.com\/mediapredict\/daffodil',\n packages=['daffodil'],\n install_requires=[\n \"parsimonious\",\n ],\n long_description='A Super-simple DSL for filtering datasets',\n classifiers=[\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Intended Audience :: Developers',\n 'Operating System :: OS Independent',\n 'Topic :: Utilities'\n ]\n)\n","subject":"Upgrade version for new optimization","message":"Upgrade version for new optimization\n\n(for real this time)","lang":"Python","license":"mit","repos":"igorkramaric\/daffodil,mediapredict\/daffodil"} {"commit":"99d3e7972e642050c4586968ba704d46d469e27c","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup, find_packages\n\nsetup(\n # packaging information that is likely to be updated between versions\n name='icecake',\n version='0.6.0',\n packages=['icecake'],\n py_modules=['cli', 'templates', 'livejs'],\n entry_points='''\n [console_scripts]\n icecake=icecake.cli:cli\n ''',\n install_requires=[\n 'Click',\n 'Jinja2',\n 'Markdown',\n 'Pygments',\n 'python-dateutil',\n 'watchdog',\n 'Werkzeug',\n ],\n\n # pypy stuff that is not likely to change between versions\n url=\"https:\/\/github.com\/cbednarski\/icecake\",\n author=\"Chris Bednarski\",\n author_email=\"banzaimonkey@gmail.com\",\n description=\"An easy and cool static site generator\",\n license=\"MIT\",\n long_description=open('README.rst').read(),\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.5',\n ],\n keywords=\"static site generator builder icecake\"\n)\n","new_contents":"from setuptools import setup, find_packages\nfrom codec import open\n\nsetup(\n # packaging information that is likely to be updated between versions\n name='icecake',\n version='0.6.0',\n packages=['icecake'],\n py_modules=['cli', 'templates', 'livejs'],\n entry_points='''\n [console_scripts]\n icecake=icecake.cli:cli\n ''',\n install_requires=[\n 'Click',\n 'Jinja2',\n 'Markdown',\n 'Pygments',\n 'python-dateutil',\n 'watchdog',\n 'Werkzeug',\n ],\n\n # pypy stuff that is not likely to change between versions\n url=\"https:\/\/github.com\/cbednarski\/icecake\",\n author=\"Chris Bednarski\",\n author_email=\"banzaimonkey@gmail.com\",\n description=\"An easy and cool static site generator\",\n license=\"MIT\",\n long_description=open('README.rst', encoding=\"utf-8\").read(),\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.5',\n ],\n keywords=\"static site generator builder icecake\"\n)\n","subject":"Fix encoding error preventing install","message":"Fix encoding error preventing install\n\nWhile running tox the installation of icecake would succeed under Python\n2.7.13 but fail under 3.5.2 with an encoding error while trying to read\nthe long description from README.rst.\n\npy35 inst-nodeps: \/icecake\/.tox\/dist\/icecake-0.6.0.zip\nERROR: invocation failed (exit code 1), logfile:\n\/icecake\/.tox\/py35\/log\/py35-16.log\nERROR: actionid: py35\nmsg: installpkg\ncmdargs: ['\/icecake\/.tox\/py35\/bin\/pip', 'install', '-U', '--no-deps',\n'\/icecake\/.tox\/dist\/icecake-0.6.0.zip']\nenv: {'TERM': 'xterm', 'VIRTUAL_ENV': '\/icecake\/.tox\/py35', 'SHLVL':\n'1', 'PYENV_HOOK_PATH':\n'\/.pyenv\/pyenv.d:\/usr\/local\/etc\/pyenv.d:\/etc\/pyenv.d:\/usr\/lib\/pyenv\/hooks:\/.pyenv\/plugins\/pyenv-virtualenv\/etc\/pyenv.d',\n'HOSTNAME': 'cdbe5b1470a0', 'PYENV_VERSION': '2.7.13:3.5.2',\n'PYENV_DIR': '\/icecake', 'PWD': '\/icecake', 'PYTHONHASHSEED':\n'1135274721', 'PATH':\n'\/icecake\/.tox\/py35\/bin:\/.pyenv\/versions\/2.7.13\/bin:\/.pyenv\/libexec:\/.pyenv\/plugins\/python-build\/bin:\/.pyenv\/plugins\/pyenv-virtualenv\/bin:\/.pyenv\/shims:\/.pyenv\/bin:\/usr\/local\/sbin:\/usr\/local\/bin:\/usr\/sbin:\/usr\/bin:\/sbin:\/bin',\n'HOME': '\/root', 'PYENV_ROOT': '\/.pyenv'}\n\nProcessing .\/.tox\/dist\/icecake-0.6.0.zip\n Complete output from command python setup.py egg_info:\n Traceback (most recent call last):\n File \"<string>\", line 1, in <module>\n File \"\/tmp\/pip-x3boup3_-build\/setup.py\", line 29, in <module>\n long_description=open('README.rst').read(),\n File \"\/icecake\/.tox\/py35\/lib\/python3.5\/encodings\/ascii.py\", line\n26, in decode\n return codecs.ascii_decode(input, self.errors)[0]\n UnicodeDecodeError: 'ascii' codec can't decode byte 0xe2 in position\n1462: ordinal not in range(128)\n\n ----------------------------------------\nCommand \"python setup.py egg_info\" failed with error code 1 in\n\/tmp\/pip-x3boup3_-build\/\n","lang":"Python","license":"mit","repos":"cbednarski\/icecake,cbednarski\/icecake"} {"commit":"afefe8afcb69988771ecd2c5fddbcafa545df888","old_file":"setup.py","new_file":"setup.py","old_contents":"\"\"\"Defines the setup for the declxml library\"\"\"\nfrom io import open\nimport os.path\nfrom setuptools import setup\n\n\ndir_path = os.path.abspath(os.path.dirname(__file__))\nreadme_path = os.path.join(dir_path, 'README.md')\nwith open(readme_path, encoding='utf-8') as readme:\n long_description = readme.read()\n\n\nsetup(\n name='declxml',\n description='Declarative XML processing library',\n long_description=long_description,\n long_description_content_type='text\/markdown',\n version='1.1.2rc4',\n url='http:\/\/declxml.readthedocs.io\/',\n project_urls={\n 'Documentation': 'http:\/\/declxml.readthedocs.io\/',\n 'Source': 'https:\/\/github.com\/gatkin\/declxml',\n 'Tracker': 'https:\/\/github.com\/gatkin\/declxml\/issues',\n },\n author='Greg Atkin',\n author_email='greg.scott.atkin@gmail.com',\n license='MIT',\n py_modules=['declxml'],\n install_requires=['typing'],\n data_files=[('', ['py.typed'])],\n zip_safe=False,\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'License :: OSI Approved :: MIT License',\n ],\n keywords='XML, Parsing, Serialization'\n)\n","new_contents":"\"\"\"Defines the setup for the declxml library\"\"\"\nfrom io import open\nimport os.path\nfrom setuptools import setup\n\n\ndir_path = os.path.abspath(os.path.dirname(__file__))\nreadme_path = os.path.join(dir_path, 'README.md')\nwith open(readme_path, encoding='utf-8') as readme:\n long_description = readme.read()\n\n\nsetup(\n name='declxml',\n description='Declarative XML processing library',\n long_description=long_description,\n long_description_content_type='text\/markdown',\n version='1.1.2',\n url='http:\/\/declxml.readthedocs.io\/',\n project_urls={\n 'Documentation': 'http:\/\/declxml.readthedocs.io\/',\n 'Source': 'https:\/\/github.com\/gatkin\/declxml',\n 'Tracker': 'https:\/\/github.com\/gatkin\/declxml\/issues',\n },\n author='Greg Atkin',\n author_email='greg.scott.atkin@gmail.com',\n license='MIT',\n py_modules=['declxml'],\n install_requires=['typing'],\n data_files=[('', ['py.typed'])],\n zip_safe=False,\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'License :: OSI Approved :: MIT License',\n ],\n keywords='XML, Parsing, Serialization'\n)\n","subject":"Bump version for PyPi release :shipit:","message":"Bump version for PyPi release :shipit:\n","lang":"Python","license":"mit","repos":"gatkin\/declxml"} {"commit":"0a86f9b44c9d3213cb965107c867f8043f062fe3","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\n\ndef read_file(fname):\n with open(fname) as f:\n return f.read()\n\n\nsetup(\n name='pytest-testrail',\n description='pytest plugin for creating TestRail runs and adding results',\n long_description=read_file('README.rst'),\n version='2.3.3',\n author='Allan Kilpatrick',\n author_email='allanklp@gmail.com',\n url='http:\/\/github.com\/allankp\/pytest-testrail\/',\n packages=[\n 'pytest_testrail',\n ],\n package_dir={'pytest_testrail': 'pytest_testrail'},\n install_requires=[\n 'pytest>=3.6',\n 'requests>=2.20.0',\n 'simplejson',\n ],\n include_package_data=True,\n entry_points={'pytest11': ['pytest-testrail = pytest_testrail.conftest']},\n)\n","new_contents":"from setuptools import setup\n\n\ndef read_file(fname):\n with open(fname) as f:\n return f.read()\n\n\nsetup(\n name='pytest-testrail',\n description='pytest plugin for creating TestRail runs and adding results',\n long_description=read_file('README.rst'),\n version='2.5.0',\n author='Allan Kilpatrick',\n author_email='allanklp@gmail.com',\n url='http:\/\/github.com\/allankp\/pytest-testrail\/',\n packages=[\n 'pytest_testrail',\n ],\n package_dir={'pytest_testrail': 'pytest_testrail'},\n install_requires=[\n 'pytest>=3.6',\n 'requests>=2.20.0',\n 'simplejson',\n ],\n include_package_data=True,\n entry_points={'pytest11': ['pytest-testrail = pytest_testrail.conftest']},\n)\n","subject":"Bump version: 2.4.0 → 2.5.0","message":"Bump version: 2.4.0 → 2.5.0\n","lang":"Python","license":"mit","repos":"dubner\/pytest-testrail,allankilpatrick\/pytest-testrail"} {"commit":"6d8b2453a77008acb6f6cde002c6bfaea2a75621","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nfrom distutils.core import setup\n\nsetup(\n name='rinse',\n version='0.0.3',\n description='Python3 SOAP client built with lxml and requests.',\n author='Tyson Clugg',\n author_email='tyson@clugg.net',\n url='http:\/\/github.com\/tysonclugg\/rinse',\n license='MIT',\n packages=['rinse'],\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"Development Status :: 3 - Alpha\",\n \"License :: OSI Approved :: MIT License\",\n ],\n)\n","new_contents":"#!\/usr\/bin\/env python\nfrom distutils.core import setup\n\nsetup(\n name='rinse',\n version='0.0.4',\n description='Python3 SOAP client built with lxml and requests.',\n author='Tyson Clugg',\n author_email='tyson@clugg.net',\n url='http:\/\/github.com\/tysonclugg\/rinse',\n license='MIT',\n packages=['rinse'],\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"Development Status :: 3 - Alpha\",\n \"License :: OSI Approved :: MIT License\",\n ],\n)\n","subject":"Remove reference to stale source (client.py).","message":"Remove reference to stale source (client.py).\n","lang":"Python","license":"mit","repos":"simudream\/rinse,thedrow\/rinse,MarkusH\/rinse,tysonclugg\/rinse,funkybob\/rinse,MarkusH\/rinse,simudream\/rinse,tysonclugg\/rinse"} {"commit":"bc0895f318a9297144e31da3647d6fc5716aafc4","old_file":"setup.py","new_file":"setup.py","old_contents":"'''\nSetup script that:\n\n\/pyquic:\n - compiles pyquic\n - copies py_quic into base directory so that we can use the module directly\n'''\nimport os\nimport shutil\n\n\nclass temp_cd():\n def __init__(self, temp_dir):\n self._temp_dir = temp_dir\n self._return_dir = os.path.dirname(os.path.realpath(__file__))\n def __enter__(self):\n os.chdir(self._temp_dir)\n def __exit__(self, type, value, traceback):\n os.chdir(self._return_dir)\n\ndef setup_pyquic():\n with temp_cd('pyquic\/py_quic'):\n os.system('make')\n\n shutil.rmtree('quic\/py_quic')\n shutil.copytree('pyquic\/py_quic', 'quic\/py_quic')\n \ndef clean_pyquic():\n shutil.rmtree('py_quic')\n os.system('git submodule update --checkout --remote -f')\n\nif __name__ == \"__main__\":\n setup_pyquic()\n","new_contents":"'''\nSetup script that:\n\n\/pyquic:\n - compiles pyquic\n - copies py_quic into base directory so that we can use the module directly\n'''\nimport os\nimport shutil\n\n\nclass temp_cd():\n def __init__(self, temp_dir):\n self._temp_dir = temp_dir\n self._return_dir = os.path.dirname(os.path.realpath(__file__))\n def __enter__(self):\n os.chdir(self._temp_dir)\n def __exit__(self, type, value, traceback):\n os.chdir(self._return_dir)\n\ndef setup_pyquic():\n with temp_cd('pyquic\/py_quic'):\n os.system('make')\n\n if os.path.exists('quic\/py_quic'):\n shutil.rmtree('quic\/py_quic')\n \n shutil.copytree('pyquic\/py_quic', 'quic\/py_quic')\n \ndef clean_pyquic():\n shutil.rmtree('py_quic')\n os.system('git submodule update --checkout --remote -f')\n\nif __name__ == \"__main__\":\n setup_pyquic()\n","subject":"Make sure this works the first time you run it","message":"Make sure this works the first time you run it\n","lang":"Python","license":"mit","repos":"skggm\/skggm,skggm\/skggm"} {"commit":"b2d6488b26f454572179943b40ef137711cb0dc5","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\n\nsetup(\n name='slacker',\n version='0.7.0',\n packages=['slacker'],\n description='Slack API client',\n author='Oktay Sancak',\n author_email='oktaysancak@gmail.com',\n url='http:\/\/github.com\/os\/slacker\/',\n install_requires=['requests >= 2.2.1'],\n license='http:\/\/www.apache.org\/licenses\/LICENSE-2.0',\n test_suite='tests',\n classifiers=(\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache Software License',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4'\n ),\n keywords='slack api'\n)\n","new_contents":"from setuptools import setup\n\n\nsetup(\n name='slacker',\n version='0.7.3',\n packages=['slacker'],\n description='Slack API client',\n author='Oktay Sancak',\n author_email='oktaysancak@gmail.com',\n url='http:\/\/github.com\/os\/slacker\/',\n install_requires=['requests >= 2.2.1'],\n license='http:\/\/www.apache.org\/licenses\/LICENSE-2.0',\n test_suite='tests',\n classifiers=(\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache Software License',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4'\n ),\n keywords='slack api'\n)\n","subject":"Set version number to 0.7.3.","message":"Set version number to 0.7.3.\n","lang":"Python","license":"apache-2.0","repos":"wasabi0522\/slacker,STANAPO\/slacker,techartorg\/slacker,wkentaro\/slacker,hreeder\/slacker,kashyap32\/slacker,os\/slacker"} {"commit":"e3e0c8dbce7f3c6fb8887f2f9cc2332020d7480b","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\n\nsetup(\n name='tangled.sqlalchemy',\n version='0.1a4.dev0',\n description='Tangled SQLAlchemy integration',\n long_description=open('README.rst').read(),\n url='http:\/\/tangledframework.org\/',\n download_url='https:\/\/github.com\/TangledWeb\/tangled.sqlalchemy\/tags',\n author='Wyatt Baldwin',\n author_email='self@wyattbaldwin.com',\n packages=[\n 'tangled',\n 'tangled.sqlalchemy',\n ],\n install_requires=[\n 'tangled>=0.1a7',\n 'SQLAlchemy',\n ],\n extras_require={\n 'dev': [\n 'tangled[dev]>=0.1a7',\n ],\n },\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n ],\n)\n","new_contents":"from setuptools import setup\n\n\nsetup(\n name='tangled.sqlalchemy',\n version='0.1a4.dev0',\n description='Tangled SQLAlchemy integration',\n long_description=open('README.rst').read(),\n url='http:\/\/tangledframework.org\/',\n download_url='https:\/\/github.com\/TangledWeb\/tangled.sqlalchemy\/tags',\n author='Wyatt Baldwin',\n author_email='self@wyattbaldwin.com',\n packages=[\n 'tangled',\n 'tangled.sqlalchemy',\n ],\n install_requires=[\n 'tangled>=0.1a9',\n 'SQLAlchemy',\n ],\n extras_require={\n 'dev': [\n 'tangled[dev]>=0.1a9',\n ],\n },\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n ],\n)\n","subject":"Upgrade tangled 0.1a7 => 0.1a9","message":"Upgrade tangled 0.1a7 => 0.1a9\n","lang":"Python","license":"mit","repos":"TangledWeb\/tangled.sqlalchemy"} {"commit":"35c3f9a339a199226b34efae9e78d15e85e5f184","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup # type: ignore[import]\n\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetup(\n name=\"objname\",\n version=\"0.11.0\",\n packages=[\"objname\"],\n package_data={\n \"objname\": [\"__init__.py\", \"py.typed\", \"_module.py\",\n \"test_objname.py\"],\n },\n\n zip_safe=False,\n author=\"Alan Cristhian Ruiz\",\n author_email=\"alan.cristh@gmail.com\",\n description=\"A library with a base class that \"\n \"stores the assigned name of an object.\",\n long_description=long_description,\n long_description_content_type=\"text\/markdown\",\n classifiers=[\n 'Development Status :: 2 - Pre-Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n 'Programming Language :: Python :: 3.9',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Object Brokering',\n 'Typing :: Typed'\n ],\n license=\"MIT\",\n keywords=\"data structure debug\",\n url=\"https:\/\/github.com\/AlanCristhian\/objname\",\n)\n","new_contents":"from setuptools import setup # type: ignore[import]\n\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\nsetup(\n name=\"objname\",\n version=\"0.12.0\",\n packages=[\"objname\"],\n package_data={\n \"objname\": [\"__init__.py\", \"py.typed\", \"_module.py\",\n \"test_objname.py\"],\n },\n\n zip_safe=False,\n author=\"Alan Cristhian Ruiz\",\n author_email=\"alan.cristh@gmail.com\",\n description=\"A library with a base class that \"\n \"stores the assigned name of an object.\",\n long_description=long_description,\n long_description_content_type=\"text\/markdown\",\n classifiers=[\n 'Development Status :: 2 - Pre-Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n 'Programming Language :: Python :: 3.9',\n 'Programming Language :: Python :: 3.10',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Object Brokering',\n 'Typing :: Typed'\n ],\n license=\"MIT\",\n keywords=\"data structure debug\",\n url=\"https:\/\/github.com\/AlanCristhian\/objname\",\n)\n","subject":"Document that python 3.10 is supported and update version number.","message":"Document that python 3.10 is supported and update version number.\n","lang":"Python","license":"mit","repos":"AlanCristhian\/namedobject,AlanCristhian\/named"} {"commit":"7e341014059c98bdd91a1c876982b8caa47a5586","old_file":"setup.py","new_file":"setup.py","old_contents":"from __future__ import with_statement\n\nfrom distutils.core import setup\n\n\ndef readme():\n try:\n with open('README.rst') as f:\n return f.read()\n except IOError:\n return\n\n\nsetup(\n name='encodingcontext',\n version='0.9.0',\n description='A bad idea about the default encoding',\n long_description=readme(),\n py_modules=['encodingcontext'],\n author='Hong Minhee',\n author_email='minhee' '@' 'dahlia.kr',\n url='https:\/\/github.com\/dahlia\/encodingcontext',\n license='MIT License',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.5',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Programming Language :: Python :: Implementation :: Stackless',\n 'Topic :: Text Processing'\n ]\n)\n","new_contents":"from __future__ import with_statement\n\nfrom distutils.core import setup\nimport re\n\n\ndef readme():\n try:\n with open('README.rst') as f:\n readme = f.read()\n except IOError:\n return\n return re.sub(\n r'''\n (?P<colon> : \\n{2,})?\n \\.\\. [ ] code-block:: \\s+ [^\\n]+ \\n\n [^ \\t]* \\n\n (?P<block>\n (?: (?: (?: \\t | [ ]{3}) [^\\n]* | [ \\t]* ) \\n)+\n )\n ''',\n lambda m: (':' + m.group('colon') if m.group('colon') else '') +\n '\\n'.join(' ' + l for l in m.group('block').splitlines()) +\n '\\n\\n',\n readme, 0, re.VERBOSE\n )\n\n\nsetup(\n name='encodingcontext',\n version='0.9.0',\n description='A bad idea about the default encoding',\n long_description=readme(),\n py_modules=['encodingcontext'],\n author='Hong Minhee',\n author_email='minhee' '@' 'dahlia.kr',\n url='https:\/\/github.com\/dahlia\/encodingcontext',\n license='MIT License',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.5',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n 'Programming Language :: Python :: Implementation :: Stackless',\n 'Topic :: Text Processing'\n ]\n)\n","subject":"Transform readme into PyPI compliant reST","message":"Transform readme into PyPI compliant reST\n","lang":"Python","license":"mit","repos":"dahlia\/encodingcontext"} {"commit":"31fde6b38329252313c40549b9188c584a2eadd7","old_file":"setup.py","new_file":"setup.py","old_contents":"from __future__ import print_function\n\ntry:\n from setuptools import setup # try first in case it's already there.\nexcept ImportError:\n from ez_setup import use_setuptools\n use_setuptools()\n from setuptools import setup\n\nsetup(\n name='vpython',\n packages=['vpython'],\n version='0.2.0b15',\n description='VPython for Jupyter Notebook',\n long_description=open('README.md').read(),\n author='John Coady \/ Ruth Chabay \/ Bruce Sherwood \/ Steve Spicklemire',\n author_email='bruce.sherwood@gmail.com',\n url='http:\/\/pypi.python.org\/pypi\/vpython\/',\n license='LICENSE.txt',\n keywords='vpython',\n classifiers=[\n 'Framework :: IPython',\n 'Development Status :: 4 - Beta',\n 'Environment :: Web Environment',\n 'Intended Audience :: End Users\/Desktop',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Topic :: Multimedia :: Graphics :: 3D Modeling',\n 'Topic :: Multimedia :: Graphics :: 3D Rendering',\n 'Topic :: Scientific\/Engineering :: Visualization',\n ],\n install_requires=['jupyter', 'vpnotebook'],\n package_data={'vpython': ['data\/*']},\n)\n","new_contents":"from __future__ import print_function\n\ntry:\n from setuptools import setup # try first in case it's already there.\nexcept ImportError:\n from ez_setup import use_setuptools\n use_setuptools()\n from setuptools import setup\n\nsetup(\n name='vpython',\n packages=['vpython'],\n version='0.2.0b16',\n description='VPython for Jupyter Notebook',\n long_description=open('README.md').read(),\n author='John Coady \/ Ruth Chabay \/ Bruce Sherwood \/ Steve Spicklemire',\n author_email='bruce.sherwood@gmail.com',\n url='http:\/\/pypi.python.org\/pypi\/vpython\/',\n license='LICENSE.txt',\n keywords='vpython',\n classifiers=[\n 'Framework :: IPython',\n 'Development Status :: 4 - Beta',\n 'Environment :: Web Environment',\n 'Intended Audience :: End Users\/Desktop',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Topic :: Multimedia :: Graphics :: 3D Modeling',\n 'Topic :: Multimedia :: Graphics :: 3D Rendering',\n 'Topic :: Scientific\/Engineering :: Visualization',\n ],\n install_requires=['jupyter', 'vpnotebook'],\n package_data={'vpython': ['data\/*']},\n)\n","subject":"Update version number to 0.2.0b16","message":"Update version number to 0.2.0b16\n","lang":"Python","license":"mit","repos":"BruceSherwood\/vpython-jupyter,BruceSherwood\/vpython-jupyter,BruceSherwood\/vpython-jupyter,mwcraig\/vpython-jupyter,mwcraig\/vpython-jupyter,mwcraig\/vpython-jupyter,BruceSherwood\/vpython-jupyter,sspickle\/vpython-jupyter,mwcraig\/vpython-jupyter,sspickle\/vpython-jupyter"} {"commit":"5fb38bfb6eae77b7024bf4d9990472f60d576826","old_file":"setup.py","new_file":"setup.py","old_contents":"import pathlib\nfrom crc import LIBRARY_VERSION\nfrom setuptools import setup\n\ncurrent = pathlib.Path(__file__).parent.resolve()\n\n\ndef readme():\n return (current \/ 'README.md').read_text(encoding='utf-8')\n\n\nif __name__ == '__main__':\n setup(\n name='crc',\n version=LIBRARY_VERSION,\n py_modules=['crc'],\n classifiers=[\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n ],\n url='https:\/\/github.com\/Nicoretti\/crc',\n license='BSD',\n python_requires='>=3.7',\n author='Nicola Coretti',\n author_email='nico.coretti@gmail.com',\n description='Library and CLI to calculate and verify all kinds of CRC checksums.',\n keywords=['CRC', 'CRC8', 'CRC16', 'CRC32', 'CRC64'],\n long_description=readme(),\n long_description_content_type='text\/markdown',\n entry_points={\n 'console_scripts': [\n 'crc=crc:main',\n ],\n }\n )\n","new_contents":"import pathlib\nfrom crc import LIBRARY_VERSION\nfrom setuptools import setup\n\ncurrent = pathlib.Path(__file__).parent.resolve()\n\n\ndef readme():\n return (current \/ 'README.md').read_text(encoding='utf-8')\n\n\nif __name__ == '__main__':\n setup(\n name='crc',\n version=LIBRARY_VERSION,\n py_modules=['crc'],\n classifiers=[\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n 'Programming Language :: Python :: 3.9',\n ],\n url='https:\/\/github.com\/Nicoretti\/crc',\n license='BSD',\n python_requires='>=3.7',\n author='Nicola Coretti',\n author_email='nico.coretti@gmail.com',\n description='Library and CLI to calculate and verify all kinds of CRC checksums.',\n keywords=['CRC', 'CRC8', 'CRC16', 'CRC32', 'CRC64'],\n long_description=readme(),\n long_description_content_type='text\/markdown',\n entry_points={\n 'console_scripts': [\n 'crc=crc:main',\n ],\n }\n )\n","subject":"Update package information about supported python versions","message":"Update package information about supported python versions\n","lang":"Python","license":"bsd-2-clause","repos":"Nicoretti\/crc"} {"commit":"40a2a4ace817f3f237d87d802d5bd286eb2bf09e","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name='syndicate',\n version='0.99.6',\n description='A wrapper for REST APIs',\n author='Justin Mayfield',\n author_email='tooker@gmail.com',\n url='https:\/\/github.com\/mayfield\/syndicate\/',\n license='MIT',\n long_description=open('README.rst').read(),\n packages=find_packages(),\n install_requires=[\n 'requests',\n 'python-dateutil',\n 'tornado',\n ],\n test_suite='test',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Software Development :: Libraries',\n ]\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup, find_packages\n\nREADME = 'README.md'\n\ndef long_desc():\n try:\n import pypandoc\n except ImportError:\n with open(README) as f:\n return f.read()\n else:\n return pypandoc.convert(README, 'rst')\n\nsetup(\n name='syndicate',\n version='0.99.7',\n description='A wrapper for REST APIs',\n author='Justin Mayfield',\n author_email='tooker@gmail.com',\n url='https:\/\/github.com\/mayfield\/syndicate\/',\n license='MIT',\n long_description=long_desc(),\n packages=find_packages(),\n install_requires=[\n 'requests',\n 'python-dateutil',\n 'tornado',\n ],\n test_suite='test',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Software Development :: Libraries',\n ]\n)\n","subject":"Make PyPi long_description look pretty with pandoc (if available).","message":"Make PyPi long_description look pretty with pandoc (if available).\n","lang":"Python","license":"mit","repos":"mayfield\/syndicate"} {"commit":"ae280f4f837a23608749e8a8de1cbba98bafc621","old_file":"examples\/mhs_atmosphere\/mhs_atmosphere_plot.py","new_file":"examples\/mhs_atmosphere\/mhs_atmosphere_plot.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Jan 9 12:52:31 2015\n\n@author: stuart\n\"\"\"\nimport os\nimport glob\n\nimport yt\n\nmodel = 'spruit'\ndatadir = os.path.expanduser('~\/mhs_atmosphere\/'+model+'\/')\n\nfiles = glob.glob(datadir+'\/*')\nfiles.sort()\n\nprint(files)\n\n\nds = yt.load(files[0])\n\nslc = yt.SlicePlot(ds, fields='density_bg', normal='x')\nslc.save('~\/yt.png')","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nCreated on Fri Jan 9 12:52:31 2015\n\n@author: stuart\n\"\"\"\nimport os\nimport glob\n\nimport yt\n\nmodel = 'spruit'\ndatadir = os.path.expanduser('~\/mhs_atmosphere\/'+model+'\/')\n\nfiles = glob.glob(datadir+'\/*')\nfiles.sort()\n\nprint(files)\n\n\nds = yt.load(files[0])\n\n# Axes flip for normal='y'\n#ds.coordinates.x_axis = {0: 2, 1: 0, 2: 1, 'x': 2, 'y': 0, 'z': 1}\n#ds.coordinates.y_axis = {0: 1, 1: 2, 2: 0, 'x': 1, 'y': 2, 'z': 0}\n\nslc = yt.SlicePlot(ds, fields='density_bg', normal='x')\nslc.save('~\/yt.png')","subject":"Add x-y flipping code for SlicePlot","message":"Add x-y flipping code for SlicePlot\n","lang":"Python","license":"bsd-2-clause","repos":"SWAT-Sheffield\/pysac,Cadair\/pysac"} {"commit":"db362bd50c7b8aa6a40a809346eff70df846f82d","old_file":"setup.py","new_file":"setup.py","old_contents":"# coding: utf-8\nfrom setuptools import setup\n\n\nsetup(\n name='pysuru',\n version='0.0.1',\n description='Python library to interact with Tsuru API',\n long_description=open('README.rst', 'r').read(),\n keywords='tsuru',\n author='Rodrigo Machado',\n author_email='rcmachado@gmail.com',\n url='https:\/\/github.com\/rcmachado\/pysuru',\n license='MIT',\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers'\n 'Topic :: Software Development :: Libraries',\n ],\n install_requires=[\n 'urllib3>=1.15'\n ]\n packages=['pysuru'],\n platforms=['linux', 'osx']\n)\n\n","new_contents":"# coding: utf-8\nfrom setuptools import setup\n\n\nsetup(\n name='pysuru',\n version='0.0.1',\n description='Python library to interact with Tsuru API',\n long_description=open('README.rst', 'r').read(),\n keywords='tsuru',\n author='Rodrigo Machado',\n author_email='rcmachado@gmail.com',\n url='https:\/\/github.com\/rcmachado\/pysuru',\n license='MIT',\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers'\n 'Topic :: Software Development :: Libraries',\n ],\n install_requires=[\n 'urllib3>=1.15',\n 'certifi'\n ]\n packages=['pysuru'],\n platforms=['linux', 'osx']\n)\n\n","subject":"Add certifi to required packages","message":"Add certifi to required packages\n","lang":"Python","license":"mit","repos":"rcmachado\/pysuru"} {"commit":"38a13415fc4c126c1e115d5af0d0ffde5bcdf08d","old_file":"setup.py","new_file":"setup.py","old_contents":"# https:\/\/jeffknupp.com\/blog\/2013\/08\/16\/open-sourcing-a-python-project-the-right-way\/\n# http:\/\/peterdowns.com\/posts\/first-time-with-pypi.html\n# Upload to PyPI Live\n# python setup.py sdist upload -r pypi\n\nfrom setuptools import setup\n\nsetup(\n name='axis',\n packages=['axis'],\n version='9',\n description='A python library for communicating with devices from Axis Communications',\n author='Robert Svensson',\n author_email='Kane610@users.noreply.github.com',\n license='MIT',\n url='https:\/\/github.com\/Kane610\/axis',\n download_url='https:\/\/github.com\/Kane610\/axis\/archive\/v9.tar.gz',\n install_requires=['requests'],\n keywords=['axis', 'vapix', 'onvif', 'event stream', 'homeassistant'],\n classifiers=[],\n)\n","new_contents":"# https:\/\/jeffknupp.com\/blog\/2013\/08\/16\/open-sourcing-a-python-project-the-right-way\/\n# http:\/\/peterdowns.com\/posts\/first-time-with-pypi.html\n# Upload to PyPI Live\n# python setup.py sdist upload -r pypi\n\nfrom setuptools import setup\n\nsetup(\n name='axis',\n packages=['axis'],\n version='10',\n description='A python library for communicating with devices from Axis Communications',\n author='Robert Svensson',\n author_email='Kane610@users.noreply.github.com',\n license='MIT',\n url='https:\/\/github.com\/Kane610\/axis',\n download_url='https:\/\/github.com\/Kane610\/axis\/archive\/v10.tar.gz',\n install_requires=['requests'],\n keywords=['axis', 'vapix', 'onvif', 'event stream', 'homeassistant'],\n classifiers=[],\n)\n","subject":"Bump version number to 10","message":"Bump version number to 10\n","lang":"Python","license":"mit","repos":"Kane610\/axis"} {"commit":"8433285fbb04f54ff1d8d0900fad4396f9e368a0","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom distutils.core import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-thumborstorage',\n version='0.92.2',\n license='MIT Licence',\n author='Stanislas Guerra',\n author_email='stanislas.guerra@gmail.com',\n description='Django custom storage for Thumbor backend.',\n long_description=README,\n url='https:\/\/github.com\/Starou\/django-thumborstorage',\n packages=['django_thumborstorage'],\n package_data={\n 'django_yaaac': []\n },\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n install_requires=['requests', 'mock'],\n)\n","new_contents":"import os\nfrom distutils.core import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), 'README.rst')).read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-thumborstorage',\n version='1.9.12',\n license='MIT Licence',\n author='Stanislas Guerra',\n author_email='stanislas.guerra@gmail.com',\n description='Django custom storage for Thumbor backend.',\n long_description=README,\n url='https:\/\/github.com\/Starou\/django-thumborstorage',\n packages=['django_thumborstorage'],\n package_data={\n 'django_yaaac': []\n },\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n install_requires=['requests', 'mock'],\n)\n","subject":"Bump version (use latest Django version supported.","message":"django-1.9: Bump version (use latest Django version supported.\n","lang":"Python","license":"mit","repos":"Starou\/django-thumborstorage,Starou\/django-thumborstorage"} {"commit":"b71ee9646bf9d7920cc12b77b15b3163cfbb95bc","old_file":"setup.py","new_file":"setup.py","old_contents":"# \/setup.py\n#\n# Installation and setup script for cmakeast\n#\n# See \/LICENCE.md for Copyright information\n\"\"\"Installation and setup script for cmakeast.\"\"\"\n\nfrom setuptools import find_packages, setup\n\nsetup(name=\"cmakeast\",\n version=\"0.0.16\",\n description=\"\"\"Parse a CMake file into an Abstract Syntax Tree.\"\"\",\n long_description_markdown_filename=\"README.md\",\n author=\"Sam Spilsbury\",\n author_email=\"smspillaz@gmail.com\",\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.1\",\n \"Programming Language :: Python :: 3.2\",\n \"Programming Language :: Python :: 3.3\",\n \"Programming Language :: Python :: 3.4\",\n \"Intended Audience :: Developers\",\n \"Topic :: Software Development :: Build Tools\",\n \"License :: OSI Approved :: MIT License\"],\n url=\"http:\/\/github.com\/polysquare\/cmake-ast\",\n license=\"MIT\",\n keywords=\"development ast cmake\",\n packages=find_packages(exclude=[\"test\"]),\n install_requires=[\"setuptools\"],\n extras_require={\n \"upload\": [\"setuptools-markdown\"]\n },\n entry_points={\n \"console_scripts\": [\n \"cmake-print-ast=cmakeast.printer:main\"\n ]\n },\n test_suite=\"nose.collector\",\n zip_safe=True,\n include_package_data=True)\n","new_contents":"# \/setup.py\n#\n# Installation and setup script for cmakeast\n#\n# See \/LICENCE.md for Copyright information\n\"\"\"Installation and setup script for cmakeast.\"\"\"\n\nfrom setuptools import find_packages, setup\n\nsetup(name=\"cmakeast\",\n version=\"0.0.17\",\n description=\"\"\"Parse a CMake file into an Abstract Syntax Tree.\"\"\",\n long_description_markdown_filename=\"README.md\",\n author=\"Sam Spilsbury\",\n author_email=\"smspillaz@gmail.com\",\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.1\",\n \"Programming Language :: Python :: 3.2\",\n \"Programming Language :: Python :: 3.3\",\n \"Programming Language :: Python :: 3.4\",\n \"Intended Audience :: Developers\",\n \"Topic :: Software Development :: Build Tools\",\n \"License :: OSI Approved :: MIT License\"],\n url=\"http:\/\/github.com\/polysquare\/cmake-ast\",\n license=\"MIT\",\n keywords=\"development ast cmake\",\n packages=find_packages(exclude=[\"test\"]),\n install_requires=[\"setuptools\"],\n extras_require={\n \"upload\": [\"setuptools-markdown\"]\n },\n entry_points={\n \"console_scripts\": [\n \"cmake-print-ast=cmakeast.printer:main\"\n ]\n },\n test_suite=\"nose.collector\",\n zip_safe=True,\n include_package_data=True)\n","subject":"Bump version: 0.0.16 -> 0.0.17","message":"Bump version: 0.0.16 -> 0.0.17\n\n[ci skip]\n","lang":"Python","license":"mit","repos":"polysquare\/cmake-ast"} {"commit":"64d4bcf0862e2715dc0de92a0621adf23dff5818","old_file":"source\/harmony\/schema\/collector.py","new_file":"source\/harmony\/schema\/collector.py","old_contents":"# :coding: utf-8\n# :copyright: Copyright (c) 2013 Martin Pengelly-Phillips\n# :license: See LICENSE.txt.\n\nfrom abc import ABCMeta, abstractmethod\n\n\nclass Collector(object):\n '''Collect and return schemas.'''\n\n __metaclass__ = ABCMeta\n\n @abstractmethod\n def collect(self):\n '''Yield collected schemas.\n\n Each schema should be a Python dictionary.\n\n '''\n\n","new_contents":"# :coding: utf-8\n# :copyright: Copyright (c) 2013 Martin Pengelly-Phillips\n# :license: See LICENSE.txt.\n\nimport os\nfrom abc import ABCMeta, abstractmethod\n\ntry:\n import json\nexcept ImportError:\n try:\n import simplejson as json\n except ImportError:\n raise ImportError('Could not import json or simplejson')\n\n\nclass Collector(object):\n '''Collect and return schemas.'''\n\n __metaclass__ = ABCMeta\n\n @abstractmethod\n def collect(self):\n '''Yield collected schemas.\n\n Each schema should be a Python dictionary.\n\n '''\n\n\nclass FilesystemCollector(Collector):\n\n def __init__(self, paths=None, recursive=True):\n '''Initialise with *paths* to search.\n\n If *recursive* is True then all subdirectories of *paths* will also be\n searched.\n\n '''\n self.paths = paths\n self.recursive = recursive\n if self.paths is None:\n self.paths = []\n super(FilesystemCollector, self).__init__()\n\n def collect(self):\n '''Yield collected schemas.'''\n for path in self.paths:\n for base, directories, filenames in os.walk(path):\n for filename in filenames:\n\n _, extension = os.path.splitext(filename)\n if extension != '.json':\n continue\n\n filepath = os.path.join(base, filename)\n with open(filepath, 'r') as file_handler:\n schema = json.load(file_handler)\n yield schema\n\n if not self.recursive:\n del directories[:]\n\n","subject":"Support collecting schemas from filesystem.","message":"Support collecting schemas from filesystem.\n","lang":"Python","license":"apache-2.0","repos":"4degrees\/harmony"} {"commit":"fa4ef8d171faaf89e81bcd36dfeec082cd86a9e8","old_file":"setup.py","new_file":"setup.py","old_contents":"from distutils.core import setup\n\nsetup(\n name = \"Cytoplasm\",\n version = \"0.04.5\",\n author = \"startling\",\n author_email = \"tdixon51793@gmail.com\",\n url = \"http:\/\/cytoplasm.somethingsido.com\",\n keywords = [\"blogging\", \"site compiler\", \"blog compiler\"],\n description = \"A static, blog-aware website generator written in python.\",\n packages = ['cytoplasm'],\n scripts = ['scripts\/cytoplasm'],\n install_requires = ['Mako'],\n classifiers = [\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet :: WWW\/HTTP :: Site Management\",\n \"Topic :: Internet :: WWW\/HTTP :: Dynamic Content :: News\/Diary\",\n ]\n)\n \n","new_contents":"from distutils.core import setup\n\nsetup(\n name = \"Cytoplasm\",\n version = \"0.05.0\",\n author = \"startling\",\n author_email = \"tdixon51793@gmail.com\",\n url = \"http:\/\/cytoplasm.somethingsido.com\",\n keywords = [\"blogging\", \"site compiler\", \"blog compiler\"],\n description = \"A static, blog-aware website generator written in python.\",\n packages = ['cytoplasm'],\n scripts = ['scripts\/cytoplasm'],\n install_requires = ['Mako'],\n classifiers = [\n \"Programming Language :: Python\",\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Topic :: Internet :: WWW\/HTTP :: Site Management\",\n \"Topic :: Internet :: WWW\/HTTP :: Dynamic Content :: News\/Diary\",\n ]\n)\n \n","subject":"Change version string to \"0.05.0\".","message":"Change version string to \"0.05.0\".\n\nThis is a medium-ish change because we kind of break compatibility with\nolder sites.\n","lang":"Python","license":"mit","repos":"startling\/cytoplasm"} {"commit":"5e9f41eae1d53cec5c90a104c249b11bfe292225","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python3\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name='django-slack',\n url=\"https:\/\/chris-lamb.co.uk\/projects\/django-slack\",\n version='5.16.0',\n description=\"Provides easy-to-use integration between Django projects and \"\n \"the Slack group chat and IM tool.\",\n author=\"Chris Lamb\",\n author_email=\"chris@chris-lamb.co.uk\",\n license=\"BSD-3-Clause\",\n packages=find_packages(),\n include_package_data=True,\n python_requires=\">=3.5\",\n install_requires=('Django>=2', 'requests'),\n)\n","new_contents":"#!\/usr\/bin\/env python3\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name='django-slack',\n url=\"https:\/\/chris-lamb.co.uk\/projects\/django-slack\",\n version='5.16.0',\n description=\"Provides easy-to-use integration between Django projects and \"\n \"the Slack group chat and IM tool.\",\n author=\"Chris Lamb\",\n author_email=\"chris@chris-lamb.co.uk\",\n license=\"BSD-3-Clause\",\n packages=find_packages(),\n include_package_data=True,\n python_requires=\">=3.6\",\n install_requires=('Django>=2', 'requests'),\n)\n","subject":"Make Python 3.6+ a hard requirement","message":"Make Python 3.6+ a hard requirement\n\nSupport for Python 3.5 was dropped in\nhttps:\/\/github.com\/lamby\/django-slack\/commit\/cd8cce9360fdceb0cb1b4228ef71c0b80a212a4b.\nPython 3.5 reached its EOL in September 2020, see\nhttps:\/\/www.python.org\/dev\/peps\/pep-0478\/.\n","lang":"Python","license":"bsd-3-clause","repos":"lamby\/django-slack"} {"commit":"c33038a85f41c2056630897c2b1b8ce590301e62","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom setuptools import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), 'README.md')).read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-swiftbrowser',\n version='0.1',\n packages=['swiftbrowser'],\n include_package_data=True,\n license='Apache License (2.0)',\n description='A simple Django app to access Openstack Swift',\n long_description=README,\n url='http:\/\/www.cschwede.com\/',\n author='Christian Schwede',\n author_email='info@cschwede.de',\n install_requires=['django', 'python-swiftclient'],\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache License (2.0)',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n)\n","new_contents":"import os\nfrom setuptools import setup\n\nREADME = open(os.path.join(os.path.dirname(__file__), 'README.md')).read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-swiftbrowser',\n version='0.1',\n packages=['swiftbrowser'],\n include_package_data=True,\n license='Apache License (2.0)',\n description='A simple Django app to access Openstack Swift',\n long_description=README,\n url='http:\/\/www.cschwede.com\/',\n author='Christian Schwede',\n author_email='info@cschwede.de',\n install_requires=['django>=1.5', 'python-swiftclient'],\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache License (2.0)',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n)\n","subject":"Add Django >= 1.5 requirement","message":"Add Django >= 1.5 requirement\n","lang":"Python","license":"apache-2.0","repos":"honza801\/django-swiftbrowser,hbhdytf\/django-swiftbrowser,hbhdytf\/django-swiftbrowser,bkawula\/django-swiftbrowser,honza801\/django-swiftbrowser,cschwede\/django-swiftbrowser,cschwede\/django-swiftbrowser,bkawula\/django-swiftbrowser,sunhongtao\/django-swiftbrowser,bkawula\/django-swiftbrowser,honza801\/django-swiftbrowser,bkawula\/django-swiftbrowser,sunhongtao\/django-swiftbrowser"} {"commit":"2349fa984e584fbfb022d01ce3d3e349dbe7870c","old_file":"setup.py","new_file":"setup.py","old_contents":"import textwrap\nfrom setuptools import setup\n\n\nsetup(name='dip',\n version='1.0.0b0',\n author='amancevice',\n author_email='smallweirdnum@gmail.com',\n packages=['dip'],\n url='http:\/\/www.smallweirdnumber.com',\n description='Install CLIs using docker-compose',\n long_description=textwrap.dedent(\n '''See GitHub_ for documentation.\n .. _GitHub: https:\/\/github.com\/amancevice\/dip'''),\n classifiers=['Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Topic :: Utilities',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python'],\n install_requires=['click>=6.7.0',\n 'colored>=1.3.4',\n 'docker-compose>=1.10.0',\n 'gitpython>=2.1.3'],\n entry_points={'console_scripts': ['dip=dip.main:dip']})\n","new_contents":"import textwrap\nfrom setuptools import setup\n\n\nsetup(name='dip',\n version='1.0.0b0',\n author='amancevice',\n author_email='smallweirdnum@gmail.com',\n packages=['dip'],\n url='http:\/\/www.smallweirdnumber.com',\n description='Install CLIs using docker-compose',\n long_description=textwrap.dedent(\n '''See GitHub_ for documentation.\n .. _GitHub: https:\/\/github.com\/amancevice\/dip'''),\n classifiers=['Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Topic :: Utilities',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python'],\n install_requires=['click>=6.7.0',\n 'colored>=1.3.4',\n 'docker-compose>=1.10.0',\n 'gitpython>=2.1.3'],\n entry_points={'console_scripts': ['dip=dip.main:dip']})\n","subject":"Remove Alpha from dev status.","message":"Remove Alpha from dev status.\n","lang":"Python","license":"mit","repos":"amancevice\/dip"} {"commit":"c7e726744b28ee0d502b51cdf8f7d6f0d9ef24e1","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\nurl = \"\"\nversion = \"0.1.0\"\nreadme = open('README.rst').read()\n\nsetup(\n name=\"dtool-create\",\n packages=[\"dtool_create\"],\n version=version,\n description=\"Dtool plugin for creating datasets and collections\",\n long_description=readme,\n include_package_data=True,\n author=\"Tjelvar Olsson\",\n author_email=\"tjelvar.olsson@jic.ac.uk\",\n url=url,\n install_requires=[\n \"Click\",\n \"click-plugins\",\n ],\n download_url=\"{}\/tarball\/{}\".format(url, version),\n license=\"MIT\"\n)\n","new_contents":"from setuptools import setup\n\nurl = \"\"\nversion = \"0.1.0\"\nreadme = open('README.rst').read()\n\nsetup(\n name=\"dtool-create\",\n packages=[\"dtool_create\"],\n version=version,\n description=\"Dtool plugin for creating datasets and collections\",\n long_description=readme,\n include_package_data=True,\n author=\"Tjelvar Olsson\",\n author_email=\"tjelvar.olsson@jic.ac.uk\",\n url=url,\n install_requires=[\n \"Click\",\n \"dtoolcore\",\n ],\n entry_points={\n \"dtool.dataset\": [\n \"create=dtool_create.dataset:create\",\n ],\n },\n download_url=\"{}\/tarball\/{}\".format(url, version),\n license=\"MIT\"\n)\n","subject":"Remove redundant click-plugins and add dtoolcore dependency; add entry point for dataset create","message":"Remove redundant click-plugins and add dtoolcore dependency; add entry point for dataset create\n","lang":"Python","license":"mit","repos":"jic-dtool\/dtool-create"} {"commit":"e42dc9fe06bad4eff195d64f031d5a2445a15cc8","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nfrom __future__ import absolute_import, division, print_function\nimport six\nfrom setuptools import setup, Command\n\nimport logging\nlogging.basicConfig(level=logging.DEBUG)\nlogging.getLogger(\"nose\").setLevel(logging.DEBUG)\n\nsetup_requires=[\"nose>=1.0\", \"coverage>=4.0\", \"Sphinx>=1.3\"]\nif six.PY2:\n setup_requires.append(\"Sphinx-PyPI-upload>=0.2\")\n\nsetup(\n name=\"awssig\",\n version=\"0.2\",\n packages=['awssig'],\n install_requires=[\"six>=1.0\"],\n setup_requires=setup_requires,\n\n # PyPI information\n author=\"David Cuthbert\",\n author_email=\"dacut@kanga.org\",\n description=\"AWS signature verification routines\",\n license=\"Apache 2.0\",\n url=\"https:\/\/github.com\/dacut\/python-aws-sig\",\n classifiers=[\n 'Development Status :: 2 - Pre-Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache Software License',\n 'Programming Language :: Python',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n keywords = ['aws', 'signature'],\n zip_safe=False,\n)\n","new_contents":"#!\/usr\/bin\/env python\nfrom __future__ import absolute_import, division, print_function\nimport six\nfrom setuptools import setup, Command\n\nimport logging\nlogging.basicConfig(level=logging.DEBUG)\nlogging.getLogger(\"nose\").setLevel(logging.DEBUG)\n\nsetup_requires=[\"nose>=1.0\", \"coverage>=4.0\", \"Sphinx>=1.3\"]\nif six.PY2:\n setup_requires.append(\"Sphinx-PyPI-upload>=0.2\")\n\nsetup(\n name=\"awssig\",\n version=\"0.2.1\",\n packages=['awssig'],\n install_requires=[\"six>=1.0\"],\n setup_requires=setup_requires,\n\n # PyPI information\n author=\"David Cuthbert\",\n author_email=\"dacut@kanga.org\",\n description=\"AWS signature verification routines\",\n license=\"Apache 2.0\",\n url=\"https:\/\/github.com\/dacut\/python-aws-sig\",\n classifiers=[\n 'Development Status :: 2 - Pre-Alpha',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache Software License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 3',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n keywords = ['aws', 'signature'],\n zip_safe=False,\n)\n","subject":"Update supported languages to indicate Python 3+ support.","message":"Update supported languages to indicate Python 3+ support.\n","lang":"Python","license":"apache-2.0","repos":"dacut\/python-aws-sig"} {"commit":"3e403ed0962b62b19247dd1021f672507d003f07","old_file":"labware\/microplates.py","new_file":"labware\/microplates.py","old_contents":"from .grid import GridContainer, GridItem\nfrom .liquids import LiquidContainer, LiquidWell\n\n\nclass Microplate(GridContainer):\n rows = 8\n cols = 12\n volume = 100\n min_vol = 50\n max_vol = 90\n height = 14.45\n length = 127.76\n width = 85.47\n diameter = 7.15\n depth = 3.25\n a1_x = 14.38\n a1_y = 11.24\n spacing = 9\n\n child_class = LiquidWell\n\n def well(self, position):\n return self.get_child(position)\n\n def calibrate(self, **kwargs):\n \"\"\"\n Coordinates should represent the center and near-bottom of well\n A1 with the pipette tip in place.\n \"\"\"\n super(Microplate, self).calibrate(**kwargs)\n\n\nclass Microplate_96(Microplate):\n pass\n\n\nclass Microplate_96_deepwell(Microplate_96):\n volume = 400\n min_vol = 50\n max_vol = 380\n height = 14.6\n depth = 10.8\n","new_contents":"from .grid import GridContainer, GridItem\nfrom .liquids import LiquidContainer, LiquidWell\n\n\nclass Microplate(GridContainer, LiquidContainer):\n rows = 8\n cols = 12\n volume = 100\n min_vol = 50\n max_vol = 90\n height = 14.45\n length = 127.76\n width = 85.47\n diameter = 7.15\n depth = 3.25\n a1_x = 14.38\n a1_y = 11.24\n spacing = 9\n\n def well(self, position):\n return self.get_child(position)\n\n def calibrate(self, **kwargs):\n \"\"\"\n Coordinates should represent the center and near-bottom of well\n A1 with the pipette tip in place.\n \"\"\"\n super(Microplate, self).calibrate(**kwargs)\n\n\nclass Microplate_96(Microplate):\n pass\n\n\nclass Microplate_96_deepwell(Microplate_96):\n volume = 400\n min_vol = 50\n max_vol = 380\n height = 14.6\n depth = 10.8\n","subject":"Refactor of LiquidContainer vs. GridContainer to support grids which don't contain liquids.","message":"Refactor of LiquidContainer vs. GridContainer to support grids which don't contain liquids.\n","lang":"Python","license":"apache-2.0","repos":"OpenTrons\/opentrons-api,OpenTrons\/opentrons-api,OpenTrons\/opentrons_sdk,OpenTrons\/opentrons-api,OpenTrons\/opentrons-api,OpenTrons\/opentrons-api,Opentrons\/labware"} {"commit":"872f76e48b5683d3562b8bcd228e681e6a5d7d80","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\nimport os\n\nsetup(\n name = \"merky\",\n version = \"0.0.3a\",\n author = \"Ethan Rowe\",\n author_email = \"ethan@the-rowes.com\",\n description = (\"JSON-oriented merkle tree utilities\"),\n license = \"MIT\",\n url = \"https:\/\/github.com\/ethanrowe\/python-merky\",\n packages = [\"merky\",\n \"merky.cases\",\n \"merky.store\",\n \"merky.test\",\n \"merky.test.misc\",\n \"merky.test.store\",\n \"merky.test.usecases\",\n ],\n long_description = \"\"\"\nMerky - compute merkle trees for JSON-friendly data.\n\"\"\",\n test_suite = \"nose.collector\",\n install_requires = [\n 'six >= 1.5',\n ],\n setup_requires = [\n 'nose',\n 'mock >= 1.0.1',\n ],\n tests_require = [\n 'nose',\n 'mock >= 1.0.1',\n ],\n)\n","new_contents":"from setuptools import setup\nimport os\n\nsetup(\n name = \"merky\",\n version = \"0.0.4\",\n author = \"Ethan Rowe\",\n author_email = \"ethan@the-rowes.com\",\n description = (\"JSON-oriented merkle tree utilities\"),\n license = \"MIT\",\n url = \"https:\/\/github.com\/ethanrowe\/python-merky\",\n packages = [\"merky\",\n \"merky.cases\",\n \"merky.store\",\n \"merky.test\",\n \"merky.test.misc\",\n \"merky.test.store\",\n \"merky.test.usecases\",\n ],\n long_description = \"\"\"\nMerky - compute merkle trees for JSON-friendly data.\n\"\"\",\n test_suite = \"nose.collector\",\n install_requires = [\n 'six >= 1.5',\n ],\n setup_requires = [\n 'nose',\n 'mock >= 1.0.1',\n ],\n tests_require = [\n 'nose',\n 'mock >= 1.0.1',\n ],\n)\n","subject":"Bump to 0.0.4 for tokenized structure walker","message":"Bump to 0.0.4 for tokenized structure walker\n","lang":"Python","license":"mit","repos":"ethanrowe\/python-merky"} {"commit":"17f236f8c8f56f525801f6b07cac22a7de4b4ad0","old_file":"setup.py","new_file":"setup.py","old_contents":"# \/setup.py\n#\n# Installation and setup script for parse-shebang\n#\n# See \/LICENCE.md for Copyright information\n\"\"\"Installation and setup script for parse-shebang.\"\"\"\n\nfrom setuptools import find_packages, setup\n\nsetup(name=\"parse-shebang\",\n version=\"0.0.13\",\n description=\"\"\"Parse shebangs and return their components.\"\"\",\n long_description_markdown_filename=\"README.md\",\n author=\"Sam Spilsbury\",\n author_email=\"smspillaz@gmail.com\",\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.1\",\n \"Programming Language :: Python :: 3.2\",\n \"Programming Language :: Python :: 3.3\",\n \"Programming Language :: Python :: 3.4\",\n \"Intended Audience :: Developers\",\n \"Topic :: System :: Shells\",\n \"Topic :: Utilities\",\n \"License :: OSI Approved :: MIT License\"],\n url=\"http:\/\/github.com\/polysquare\/python-parse-shebang\",\n license=\"MIT\",\n keywords=\"development\",\n packages=find_packages(exclude=[\"test\"]),\n install_requires=[\"setuptools\"],\n extras_require={\n \"upload\": [\"setuptools-markdown\"]\n },\n test_suite=\"nose.collector\",\n zip_safe=True,\n include_package_data=True)\n","new_contents":"# \/setup.py\n#\n# Installation and setup script for parse-shebang\n#\n# See \/LICENCE.md for Copyright information\n\"\"\"Installation and setup script for parse-shebang.\"\"\"\n\nfrom setuptools import find_packages, setup\n\nsetup(name=\"parse-shebang\",\n version=\"0.0.14\",\n description=\"\"\"Parse shebangs and return their components.\"\"\",\n long_description_markdown_filename=\"README.md\",\n author=\"Sam Spilsbury\",\n author_email=\"smspillaz@gmail.com\",\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python :: 2\",\n \"Programming Language :: Python :: 2.7\",\n \"Programming Language :: Python :: 3\",\n \"Programming Language :: Python :: 3.1\",\n \"Programming Language :: Python :: 3.2\",\n \"Programming Language :: Python :: 3.3\",\n \"Programming Language :: Python :: 3.4\",\n \"Intended Audience :: Developers\",\n \"Topic :: System :: Shells\",\n \"Topic :: Utilities\",\n \"License :: OSI Approved :: MIT License\"],\n url=\"http:\/\/github.com\/polysquare\/python-parse-shebang\",\n license=\"MIT\",\n keywords=\"development\",\n packages=find_packages(exclude=[\"test\"]),\n install_requires=[\"setuptools\"],\n extras_require={\n \"upload\": [\"setuptools-markdown\"]\n },\n test_suite=\"nose.collector\",\n zip_safe=True,\n include_package_data=True)\n","subject":"Bump version: 0.0.13 -> 0.0.14","message":"Bump version: 0.0.13 -> 0.0.14\n\n[ci skip]\n","lang":"Python","license":"mit","repos":"polysquare\/python-parse-shebang"} {"commit":"605af19623b2536752304275f2403ce2b4fa52f8","old_file":"bluebottle\/test\/factory_models\/projects.py","new_file":"bluebottle\/test\/factory_models\/projects.py","old_contents":"import factory\nimport logging\n\nfrom django.conf import settings\n\nfrom bluebottle.projects.models import (\n Project, ProjectTheme, ProjectDetailField, ProjectBudgetLine)\nfrom .accounts import BlueBottleUserFactory\n\n# Suppress debug information for Factory Boy\nlogging.getLogger('factory').setLevel(logging.WARN)\n\n\nclass ProjectFactory(factory.DjangoModelFactory):\n FACTORY_FOR = Project\n\n owner = factory.SubFactory(BlueBottleUserFactory)\n phase = settings.PROJECT_PHASES[0][1][0][0]\n title = factory.Sequence(lambda n: 'Project_{0}'.format(n))\n\n\nclass ProjectThemeFactory(factory.DjangoModelFactory):\n FACTORY_FOR = ProjectTheme\n\n name = factory.Sequence(lambda n: 'Theme_{0}'.format(n))\n name_nl = name\n slug = name\n description = 'ProjectTheme factory model'\n\n\nclass ProjectDetailFieldFactory(factory.DjangoModelFactory):\n FACTORY_FOR = ProjectDetailField\n\n name = factory.Sequence(lambda n: 'Field_{0}'.format(n))\n description = 'DetailField factory model'\n slug = name\n type = 'text'\n\n\nclass ProjectBudgetLineFactory(factory.DjangoModelFactory):\n FACTORY_FOR = ProjectBudgetLine\n\n project = factory.SubFactory(ProjectFactory)\n amount = 100000\n","new_contents":"import factory\nimport logging\n\nfrom django.conf import settings\n\nfrom bluebottle.projects.models import (\n Project, ProjectTheme, ProjectDetailField, ProjectBudgetLine)\nfrom .accounts import BlueBottleUserFactory\n\n# Suppress debug information for Factory Boy\nlogging.getLogger('factory').setLevel(logging.WARN)\n\n\nclass ProjectFactory(factory.DjangoModelFactory):\n FACTORY_FOR = Project\n\n owner = factory.SubFactory(BlueBottleUserFactory)\n title = factory.Sequence(lambda n: 'Project_{0}'.format(n))\n\n\nclass ProjectThemeFactory(factory.DjangoModelFactory):\n FACTORY_FOR = ProjectTheme\n\n name = factory.Sequence(lambda n: 'Theme_{0}'.format(n))\n name_nl = name\n slug = name\n description = 'ProjectTheme factory model'\n\n\nclass ProjectDetailFieldFactory(factory.DjangoModelFactory):\n FACTORY_FOR = ProjectDetailField\n\n name = factory.Sequence(lambda n: 'Field_{0}'.format(n))\n description = 'DetailField factory model'\n slug = name\n type = 'text'\n\n\nclass ProjectBudgetLineFactory(factory.DjangoModelFactory):\n FACTORY_FOR = ProjectBudgetLine\n\n project = factory.SubFactory(ProjectFactory)\n amount = 100000\n","subject":"Remove phase from project factory","message":"Remove phase from project factory\n","lang":"Python","license":"bsd-3-clause","repos":"onepercentclub\/bluebottle,jfterpstra\/bluebottle,onepercentclub\/bluebottle,onepercentclub\/bluebottle,jfterpstra\/bluebottle,onepercentclub\/bluebottle,onepercentclub\/bluebottle,jfterpstra\/bluebottle,jfterpstra\/bluebottle"} {"commit":"3cc8dc8de4fe75f0abf92f89979b0b7e48d9c137","old_file":"splearn\/ensemble\/tests\/__init__.py","new_file":"splearn\/ensemble\/tests\/__init__.py","old_contents":"import numpy as np\nfrom nose.tools import assert_true\nfrom sklearn.ensemble import RandomForestClassifier\nfrom splearn.ensemble import SparkRandomForestClassifier\nfrom splearn.utils.testing import SplearnTestCase\nfrom splearn.utils.validation import check_rdd_dtype\n\n\nclass TestSparkRandomForest(SplearnTestCase):\n\n def test_same_predictions(self):\n X, y, Z = self.make_classification(2, 10000)\n\n local = RandomForestClassifier()\n dist = SparkRandomForestClassifier()\n\n y_local = local.fit(X, y).predict(X)\n y_dist = dist.fit(Z, classes=np.unique(y)).predict(Z[:, 'X'])\n y_conv = dist.to_scikit().predict(X)\n\n assert_true(check_rdd_dtype(y_dist, (np.ndarray,)))\n assert(sum(y_local != y_dist.toarray()) < len(y_local) * 2.\/100.)\n assert(sum(y_local != y_conv) < len(y_local) * 2.\/100.)\n","new_contents":"import numpy as np\nfrom nose.tools import assert_true\nfrom sklearn.ensemble import RandomForestClassifier\nfrom splearn.ensemble import SparkRandomForestClassifier\nfrom splearn.utils.testing import SplearnTestCase\nfrom splearn.utils.validation import check_rdd_dtype\n\n\nclass TestSparkRandomForest(SplearnTestCase):\n\n def test_same_predictions(self):\n X, y, Z = self.make_classification(2, 10000)\n\n local = RandomForestClassifier()\n dist = SparkRandomForestClassifier()\n\n y_local = local.fit(X, y).predict(X)\n y_dist = dist.fit(Z, classes=np.unique(y)).predict(Z[:, 'X'])\n y_conv = dist.to_scikit().predict(X)\n\n assert_true(check_rdd_dtype(y_dist, (np.ndarray,)))\n assert(sum(y_local != y_dist.toarray()) < len(y_local) * 5.\/100.)\n assert(sum(y_local != y_conv) < len(y_local) * 5.\/100.)\n","subject":"Raise allowed difference for RandomForest","message":"Raise allowed difference for RandomForest\n","lang":"Python","license":"apache-2.0","repos":"lensacom\/sparkit-learn,taynaud\/sparkit-learn,taynaud\/sparkit-learn,lensacom\/sparkit-learn"} {"commit":"451951b311ef6e2bb76348a116dc0465f735348e","old_file":"pytest_watch\/config.py","new_file":"pytest_watch\/config.py","old_contents":"try:\n from configparser import ConfigParser\nexcept ImportError:\n from ConfigParser import ConfigParser\n\nimport pytest\n\n\nCLI_OPTION_PREFIX = '--'\n\n\nclass CollectConfig(object):\n \"\"\"\n A pytest plugin to gets the configuration file.\n \"\"\"\n def __init__(self):\n self.path = None\n\n def pytest_cmdline_main(self, config):\n self.path = str(config.inifile)\n\n\ndef merge_config(args):\n collect_config = CollectConfig()\n pytest.main(['--collect-only'], plugins=[collect_config])\n if not collect_config.path:\n return\n\n config = ConfigParser()\n config.read(collect_config.path)\n if not config.has_section('pytest-watch'):\n return\n\n for cli_name in args:\n if not cli_name.startswith(CLI_OPTION_PREFIX):\n continue\n config_name = cli_name[len(CLI_OPTION_PREFIX):]\n\n # Let CLI options take precedence\n if args[cli_name]:\n continue\n\n # Find config option\n if not config.has_option('pytest-watch', config_name):\n continue\n\n # Merge config option using the expected type\n if isinstance(args[cli_name], bool):\n args[cli_name] = config.getboolean('pytest-watch', config_name)\n else:\n args[cli_name] = config.get('pytest-watch', config_name)\n","new_contents":"try:\n from configparser import ConfigParser\nexcept ImportError:\n from ConfigParser import ConfigParser\n\nimport pytest\n\n\nCLI_OPTION_PREFIX = '--'\n\n\nclass CollectConfig(object):\n \"\"\"\n A pytest plugin to gets the configuration file.\n \"\"\"\n def __init__(self):\n self.path = None\n\n def pytest_cmdline_main(self, config):\n if config.inifile:\n self.path = str(config.inifile)\n\n\ndef merge_config(args):\n collect_config = CollectConfig()\n pytest.main(['--collect-only'], plugins=[collect_config])\n if not collect_config.path:\n return\n\n config = ConfigParser()\n config.read(collect_config.path)\n if not config.has_section('pytest-watch'):\n return\n\n for cli_name in args:\n if not cli_name.startswith(CLI_OPTION_PREFIX):\n continue\n config_name = cli_name[len(CLI_OPTION_PREFIX):]\n\n # Let CLI options take precedence\n if args[cli_name]:\n continue\n\n # Find config option\n if not config.has_option('pytest-watch', config_name):\n continue\n\n # Merge config option using the expected type\n if isinstance(args[cli_name], bool):\n args[cli_name] = config.getboolean('pytest-watch', config_name)\n else:\n args[cli_name] = config.get('pytest-watch', config_name)\n","subject":"Fix running when pytest.ini is not present.","message":"Fix running when pytest.ini is not present.\n","lang":"Python","license":"mit","repos":"joeyespo\/pytest-watch"} {"commit":"b690d4be60d9c72eb805b779c61f6d3001881bd2","old_file":"raven\/conf\/__init__.py","new_file":"raven\/conf\/__init__.py","old_contents":"\"\"\"\nraven.conf\n~~~~~~~~~~\n\n:copyright: (c) 2010 by the Sentry Team, see AUTHORS for more details.\n:license: BSD, see LICENSE for more details.\n\"\"\"\n\nimport urlparse\n\n\ndef load(dsn, scope):\n \"\"\"\n Parses a Sentry compatible DSN and loads it\n into the given scope.\n\n >>> import raven\n >>> dsn = 'https:\/\/public_key:secret_key@sentry.local\/project_id'\n >>> raven.load(dsn, locals())\n \"\"\"\n url = urlparse.urlparse(dsn)\n if url.scheme not in ('http', 'https'):\n raise ValueError('Unsupported Sentry DSN scheme: %r' % url.scheme)\n netloc = url.hostname\n if url.port and url.port != 80:\n netloc += ':%s' % url.port\n path_bits = url.path.rsplit('\/', 1)\n if len(path_bits) > 1:\n path = path_bits[0]\n else:\n path = ''\n project = path_bits[-1]\n scope.update({\n 'SENTRY_SERVERS': ['%s:\/\/%s%s\/api\/store\/' % (url.scheme, netloc, path)],\n 'SENTRY_PROJECT': project,\n 'SENTRY_PUBLIC_KEY': url.username,\n 'SENTRY_SECRET_KEY': url.password,\n })\n","new_contents":"\"\"\"\nraven.conf\n~~~~~~~~~~\n\n:copyright: (c) 2010 by the Sentry Team, see AUTHORS for more details.\n:license: BSD, see LICENSE for more details.\n\"\"\"\n\nimport urlparse\n\n\ndef load(dsn, scope):\n \"\"\"\n Parses a Sentry compatible DSN and loads it\n into the given scope.\n\n >>> import raven\n >>> dsn = 'https:\/\/public_key:secret_key@sentry.local\/project_id'\n >>> raven.load(dsn, locals())\n \"\"\"\n url = urlparse.urlparse(dsn)\n if url.scheme not in ('http', 'https', 'udp'):\n raise ValueError('Unsupported Sentry DSN scheme: %r' % url.scheme)\n netloc = url.hostname\n if url.port and url.port != 80:\n netloc += ':%s' % url.port\n path_bits = url.path.rsplit('\/', 1)\n if len(path_bits) > 1:\n path = path_bits[0]\n else:\n path = ''\n project = path_bits[-1]\n scope.update({\n 'SENTRY_SERVERS': ['%s:\/\/%s%s\/api\/store\/' % (url.scheme, netloc, path)],\n 'SENTRY_PROJECT': project,\n 'SENTRY_PUBLIC_KEY': url.username,\n 'SENTRY_SECRET_KEY': url.password,\n })\n","subject":"Support udp scheme in raven.load","message":"Support udp scheme in raven.load\n","lang":"Python","license":"bsd-3-clause","repos":"Goldmund-Wyldebeast-Wunderliebe\/raven-python,someonehan\/raven-python,dbravender\/raven-python,jbarbuto\/raven-python,ticosax\/opbeat_python,akheron\/raven-python,smarkets\/raven-python,getsentry\/raven-python,patrys\/opbeat_python,jmp0xf\/raven-python,jmagnusson\/raven-python,recht\/raven-python,hzy\/raven-python,akalipetis\/raven-python,danriti\/raven-python,tarkatronic\/opbeat_python,recht\/raven-python,nikolas\/raven-python,ewdurbin\/raven-python,nikolas\/raven-python,ewdurbin\/raven-python,jbarbuto\/raven-python,Photonomie\/raven-python,tarkatronic\/opbeat_python,arthurlogilab\/raven-python,danriti\/raven-python,smarkets\/raven-python,johansteffner\/raven-python,ewdurbin\/raven-python,lepture\/raven-python,ronaldevers\/raven-python,akheron\/raven-python,nikolas\/raven-python,openlabs\/raven,recht\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,inspirehep\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,beniwohli\/apm-agent-python,inspirehep\/raven-python,johansteffner\/raven-python,akalipetis\/raven-python,inspirehep\/raven-python,nikolas\/raven-python,ticosax\/opbeat_python,johansteffner\/raven-python,lopter\/raven-python-old,lepture\/raven-python,jbarbuto\/raven-python,someonehan\/raven-python,ronaldevers\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,dbravender\/raven-python,smarkets\/raven-python,icereval\/raven-python,percipient\/raven-python,dbravender\/raven-python,percipient\/raven-python,lepture\/raven-python,hzy\/raven-python,collective\/mr.poe,akheron\/raven-python,jmp0xf\/raven-python,icereval\/raven-python,dirtycoder\/opbeat_python,Photonomie\/raven-python,jmagnusson\/raven-python,daikeren\/opbeat_python,beniwohli\/apm-agent-python,daikeren\/opbeat_python,arthurlogilab\/raven-python,inspirehep\/raven-python,icereval\/raven-python,danriti\/raven-python,jbarbuto\/raven-python,hzy\/raven-python,akalipetis\/raven-python,daikeren\/opbeat_python,patrys\/opbeat_python,getsentry\/raven-python,someonehan\/raven-python,ronaldevers\/raven-python,smarkets\/raven-python,dirtycoder\/opbeat_python,beniwohli\/apm-agent-python,patrys\/opbeat_python,jmp0xf\/raven-python,beniwohli\/apm-agent-python,ticosax\/opbeat_python,alex\/raven,percipient\/raven-python,icereval\/raven-python,arthurlogilab\/raven-python,arthurlogilab\/raven-python,dirtycoder\/opbeat_python,jmagnusson\/raven-python,tarkatronic\/opbeat_python,getsentry\/raven-python,Photonomie\/raven-python,patrys\/opbeat_python"} {"commit":"0f40869157ef56df0ff306fb510be4401b5cbe5d","old_file":"test\/low_level\/test_frame_identifiers.py","new_file":"test\/low_level\/test_frame_identifiers.py","old_contents":"import inspect\n\nfrom pyinstrument.low_level import stat_profile as stat_profile_c\nfrom pyinstrument.low_level import stat_profile_python\n\n\nclass AClass:\n def get_frame_identfier_for_a_method(self, getter_function):\n frame = inspect.currentframe()\n assert frame\n return getter_function(frame)\n\n @classmethod\n def get_frame_identfier_for_a_class_method(cls, getter_function):\n frame = inspect.currentframe()\n assert frame\n return getter_function(frame)\n\n\ndef test_frame_identifier():\n frame = inspect.currentframe()\n\n assert frame\n assert stat_profile_c.get_frame_identifier(frame) == stat_profile_python.get_frame_identifier(\n frame\n )\n\n\ndef test_frame_identifier_for_method():\n instance = AClass()\n assert instance.get_frame_identfier_for_a_method(\n stat_profile_c.get_frame_identifier\n ) == instance.get_frame_identfier_for_a_method(stat_profile_python.get_frame_identifier)\n","new_contents":"import inspect\n\nfrom pyinstrument.low_level import stat_profile as stat_profile_c\nfrom pyinstrument.low_level import stat_profile_python\n\n\nclass AClass:\n def get_frame_identifier_for_a_method(self, getter_function):\n frame = inspect.currentframe()\n assert frame\n return getter_function(frame)\n\n def get_frame_identifier_with_cell_variable(self, getter_function):\n frame = inspect.currentframe()\n assert frame\n\n def an_inner_function():\n # reference self to make it a cell variable\n if self:\n pass\n\n return getter_function(frame)\n\n @classmethod\n def get_frame_identifier_for_a_class_method(cls, getter_function):\n frame = inspect.currentframe()\n assert frame\n return getter_function(frame)\n\n\ndef test_frame_identifier():\n frame = inspect.currentframe()\n\n assert frame\n assert stat_profile_c.get_frame_identifier(frame) == stat_profile_python.get_frame_identifier(\n frame\n )\n\n\ndef test_frame_identifiers():\n instance = AClass()\n\n test_functions = [\n instance.get_frame_identifier_for_a_method,\n AClass.get_frame_identifier_for_a_class_method,\n instance.get_frame_identifier_with_cell_variable,\n ]\n\n for test_function in test_functions:\n assert test_function(stat_profile_c.get_frame_identifier) == test_function(\n stat_profile_python.get_frame_identifier\n )\n","subject":"Add test for a cell variable","message":"Add test for a cell variable\n","lang":"Python","license":"bsd-3-clause","repos":"joerick\/pyinstrument,joerick\/pyinstrument,joerick\/pyinstrument,joerick\/pyinstrument,joerick\/pyinstrument,joerick\/pyinstrument"} {"commit":"8a607ab7c064f3f593b4cecaa8e4262bec9326fa","old_file":"corehq\/apps\/es\/forms.py","new_file":"corehq\/apps\/es\/forms.py","old_contents":"from .es_query import HQESQuery\nfrom . import filters\n\n\nclass FormES(HQESQuery):\n index = 'forms'\n default_filters = {\n 'is_xform_instance': {\"term\": {\"doc_type\": \"xforminstance\"}},\n 'has_xmlns': {\"not\": {\"missing\": {\"field\": \"xmlns\"}}},\n 'has_user': {\"not\": {\"missing\": {\"field\": \"form.meta.userID\"}}},\n }\n @property\n def builtin_filters(self):\n return [\n xmlns,\n app,\n submitted,\n completed,\n user_id,\n ] + super(FormES, self).builtin_filters\n\n def user_facet(self):\n return self.terms_facet('form.meta.userID', 'user')\n\n def domain_facet(self):\n return self.terms_facet('domain', 'domain')\n\n\ndef xmlns(xmlns):\n return filters.term('xmlns.exact', xmlns)\n\n\ndef app(app_id):\n return filters.term('app_id', app_id)\n\n\ndef submitted(gt=None, gte=None, lt=None, lte=None):\n return filters.date_range('received_on', gt, gte, lt, lte)\n\n\ndef completed(gt=None, gte=None, lt=None, lte=None):\n return filters.date_range('form.meta.timeEnd', gt, gte, lt, lte)\n\n\ndef user_id(user_ids):\n return filters.term('form.meta.userID', list(user_ids))\n","new_contents":"from .es_query import HQESQuery\nfrom . import filters\n\n\nclass FormES(HQESQuery):\n index = 'forms'\n default_filters = {\n 'is_xform_instance': {\"term\": {\"doc_type\": \"xforminstance\"}},\n 'has_xmlns': {\"not\": {\"missing\": {\"field\": \"xmlns\"}}},\n 'has_user': {\"not\": {\"missing\": {\"field\": \"form.meta.userID\"}}},\n }\n @property\n def builtin_filters(self):\n return [\n xmlns,\n app,\n submitted,\n completed,\n user_id,\n ] + super(FormES, self).builtin_filters\n\n def user_facet(self):\n return self.terms_facet('form.meta.userID', 'user')\n\n def domain_facet(self):\n return self.terms_facet('domain', 'domain', 1000000)\n\n\ndef xmlns(xmlns):\n return filters.term('xmlns.exact', xmlns)\n\n\ndef app(app_id):\n return filters.term('app_id', app_id)\n\n\ndef submitted(gt=None, gte=None, lt=None, lte=None):\n return filters.date_range('received_on', gt, gte, lt, lte)\n\n\ndef completed(gt=None, gte=None, lt=None, lte=None):\n return filters.date_range('form.meta.timeEnd', gt, gte, lt, lte)\n\n\ndef user_id(user_ids):\n return filters.term('form.meta.userID', list(user_ids))\n","subject":"Return more than 10 domains","message":"Return more than 10 domains\n","lang":"Python","license":"bsd-3-clause","repos":"qedsoftware\/commcare-hq,puttarajubr\/commcare-hq,puttarajubr\/commcare-hq,puttarajubr\/commcare-hq,dimagi\/commcare-hq,dimagi\/commcare-hq,qedsoftware\/commcare-hq,qedsoftware\/commcare-hq,qedsoftware\/commcare-hq,dimagi\/commcare-hq,puttarajubr\/commcare-hq,dimagi\/commcare-hq,dimagi\/commcare-hq,qedsoftware\/commcare-hq"} {"commit":"0d3ae906a9382aca5e00ff403986591223e3a1a7","old_file":"pic\/flash.py","new_file":"pic\/flash.py","old_contents":"\nclass ProgramMemory(object):\n\n def __init__(self, size: int, programCounter):\n self._operations = [None] * size\n self._programCounter = programCounter\n\n @property\n def operations(self) -> list:\n return self._operations\n\n @property\n def programCounter(self):\n return self._programCounter\n\n def nextOp(self):\n return self.operations[self.programCounter.value]\n","new_contents":"\nclass ProgramMemory(object):\n\n def __init__(self, size: int, programCounter):\n self._operations = [None] * size\n self._programCounter = programCounter\n\n @property\n def operations(self) -> list:\n return self._operations\n\n @property\n def programCounter(self):\n return self._programCounter\n\n def nextOp(self):\n return self.operations[self.programCounter.address]\n","subject":"Change program counter current address reference.","message":"Change program counter current address reference.\n","lang":"Python","license":"mit","repos":"SuperOxigen\/pic16f917-simulator"} {"commit":"d1da03b40e9a10a07b67eeb76a0bef8cc704a40c","old_file":"utils\/exporter.py","new_file":"utils\/exporter.py","old_contents":"import plotly as py\nfrom os import makedirs\n\nfrom utils.names import output_file_name\n\n_out_dir = 'graphs\/'\n\ndef export(fig, module, dates):\n graph_dir = '{}{}\/'.format(_out_dir, str(module))\n makedirs(graph_dir, exist_ok=True)\n py.offline.plot(fig, filename=graph_dir + output_file_name(module, dates))\n\n","new_contents":"import plotly as py\nfrom os import makedirs\n\nfrom utils.names import output_file_name\n\n_out_dir = 'graphs\/'\n\ndef export(fig, module, dates):\n graph_dir = '{}{}\/'.format(_out_dir, str(module))[:-3] # remove .py extension from dir names\n makedirs(graph_dir, exist_ok=True)\n py.offline.plot(fig, filename=graph_dir + output_file_name(module, dates))\n\n","subject":"Remove .py extension from graph dir names","message":"Remove .py extension from graph dir names\n","lang":"Python","license":"mit","repos":"f-jiang\/sleep-pattern-grapher"} {"commit":"2f0d71024ab9fb3ab4b97e086741b0dd8564d29e","old_file":"dadd\/master\/handlers.py","new_file":"dadd\/master\/handlers.py","old_contents":"from flask import send_file, request, redirect, url_for\n\nfrom dadd.master import app\nfrom dadd.master.files import FileStorage\n\n\n@app.route('\/')\ndef index():\n return redirect(url_for('admin.index'))\n\n\n@app.route('\/files\/<path>', methods=['PUT', 'GET'])\ndef files(path):\n storage = FileStorage(app.config['STORAGE_DIR'])\n\n if request.method == 'PUT':\n storage.save(path, request.stream)\n resp = app.make_response('Updated %s' % path)\n resp.status_code = 202\n return resp\n\n return send_file(storage.read(path))\n","new_contents":"from flask import send_file, request, redirect, url_for\n\nfrom dadd.master import app\nfrom dadd.master.files import FileStorage\n\n\n@app.route('\/')\ndef index():\n return redirect(url_for('admin.index'))\n\n\n@app.route('\/files\/<path:path>', methods=['PUT', 'GET'])\ndef files(path):\n storage = FileStorage(app.config['STORAGE_DIR'])\n\n if request.method == 'PUT':\n storage.save(path, request.stream)\n resp = app.make_response('Updated %s' % path)\n resp.status_code = 202\n return resp\n\n return send_file(storage.read(path))\n","subject":"Fix files path argument type.","message":"Fix files path argument type.\n","lang":"Python","license":"bsd-3-clause","repos":"ionrock\/dadd,ionrock\/dadd,ionrock\/dadd,ionrock\/dadd"} {"commit":"8f753ba1bc3f17146841d1f83f493adb8ea480e1","old_file":"voteit\/stancer.py","new_file":"voteit\/stancer.py","old_contents":"\n\n\n\ndef generate_stances(blocs=[], filters={}):\n\n return \"banana!\"\n","new_contents":"from bson.code import Code\n\nfrom voteit.core import votes\n\nREDUCE = Code(\"\"\"\nfunction(obj, prev) {\n if (!prev.votes.hasOwnProperty(obj.option)) {\n prev.votes[obj.option] = 1;\n } else {\n prev.votes[obj.option]++;\n }\n \/\/prev.count++;\n};\n\"\"\")\n\n\ndef generate_stances(blocs=[], filters={}):\n data = votes.group(blocs, filters, {\"votes\": {}}, REDUCE)\n print data\n return data\n","subject":"Replace “banana” with business logic (kind of).","message":"Replace “banana” with business logic (kind of). ","lang":"Python","license":"mit","repos":"tmtmtmtm\/voteit-api,pudo\/voteit-server,pudo-attic\/voteit-server"} {"commit":"55069f1635a32c57faf7c8afcbfc9b88df093601","old_file":"bin\/monitor\/find_invalid_pipeline_state.py","new_file":"bin\/monitor\/find_invalid_pipeline_state.py","old_contents":"import arrow\nimport logging\nimport argparse\nimport emission.core.wrapper.pipelinestate as ecwp\nimport emission.core.get_database as edb\n\n# Run in containers using:\n# sudo docker exec $CONTAINER bash -c 'cd e-mission-server; source setup\/activate.sh; .\/e-mission-py.bash bin\/debug\/find_invalid_pipeline_state.py'\n\ndef print_all_invalid_state():\n all_invalid_states = edb.get_pipeline_state_db().find({\"curr_run_ts\": {\"$ne\": None}})\n for invalid_state in all_invalid_states:\n print(f\"{invalid_state.user_id}: {ecwp.PipelineStage(invalid_state.pipeline_stage)} set to {arrow.get(invalid_state.curr_run_ts)}\")\n\nif __name__ == '__main__':\n logging.basicConfig(level=logging.DEBUG)\n parser = argparse.ArgumentParser(prog=\"find_invalid_pipeline_state\")\n args = parser.parse_args()\n print_all_invalid_state()\n","new_contents":"import arrow\nimport logging\nimport argparse\nimport emission.core.wrapper.pipelinestate as ecwp\nimport emission.core.get_database as edb\n\n# Run in containers using:\n# sudo docker exec $CONTAINER bash -c 'cd e-mission-server; source setup\/activate.sh; .\/e-mission-py.bash bin\/debug\/find_invalid_pipeline_state.py'\n\ndef print_all_invalid_state():\n all_invalid_states = [ecwp.PipelineState(p) for p in edb.get_pipeline_state_db().find({\"curr_run_ts\": {\"$ne\": None}})]\n for invalid_state in all_invalid_states:\n print(f\"{invalid_state.user_id}: {ecwp.PipelineStages(invalid_state.pipeline_stage)} set to {arrow.get(invalid_state.curr_run_ts)}\")\n\nif __name__ == '__main__':\n logging.basicConfig(level=logging.DEBUG)\n parser = argparse.ArgumentParser(prog=\"find_invalid_pipeline_state\")\n args = parser.parse_args()\n print_all_invalid_state()\n","subject":"Convert entries to PipelineState objects","message":"Convert entries to PipelineState objects\n\n+ typo in wrapper class name\n","lang":"Python","license":"bsd-3-clause","repos":"shankari\/e-mission-server,shankari\/e-mission-server,e-mission\/e-mission-server,e-mission\/e-mission-server,e-mission\/e-mission-server,shankari\/e-mission-server,e-mission\/e-mission-server,shankari\/e-mission-server"} {"commit":"a642934b9e284009ac92057d2163515b2b581acf","old_file":"conanfile.py","new_file":"conanfile.py","old_contents":"from conans import ConanFile\n\nclass GslLiteConan(ConanFile):\n version = \"0.34.0\"\n name = \"gsl-lite\"\n description = \"A single-file header-only version of ISO C++ Guidelines Support Library (GSL) for C++98, C++11 and later\"\n license = \"MIT License. https:\/\/opensource.org\/licenses\/MIT\"\n url = \"https:\/\/github.com\/martinmoene\/gsl-lite.git\"\n exports_sources = \"include\/gsl\/*\", \"include\/*\", \"LICENSE\"\n build_policy = \"missing\" \n author = \"Martin Moene\"\n\n def build(self):\n \"\"\"Avoid warning on build step\"\"\"\n pass\n\n def package(self):\n \"\"\"Provide pkg\/include\/gsl\/*.hpp\"\"\"\n self.copy(\"*.hpp\")\n self.copy(\"include\/gsl\/gsl\")\n\n def package_info(self):\n self.info.header_only()\n","new_contents":"from conans import ConanFile, CMake\n\nclass GslLiteConan(ConanFile):\n version = \"0.34.0\"\n name = \"gsl-lite\"\n description = \"A single-file header-only version of ISO C++ Guidelines Support Library (GSL) for C++98, C++11 and later\"\n license = \"MIT License. https:\/\/opensource.org\/licenses\/MIT\"\n url = \"https:\/\/github.com\/martinmoene\/gsl-lite.git\"\n exports_sources = \"include\/gsl\/*\", \"CMakeLists.txt\", \"cmake\/*\", \"LICENSE\"\n build_policy = \"missing\"\n author = \"Martin Moene\"\n\n def build(self):\n \"\"\"Avoid warning on build step\"\"\"\n pass\n\n def package(self):\n \"\"\"Run CMake install\"\"\"\n cmake = CMake(self)\n cmake.definitions[\"GSL_LITE_OPT_BUILD_TESTS\"] = \"OFF\"\n cmake.definitions[\"GSL_LITE_OPT_BUILD_EXAMPLES\"] = \"OFF\"\n cmake.configure()\n cmake.install()\n\n def package_info(self):\n self.info.header_only()\n","subject":"Improve conan\/cmake interoperation (nonstd-lite issue 32)","message":"Improve conan\/cmake interoperation (nonstd-lite issue 32)\n\n- https:\/\/github.com\/martinmoene\/nonstd-lite-project\/issues\/32\n","lang":"Python","license":"mit","repos":"martinmoene\/gsl-lite,martinmoene\/gsl-lite,martinmoene\/gsl-lite"} {"commit":"c78bb93f6b9a1347458bb9acaf26954b77bc4aab","old_file":"conanfile.py","new_file":"conanfile.py","old_contents":"from conans import ConanFile\nfrom conans.tools import download, unzip\nimport os\n\nVERSION = \"0.0.3\"\n\n\nclass CMakeHeaderLanguageConan(ConanFile):\n name = \"cmake-header-language\"\n version = os.environ.get(\"CONAN_VERSION_OVERRIDE\", VERSION)\n generators = \"cmake\"\n requires = (\"cmake-include-guard\/master@smspillaz\/cmake-include-guard\",)\n url = \"http:\/\/github.com\/polysquare\/cmake-header-language\"\n license = \"MIT\"\n options = {\n \"dev\": [True, False]\n }\n default_options = \"dev=False\"\n\n def requirements(self):\n if self.options.dev:\n self.requires(\"cmake-module-common\/master@smspillaz\/cmake-module-common\")\n\n def source(self):\n zip_name = \"cmake-header-language.zip\"\n download(\"https:\/\/github.com\/polysquare\/\"\n \"cmake-header-language\/archive\/{version}.zip\"\n \"\".format(version=\"v\" + VERSION),\n zip_name)\n unzip(zip_name)\n os.unlink(zip_name)\n\n def package(self):\n self.copy(pattern=\"*.cmake\",\n dst=\"cmake\/cmake-header-language\",\n src=\"cmake-header-language-\" + VERSION,\n keep_path=True)\n","new_contents":"from conans import ConanFile\nfrom conans.tools import download, unzip\nimport os\n\nVERSION = \"0.0.4\"\n\n\nclass CMakeHeaderLanguageConan(ConanFile):\n name = \"cmake-header-language\"\n version = os.environ.get(\"CONAN_VERSION_OVERRIDE\", VERSION)\n generators = \"cmake\"\n requires = (\"cmake-include-guard\/master@smspillaz\/cmake-include-guard\",)\n url = \"http:\/\/github.com\/polysquare\/cmake-header-language\"\n license = \"MIT\"\n options = {\n \"dev\": [True, False]\n }\n default_options = \"dev=False\"\n\n def requirements(self):\n if self.options.dev:\n self.requires(\"cmake-module-common\/master@smspillaz\/cmake-module-common\")\n\n def source(self):\n zip_name = \"cmake-header-language.zip\"\n download(\"https:\/\/github.com\/polysquare\/\"\n \"cmake-header-language\/archive\/{version}.zip\"\n \"\".format(version=\"v\" + VERSION),\n zip_name)\n unzip(zip_name)\n os.unlink(zip_name)\n\n def package(self):\n self.copy(pattern=\"*.cmake\",\n dst=\"cmake\/cmake-header-language\",\n src=\"cmake-header-language-\" + VERSION,\n keep_path=True)\n","subject":"Bump version: 0.0.3 -> 0.0.4","message":"Bump version: 0.0.3 -> 0.0.4\n\n[ci skip]\n","lang":"Python","license":"mit","repos":"polysquare\/cmake-header-language"} {"commit":"a689ce5e38c4d4a5ce982037c5c15396f361c7d6","old_file":"contrib\/examples\/sensors\/echo_flask_app.py","new_file":"contrib\/examples\/sensors\/echo_flask_app.py","old_contents":"from flask import request, Flask\n\nfrom st2reactor.sensor.base import Sensor\n\n\nclass EchoFlaskSensor(Sensor):\n def __init__(self, sensor_service, config):\n super(EchoFlaskSensor, self).__init__(\n sensor_service=sensor_service,\n config=config\n )\n\n self._host = '127.0.0.1'\n self._port = 5000\n self._path = '\/echo'\n\n self._log = self._sensor_service.get_logger(__name__)\n self._app = Flask(__name__)\n\n def setup(self):\n pass\n\n def run(self):\n @self._app.route(self._path, methods=['POST'])\n def echo():\n payload = request.get_json(force=True)\n self._sensor_service.dispatch(trigger=\"examples.echo_flask\",\n payload=payload)\n return request.data\n\n self._log.info('Listening for payload on http:\/\/{}:{}{}'.format(\n self._host, self._port, self._path))\n self._app.run(host=self._host, port=self._port, threaded=True)\n\n def cleanup(self):\n pass\n\n def add_trigger(self, trigger):\n # This method is called when trigger is created\n pass\n\n def update_trigger(self, trigger):\n # This method is called when trigger is updated\n pass\n\n def remove_trigger(self, trigger):\n # This method is called when trigger is deleted\n pass\n","new_contents":"from flask import request, Flask\n\nfrom st2reactor.sensor.base import Sensor\n\n\nclass EchoFlaskSensor(Sensor):\n def __init__(self, sensor_service, config):\n super(EchoFlaskSensor, self).__init__(\n sensor_service=sensor_service,\n config=config\n )\n\n self._host = '127.0.0.1'\n self._port = 5000\n self._path = '\/echo'\n\n self._log = self._sensor_service.get_logger(__name__)\n self._app = Flask(__name__)\n\n def setup(self):\n pass\n\n def run(self):\n @self._app.route(self._path, methods=['POST'])\n def echo():\n payload = request.get_json(force=True)\n self._sensor_service.dispatch(trigger=\"examples.echoflasksensor\",\n payload=payload)\n return request.data\n\n self._log.info('Listening for payload on http:\/\/{}:{}{}'.format(\n self._host, self._port, self._path))\n self._app.run(host=self._host, port=self._port, threaded=True)\n\n def cleanup(self):\n pass\n\n def add_trigger(self, trigger):\n # This method is called when trigger is created\n pass\n\n def update_trigger(self, trigger):\n # This method is called when trigger is updated\n pass\n\n def remove_trigger(self, trigger):\n # This method is called when trigger is deleted\n pass\n","subject":"Make trigger name in .py match .yaml","message":"Make trigger name in .py match .yaml","lang":"Python","license":"apache-2.0","repos":"StackStorm\/st2,nzlosh\/st2,StackStorm\/st2,StackStorm\/st2,nzlosh\/st2,Plexxi\/st2,Plexxi\/st2,nzlosh\/st2,Plexxi\/st2,Plexxi\/st2,nzlosh\/st2,StackStorm\/st2"} {"commit":"6e8e7a067419166afd632aa63ecb743dd6c3a162","old_file":"geokey_dataimports\/tests\/test_model_helpers.py","new_file":"geokey_dataimports\/tests\/test_model_helpers.py","old_contents":"# coding=utf-8\nfrom io import BytesIO\nfrom django.test import TestCase\n\n\nfrom geokey_dataimports.helpers.model_helpers import import_from_csv\n\n\nclass ImportFromCSVTest(TestCase):\n \"\"\"Tests to check that characters can be imported from CSV files.\n\n Notes that these tests are probably not possible or relevant under Python 3.\n \"\"\"\n def test_import_csv_basic_chars(self):\n \"\"\"Basic ASCII characters can be imported.\"\"\"\n mock_csv = BytesIO(\"abc,cde,efg\\n123,456,789\")\n features = []\n import_from_csv(features=features, fields=[], file=mock_csv)\n print(features)\n self.assertEquals(features[0]['properties'], {'cde': '456', 'abc': '123', 'efg': '789'})\n\n def test_import_csv_non_ascii_chars(self):\n \"\"\"Non-ASCII unicode characters can be imported.\"\"\"\n mock_csv = BytesIO(\"abc,àde,e£g\\n¡23,45Ç,Æ8é\")\n features = []\n import_from_csv(features=features, fields=[], file=mock_csv)\n print(features)\n self.assertEquals(features[0]['properties'], {'àde': '45Ç', 'abc': '¡23', 'e£g': 'Æ8é'})\n","new_contents":"# coding=utf-8\nfrom cStringIO import StringIO\nfrom django.test import TestCase\n\n\nfrom geokey_dataimports.helpers.model_helpers import import_from_csv\n\n\nclass ImportFromCSVTest(TestCase):\n \"\"\"Tests to check that characters can be imported from CSV files.\n\n Notes that these tests are probably not possible or relevant under Python 3.\n \"\"\"\n def test_import_csv_basic_chars(self):\n \"\"\"Basic ASCII characters can be imported.\"\"\"\n input_dict = {u'abc': u'123', u'cde': u'456', u'efg': u'789'}\n mock_csv = StringIO(\"abc,cde,efg\\n123,456,789\")\n features = []\n import_from_csv(features=features, fields=[], file_obj=mock_csv)\n for k, v in input_dict.items():\n self.assertEquals(v, features[0]['properties'][k])\n\n def test_import_csv_non_ascii_chars(self):\n \"\"\"Non-ASCII unicode characters can be imported.\"\"\"\n input_dict = {u'à': u'¡', u'£': u'Ç'}\n mock_csv = StringIO(\"à,£\\n¡,Ç\")\n features = []\n import_from_csv(features=features, fields=[], file_obj=mock_csv)\n for k, v in input_dict.items():\n self.assertEquals(v, features[0]['properties'][k])\n\n","subject":"Simplify test data for easier comparison.","message":"Simplify test data for easier comparison.\n","lang":"Python","license":"mit","repos":"ExCiteS\/geokey-dataimports,ExCiteS\/geokey-dataimports,ExCiteS\/geokey-dataimports"} {"commit":"b8e23194ff0c24bd9460629aff18f69d7a868f6d","old_file":"likelihood.py","new_file":"likelihood.py","old_contents":"import math\n#Log-likelihood\ndef ll(ciphertext,perm,mat,k):\n s=0.0\n for i in range(len(ciphertext)-(k-1)):\n kmer = tuple([perm[c] for c in ciphertext[i:i+k]])\n s = s + math.log(mat[kmer])\n return s\n","new_contents":"import math\n#Log-likelihood\ndef ll(ciphertext,perm,mat,k):\n if k==1:\n return ll_k1(ciphertext,perm,mat)\n if k==2:\n return ll_k2(ciphertext,perm,mat)\n if k==3:\n return ll_k3(ciphertext,perm,mat)\n s=0.0\n for i in range(len(ciphertext)-(k-1)):\n kmer = tuple([perm[c] for c in ciphertext[i:i+k]])\n s = s + math.log(mat[kmer])\n return s\n\n##Log-likelihood - hard-coded version for k=1\ndef ll_k1(ciphertext,perm,mat):\n s=0.0\n for i in range(len(ciphertext)):\n uple = (perm[ciphertext[i]],)\n s = s + math.log(mat[uple])\n return s\n\n##Log-likelihood - hard-coded version for k=2\ndef ll_k2(ciphertext,perm,mat):\n s=0.0\n for i in range(len(ciphertext)-1):\n pair = (perm[ciphertext[i]],perm[ciphertext[i+1]])\n s = s + math.log(mat[pair])\n return s\n\n##Log-likelihood - hard-coded version for k=3\ndef ll_k3(ciphertext,perm,mat):\n s=0.0\n for i in range(len(ciphertext)-2):\n triplet = (perm[ciphertext[i]],perm[ciphertext[i+1]],perm[ciphertext[i+2]])\n s = s + math.log(mat[triplet])\n return s\n","subject":"Add hard-coded versions of ll function for k=1,2,3 for speed","message":"Add hard-coded versions of ll function for k=1,2,3 for speed\n","lang":"Python","license":"mit","repos":"gputzel\/decode"} {"commit":"50596484c1214a73d3722af116ccea7fa258fb11","old_file":"direnaj\/direnaj_api\/celery_app\/server_endpoint.py","new_file":"direnaj\/direnaj_api\/celery_app\/server_endpoint.py","old_contents":"__author__ = 'onur'\n\nfrom celery import Celery\n\nimport direnaj_api.config.server_celeryconfig as celeryconfig\n\napp_object = Celery()\n\napp_object.config_from_object(celeryconfig)\n\n@app_object.task\ndef deneme(x, seconds):\n print \"Sleeping for printing %s for %s seconds..\" % (x, seconds)\n import time\n time.sleep(seconds)\n print x\n\n#from celery.schedules import crontab\n#from celery.task import periodic_task\nfrom direnaj_api.utils.direnajmongomanager import create_batch_from_watchlist\n\n#@periodic_task(run_every=crontab(minute='*\/1'))\n@app_object.task\ndef check_watchlist_and_dispatch_tasks():\n batch_size = 10\n res_array = create_batch_from_watchlist(app_object, batch_size)\n\n\nif __name__ == \"__main__\":\n app_object.start()","new_contents":"__author__ = 'onur'\n\nfrom celery import Celery\n\nimport direnaj_api.config.server_celeryconfig as celeryconfig\n\napp_object = Celery()\n\napp_object.config_from_object(celeryconfig)\n\n@app_object.task\ndef deneme(x, seconds):\n print \"Sleeping for printing %s for %s seconds..\" % (x, seconds)\n import time\n time.sleep(seconds)\n print x\n\n#from celery.schedules import crontab\n#from celery.task import periodic_task\nfrom direnaj_api.utils.direnajmongomanager import create_batch_from_watchlist\n\n#@periodic_task(run_every=crontab(minute='*\/1'))\n@app_object.task(name='check_watchlist_and_dispatch_tasks')\ndef check_watchlist_and_dispatch_tasks():\n batch_size = 10\n res_array = create_batch_from_watchlist(app_object, batch_size)\n\n\nif __name__ == \"__main__\":\n app_object.start()","subject":"Fix for periodic task scheduler (4)","message":"Fix for periodic task scheduler (4)\n","lang":"Python","license":"mit","repos":"boun-cmpe-soslab\/drenaj,boun-cmpe-soslab\/drenaj,boun-cmpe-soslab\/drenaj,boun-cmpe-soslab\/drenaj"} {"commit":"6d2118a87dfb811015727970b1eda74c15769e06","old_file":"distutilazy\/__init__.py","new_file":"distutilazy\/__init__.py","old_contents":"\"\"\"\ndistutilazy\n-----------\n\nExtra distutils command classes.\n\n:license: MIT, see LICENSE for more details.\n\"\"\"\n\nfrom os.path import dirname, abspath\nimport sys\n\n__version__ = '0.4.0'\n__all__ = ['clean', 'pyinstaller', 'command']\n\nbase_dir = abspath(dirname(dirname(__file__)))\nif base_dir not in sys.path:\n if len(sys.path):\n sys.path.insert(1, base_dir)\n else:\n sys.path.append(base_dir)\n","new_contents":"\"\"\"\ndistutilazy\n-----------\n\nExtra distutils command classes.\n\n:license: MIT, see LICENSE for more details.\n\"\"\"\n\nfrom os.path import dirname, abspath\nimport sys\n\n__version__ = \"0.4.0\"\n__all__ = (\"clean\", \"pyinstaller\", \"command\")\n\nbase_dir = abspath(dirname(dirname(__file__)))\nif base_dir not in sys.path:\n if len(sys.path):\n sys.path.insert(1, base_dir)\n else:\n sys.path.append(base_dir)\n","subject":"Replace ' with \" to keep it consistant with other modules","message":"Replace ' with \" to keep it consistant with other modules\n","lang":"Python","license":"mit","repos":"farzadghanei\/distutilazy"} {"commit":"2b71acfdff19fad51bdabf9284ccadb63960feff","old_file":"src\/ggrc\/settings\/default.py","new_file":"src\/ggrc\/settings\/default.py","old_contents":"# Copyright (C) 2013 Google Inc., authors, and contributors <see AUTHORS file>\n# Licensed under http:\/\/www.apache.org\/licenses\/LICENSE-2.0 <see LICENSE file>\n# Created By: dan@reciprocitylabs.com\n# Maintained By: dan@reciprocitylabs.com\n\nDEBUG = False\nTESTING = False\n\n# Flask-SQLAlchemy fix to be less than `wait_time` in \/etc\/mysql\/my.cnf\nSQLALCHEMY_POOL_RECYCLE = 120\n\n# Settings in app.py\nAUTOBUILD_ASSETS = False\nENABLE_JASMINE = False\nDEBUG_ASSETS = False\nFULLTEXT_INDEXER = None\nUSER_PERMISSIONS_PROVIDER = None\nEXTENSIONS = []\n\n# Deployment-specific variables\nCOMPANY = \"Company, Inc.\"\nCOMPANY_LOGO_TEXT = \"Company GRC\"\nVERSION = \"s14\"\n\n# Initialize from environment if present\nimport os\nSQLALCHEMY_DATABASE_URI = os.environ.get('GGRC_DATABASE_URI', '')\nSECRET_KEY = os.environ.get('GGRC_SECRET_KEY', 'Replace-with-something-secret')\n","new_contents":"# Copyright (C) 2013 Google Inc., authors, and contributors <see AUTHORS file>\n# Licensed under http:\/\/www.apache.org\/licenses\/LICENSE-2.0 <see LICENSE file>\n# Created By: dan@reciprocitylabs.com\n# Maintained By: dan@reciprocitylabs.com\n\nDEBUG = False\nTESTING = False\n\n# Flask-SQLAlchemy fix to be less than `wait_time` in \/etc\/mysql\/my.cnf\nSQLALCHEMY_POOL_RECYCLE = 120\n\n# Settings in app.py\nAUTOBUILD_ASSETS = False\nENABLE_JASMINE = False\nDEBUG_ASSETS = False\nFULLTEXT_INDEXER = None\nUSER_PERMISSIONS_PROVIDER = None\nEXTENSIONS = []\n\n# Deployment-specific variables\nCOMPANY = \"Company, Inc.\"\nCOMPANY_LOGO_TEXT = \"Company GRC\"\nVERSION = \"s14.1\"\n\n# Initialize from environment if present\nimport os\nSQLALCHEMY_DATABASE_URI = os.environ.get('GGRC_DATABASE_URI', '')\nSECRET_KEY = os.environ.get('GGRC_SECRET_KEY', 'Replace-with-something-secret')\n","subject":"Bump version for 14.1 patch","message":"Bump version for 14.1 patch\n","lang":"Python","license":"apache-2.0","repos":"prasannav7\/ggrc-core,vladan-m\/ggrc-core,VinnieJohns\/ggrc-core,AleksNeStu\/ggrc-core,jmakov\/ggrc-core,hasanalom\/ggrc-core,edofic\/ggrc-core,prasannav7\/ggrc-core,hyperNURb\/ggrc-core,uskudnik\/ggrc-core,prasannav7\/ggrc-core,josthkko\/ggrc-core,j0gurt\/ggrc-core,vladan-m\/ggrc-core,josthkko\/ggrc-core,vladan-m\/ggrc-core,j0gurt\/ggrc-core,uskudnik\/ggrc-core,j0gurt\/ggrc-core,josthkko\/ggrc-core,NejcZupec\/ggrc-core,selahssea\/ggrc-core,plamut\/ggrc-core,jmakov\/ggrc-core,hyperNURb\/ggrc-core,kr41\/ggrc-core,jmakov\/ggrc-core,andrei-karalionak\/ggrc-core,hyperNURb\/ggrc-core,jmakov\/ggrc-core,kr41\/ggrc-core,uskudnik\/ggrc-core,AleksNeStu\/ggrc-core,plamut\/ggrc-core,edofic\/ggrc-core,plamut\/ggrc-core,andrei-karalionak\/ggrc-core,hyperNURb\/ggrc-core,VinnieJohns\/ggrc-core,AleksNeStu\/ggrc-core,josthkko\/ggrc-core,edofic\/ggrc-core,edofic\/ggrc-core,andrei-karalionak\/ggrc-core,AleksNeStu\/ggrc-core,uskudnik\/ggrc-core,VinnieJohns\/ggrc-core,kr41\/ggrc-core,j0gurt\/ggrc-core,uskudnik\/ggrc-core,VinnieJohns\/ggrc-core,hasanalom\/ggrc-core,NejcZupec\/ggrc-core,plamut\/ggrc-core,hasanalom\/ggrc-core,hasanalom\/ggrc-core,vladan-m\/ggrc-core,selahssea\/ggrc-core,selahssea\/ggrc-core,vladan-m\/ggrc-core,jmakov\/ggrc-core,selahssea\/ggrc-core,NejcZupec\/ggrc-core,kr41\/ggrc-core,hasanalom\/ggrc-core,andrei-karalionak\/ggrc-core,NejcZupec\/ggrc-core,hyperNURb\/ggrc-core,prasannav7\/ggrc-core"} {"commit":"8ffe217fe512296d41ae474c9d145ee2de599eac","old_file":"src\/constants.py","new_file":"src\/constants.py","old_contents":"class AppUrl:\n\t\"\"\"Class for storing all the URLs used in the application\"\"\"\n\tBASE = \"http:\/\/www.thehindu.com\/\"\n\tOP_BASE = BASE + \"opinion\/\"\n\tOPINION = OP_BASE + \"?service=rss\"\n\tEDITORIAL = OP_BASE + \"editorial\/?service=rss\"\n\tSAMPLE = BASE +\t\"op-ed\/a-super-visit-in-the-season-of-hope\/article7214799.ece\"\n\tRSS_ARGS = \"?utm_source=RSS_Feed&utm_medium=RSS&utm_campaign=RSS_Syndication\"\n\nclass Kind:\n\t#BLOGS\t\t=\t'blogs'\n\t#CARTOON\t=\t'cartoon'\n\tCOLUMNS\t\t=\t'columns'\n\tEDITORIAL\t=\t'editorial'\n\tINTERVIEW\t=\t'interview'\n\tLEAD\t\t=\t'lead'\n\t#LETTERS\t=\t'letters'\n\tOP_ED\t\t=\t'op-ed'\n\tOPEN_PAGE\t=\t'open-page'\n\t#READERS_ED\t=\t'Readers-Editor'\n\t#SUNDAY_ANCHOR\t=\t'sunday-anchor'\n\nclass Tags:\n\taccepted = ['a', 'b', 'i', 'p']\n","new_contents":"class AppUrl:\n\t\"\"\"Class for storing all the URLs used in the application\"\"\"\n\tBASE = \"http:\/\/www.thehindu.com\/\"\n\tOP_BASE = BASE + \"opinion\/\"\n\tOPINION = OP_BASE + \"?service=rss\"\n\tEDITORIAL = OP_BASE + \"editorial\/?service=rss\"\n\tSAMPLE = BASE +\t\"op-ed\/a-super-visit-in-the-season-of-hope\/article7214799.ece\"\n\tRSS_ARGS = \"?utm_source=RSS_Feed&utm_medium=RSS&utm_campaign=RSS_Syndication\"\n\nclass Kind:\n\tBLOGS\t\t=\t'blogs'\n\tCARTOON\t\t=\t'cartoon'\n\tCOLUMNS\t\t=\t'columns'\n\tEDITORIAL\t=\t'editorial'\n\tINTERVIEW\t=\t'interview'\n\tLEAD\t\t=\t'lead'\n\tLETTERS\t\t=\t'letters'\n\tOP_ED\t\t=\t'op-ed'\n\tOPEN_PAGE\t=\t'open-page'\n\tREADERS_ED\t=\t'Readers-Editor'\n\tSUNDAY_ANCHOR\t=\t'sunday-anchor'\n\tSUPPORTED\t=\t[COLUMNS, EDITORIAL, INTERVIEW, LEAD, OP_ED,\n\t\t\tOPEN_PAGE]\n\nclass Tags:\n\taccepted = ['a', 'b', 'i', 'p']\n","subject":"Include all but have a supported set","message":"Include all but have a supported set\n\nRather than commenting out the values, this seems more sensible. And pretty.. of\ncourse.\n\nSigned-off-by: Venkatesh Shukla <8349e50bec2939976da648e286d7e261bcd17fa3@iitbhu.ac.in>\n","lang":"Python","license":"mit","repos":"venkateshshukla\/th-editorials-server"} {"commit":"1fabf53957c4951e36f756c94bea4007cd6a5d6e","old_file":"run_server.py","new_file":"run_server.py","old_contents":"#!\/usr\/bin\/env python3\nimport subprocess\nimport sys\n\n\ndef main():\n ip = '127.0.0.1'\n port = 5000\n workers_count = 4\n if len(sys.argv) > 1:\n for arg in sys.argv[1:]:\n if ':' in arg:\n ip, port = arg.split(':')\n port = int(port)\n break\n if '.' in arg:\n ip = arg\n if arg.isdigit():\n port = int(arg)\n\n print('FluCalc started on {ip}:{port}'.format(ip=ip, port=port))\n subprocess.run('gunicorn -w {workers_count} -b {ip}:{port} flucalc.server:app'.format(\n workers_count=workers_count, ip=ip, port=port\n ), shell=True)\n\n\nif __name__ == '__main__':\n main()\n\n","new_contents":"#!\/usr\/bin\/env python3\nimport subprocess\nimport sys\n\n\ndef main():\n ip = '127.0.0.1'\n port = 5000\n workers_count = 4\n if len(sys.argv) > 1:\n for arg in sys.argv[1:]:\n if ':' in arg:\n ip, port = arg.split(':')\n port = int(port)\n break\n if '.' in arg:\n ip = arg\n if arg.isdigit():\n port = int(arg)\n\n print('FluCalc started on http:\/\/{ip}:{port}'.format(ip=ip, port=port))\n subprocess.run('gunicorn -w {workers_count} -b {ip}:{port} flucalc.server:app'.format(\n workers_count=workers_count, ip=ip, port=port\n ), shell=True)\n\n\nif __name__ == '__main__':\n main()\n\n","subject":"Improve print message for the server address","message":"Improve print message for the server address\n","lang":"Python","license":"mit","repos":"bondarevts\/flucalc,bondarevts\/flucalc,bondarevts\/flucalc"} {"commit":"af4bdd9339e3905f6489577afe7fac33475e3884","old_file":"src\/services\/listener_bot.py","new_file":"src\/services\/listener_bot.py","old_contents":"from logging import info, basicConfig, INFO\nfrom time import sleep\nfrom src.slack.slack_bot import Bot\n\n\ndef main():\n basicConfig(level=INFO)\n bot = Bot()\n\n info('Connection Slack')\n if not bot.slack_client.rtm_connect():\n info('Could not connect in web_socket')\n exit()\n else:\n info('Connected')\n\n while True:\n sleep(0.5)\n for message in bot.get_direct_messages():\n info(message)\n bot.send_message_in_channel(message.message, message.channel)\n\n\nif __name__ == '__main__':\n main()\n","new_contents":"from logging import info, basicConfig, INFO\nfrom time import sleep\nfrom src.slack.slack_bot import Bot\n\n\ndef main():\n basicConfig(level=INFO)\n bot = Bot()\n\n info('Connection Slack')\n if not bot.slack_client.rtm_connect():\n info('Could not connect in web_socket')\n exit()\n else:\n info('Connected')\n\n while True:\n sleep(1)\n for message in bot.get_direct_messages():\n info(message)\n bot.send_message_in_channel(message.message, message.channel)\n\n\nif __name__ == '__main__':\n main()\n","subject":"Improve waiting time to bot listener","message":"Improve waiting time to bot listener\n","lang":"Python","license":"bsd-3-clause","repos":"otherpirate\/dbaas-slack-bot"} {"commit":"f1506a36ab4d9970fb7661aaf2a0f7da05812cec","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name='mnp',\n version='1.0.0',\n author='Heryandi',\n author_email='heryandi@gmail.com',\n packages=find_packages(exclude='test'),\n scripts=[],\n url='https:\/\/github.com\/heryandi\/mnp',\n license='MIT',\n description='Wrapper tools to manage Mininet package',\n long_description=open('README.rst').read(),\n install_requires=[\n \"pip\",\n \"requests\",\n \"setuptools\",\n ],\n entry_points={'console_scripts': [\n 'mnp = mnp:main',\n ]}\n)\n","new_contents":"import os\n\nfrom setuptools import setup, find_packages\n\nsetup(\n name=\"mnp\",\n version=\"1.0.0\",\n author=\"Heryandi\",\n author_email=\"heryandi@gmail.com\",\n packages=find_packages(exclude=\"test\"),\n scripts=[],\n url=\"https:\/\/github.com\/heryandi\/mnp\",\n license=\"MIT\",\n description=\"Tools to manage Mininet package\",\n long_description=open(\"README.rst\").read(),\n install_requires=[\n \"pip\",\n \"requests\",\n \"setuptools\",\n ],\n entry_points={\"console_scripts\": [\n \"mnp = mnp:main\",\n ]},\n classifiers=[\n \"Mininet :: Tool\",\n ],\n keywords=\"command-line commandline mininet package packaging tool\"\n)\n","subject":"Add classifier and keyword to metadata","message":"Add classifier and keyword to metadata\n","lang":"Python","license":"mit","repos":"heryandi\/mnp"} {"commit":"b22c7b78586ee07d686d06ccfd6213ca35afef1b","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\n\n\nsetup(\n name='slacker',\n version='0.9.29',\n packages=['slacker'],\n description='Slack API client',\n author='Oktay Sancak',\n author_email='oktaysancak@gmail.com',\n url='http:\/\/github.com\/os\/slacker\/',\n install_requires=['requests >= 2.2.1'],\n license='http:\/\/www.apache.org\/licenses\/LICENSE-2.0',\n test_suite='tests',\n classifiers=(\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache Software License',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4'\n ),\n keywords='slack api'\n)\n","new_contents":"from setuptools import setup\n\n\nsetup(\n name='slacker',\n version='0.9.30',\n packages=['slacker'],\n description='Slack API client',\n author='Oktay Sancak',\n author_email='oktaysancak@gmail.com',\n url='http:\/\/github.com\/os\/slacker\/',\n install_requires=['requests >= 2.2.1'],\n license='http:\/\/www.apache.org\/licenses\/LICENSE-2.0',\n test_suite='tests',\n classifiers=(\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: Apache Software License',\n 'Natural Language :: English',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4'\n ),\n keywords='slack api'\n)\n","subject":"Set version number to 0.9.30","message":"Set version number to 0.9.30\n","lang":"Python","license":"apache-2.0","repos":"os\/slacker"} {"commit":"e7e519f1d1636240022b7961fe949bc41a97c091","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport shutil\nimport sys\n\nVERSION = '0.0.0git'\n\nsetup_info = dict(\n\n name = 'nimp',\n version = VERSION,\n author = 'Dontnod Entertainment',\n description = 'DNE build tool',\n\n packages = [\n 'nimp',\n 'nimp\/commands',\n 'nimp',\n ],\n\n install_requires = [\n 'glob2',\n ],\n\n entry_points = {\n 'console_scripts' : [ 'nimp = nimp.nimp_cli:main' ],\n },\n)\n\nsetuptools_info = dict(\n zip_safe = True,\n)\n\nfrom setuptools import setup\nsetup(**setup_info)\n\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport shutil\nimport sys\n\nVERSION = '0.0.1'\n\nsetup_info = dict(\n\n name = 'nimp-cli',\n version = VERSION,\n author = 'Dontnod Entertainment',\n description = 'Multipurpose build tool',\n\n packages = [\n 'nimp',\n 'nimp\/commands',\n 'nimp',\n ],\n\n install_requires = [\n 'glob2',\n ],\n\n entry_points = {\n 'console_scripts' : [ 'nimp = nimp.nimp_cli:main' ],\n },\n)\n\nsetuptools_info = dict(\n zip_safe = True,\n)\n\nfrom setuptools import setup\nsetup(**setup_info)\n\n","subject":"Call the package nimp-cli because nimp is already taken.","message":"Call the package nimp-cli because nimp is already taken.\n","lang":"Python","license":"mit","repos":"dontnod\/nimp"} {"commit":"03cb3e001a25467319d0d82a5fc95e1c07ea3dd4","old_file":"setup.py","new_file":"setup.py","old_contents":"from distutils.core import setup\nimport multi_schema\n\nsetup(\n name = \"django-multi-schema\",\n version = multi_schema.__version__,\n description = \"Postgres schema support in django.\",\n url = \"http:\/\/hg.schinckel.net\/django-multi-schema\",\n author = \"Matthew Schinckel\",\n author_email = \"matt@schinckel.net\",\n packages = [\n \"multi_schema\",\n ],\n classifiers = [\n 'Programming Language :: Python',\n 'Operating System :: OS Independent',\n 'Framework :: Django',\n ],\n)\n","new_contents":"from distutils.core import setup\nimport multi_schema\n\nsetup(\n name = \"django-multi-schema\",\n version = multi_schema.__version__,\n description = \"Postgres schema support in django.\",\n url = \"http:\/\/hg.schinckel.net\/django-multi-schema\",\n author = \"Matthew Schinckel\",\n author_email = \"matt@schinckel.net\",\n packages = [\n \"multi_schema\",\n ],\n install_requires = [\n 'south==0.7.4', \n ],\n classifiers = [\n 'Programming Language :: Python',\n 'Operating System :: OS Independent',\n 'Framework :: Django',\n ],\n)\n","subject":"Add south as a dependency, so we can apply a version. Does not need to be installed in INSTALLED_APPS.","message":"Add south as a dependency, so we can apply a version.\nDoes not need to be installed in INSTALLED_APPS.\n","lang":"Python","license":"bsd-3-clause","repos":"luzfcb\/django-boardinghouse,luzfcb\/django-boardinghouse,luzfcb\/django-boardinghouse"} {"commit":"f97e5585386e7e9417689cceff1bf49386473551","old_file":"setup.py","new_file":"setup.py","old_contents":"from ez_setup import use_setuptools # https:\/\/pypi.python.org\/pypi\/setuptools\nuse_setuptools()\nfrom setuptools import setup, find_packages\nfrom packager import __version__\n\n# Get the long description from the README file.\ndef get_long_description():\n from codecs import open\n from os import path\n\n here = path.abspath(path.dirname(__file__))\n try:\n with open(path.join(here, 'README.md'), encoding='utf-8') as f:\n long_description = f.read()\n except:\n return []\n else:\n return long_description\n\nsetup(\n name='packagebuilder',\n version=__version__,\n description='Tools for building rpm and deb packages for CSDMS software',\n long_description=get_long_description(),\n url='https:\/\/github.com\/csdms\/packagebuilder',\n author='Mark Piper',\n author_email='mark.piper@colorado.edu',\n license='MIT',\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'Topic :: Software Development :: Build Tools',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 2.7',\n ],\n keywords='CSDMS, earth system modeling, packaging, Linux, rpm, deb',\n packages=find_packages(),\n install_requires=['nose'],\n package_data={\n 'packager': ['repositories.txt'],\n },\n entry_points={\n 'console_scripts': [\n 'build_rpm=packager.rpm.build:main',\n ],\n },\n )\n","new_contents":"from ez_setup import use_setuptools # https:\/\/pypi.python.org\/pypi\/setuptools\nuse_setuptools()\nfrom setuptools import setup, find_packages\nfrom packager import __version__\n\n# Get the long description from the README file.\ndef get_long_description():\n from codecs import open\n from os import path\n\n here = path.abspath(path.dirname(__file__))\n try:\n with open(path.join(here, 'README.md'), encoding='utf-8') as f:\n long_description = f.read()\n except:\n return []\n else:\n return long_description\n\nsetup(\n name='packagebuilder',\n version=__version__,\n description='Tools for building rpm and deb packages for CSDMS software',\n long_description=get_long_description(),\n url='https:\/\/github.com\/csdms\/packagebuilder',\n author='Mark Piper',\n author_email='mark.piper@colorado.edu',\n license='MIT',\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Intended Audience :: Developers',\n 'Topic :: Software Development :: Build Tools',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 2.7',\n ],\n keywords='CSDMS, earth system modeling, packaging, Linux, rpm, deb',\n packages=find_packages(exclude=['*test']),\n install_requires=['nose'],\n package_data={\n 'packager': ['repositories.txt'],\n },\n entry_points={\n 'console_scripts': [\n 'build_rpm=packager.rpm.build:main',\n ],\n },\n )\n","subject":"Exclude tests from set of installed packages","message":"Exclude tests from set of installed packages\n","lang":"Python","license":"mit","repos":"csdms\/packagebuilder"} {"commit":"f0579373b2c3261b00228d4c30832f51d713a7c2","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom setuptools import setup\n\nwith open(os.path.join(os.path.dirname(__file__), 'README.rst')) as readme:\n README = readme.read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-oidc-provider',\n version='0.0.4',\n packages=[\n 'oidc_provider', 'oidc_provider\/lib', 'oidc_provider\/lib\/endpoints',\n 'oidc_provider\/lib\/utils', 'oidc_provider\/tests',\n ],\n include_package_data=True,\n license='MIT License',\n description='OpenID Connect Provider implementation for Django.',\n long_description=README,\n url='http:\/\/github.com\/juanifioren\/django-oidc-provider',\n author='Juan Ignacio Fiorentino',\n author_email='juanifioren@gmail.com',\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n install_requires=[\n 'pyjwt==1.1.0',\n ],\n)\n","new_contents":"import os\nfrom setuptools import setup\n\nwith open(os.path.join(os.path.dirname(__file__), 'README.rst')) as readme:\n README = readme.read()\n\n# allow setup.py to be run from any path\nos.chdir(os.path.normpath(os.path.join(os.path.abspath(__file__), os.pardir)))\n\nsetup(\n name='django-oidc-provider',\n version='0.0.4',\n packages=[\n 'oidc_provider', 'oidc_provider\/lib', 'oidc_provider\/lib\/endpoints',\n 'oidc_provider\/lib\/utils', 'oidc_provider\/tests', 'oidc_provider\/migrations',\n ],\n include_package_data=True,\n license='MIT License',\n description='OpenID Connect Provider implementation for Django.',\n long_description=README,\n url='http:\/\/github.com\/juanifioren\/django-oidc-provider',\n author='Juan Ignacio Fiorentino',\n author_email='juanifioren@gmail.com',\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Topic :: Internet :: WWW\/HTTP',\n 'Topic :: Internet :: WWW\/HTTP :: Dynamic Content',\n ],\n install_requires=[\n 'pyjwt==1.1.0',\n ],\n)\n","subject":"Add migrations folder to build.","message":"Add migrations folder to build.\n","lang":"Python","license":"mit","repos":"django-py\/django-openid-provider,nmohoric\/django-oidc-provider,wojtek-fliposports\/django-oidc-provider,wayward710\/django-oidc-provider,nmohoric\/django-oidc-provider,ByteInternet\/django-oidc-provider,bunnyinc\/django-oidc-provider,ByteInternet\/django-oidc-provider,juanifioren\/django-oidc-provider,torreco\/django-oidc-provider,Sjord\/django-oidc-provider,torreco\/django-oidc-provider,django-py\/django-openid-provider,juanifioren\/django-oidc-provider,Sjord\/django-oidc-provider,bunnyinc\/django-oidc-provider,wayward710\/django-oidc-provider,wojtek-fliposports\/django-oidc-provider"} {"commit":"c1955ceeb08f960cef631e6a812862052d42d85c","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nfrom setuptools import find_packages, setup\n\nsetup(\n name='blanc-contentfiles',\n version='0.2.1',\n description='Blanc Content Files',\n long_description=open('README.rst').read(),\n url='https:\/\/github.com\/blancltd\/blanc-contentfiles',\n maintainer='Alex Tomkins',\n maintainer_email='alex@blanc.ltd.uk',\n platforms=['any'],\n packages=find_packages(),\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n ],\n license='BSD',\n)\n","new_contents":"#!\/usr\/bin\/env python\nfrom setuptools import find_packages, setup\n\nsetup(\n name='blanc-contentfiles',\n version='0.2.1',\n description='Blanc Content Files',\n long_description=open('README.rst').read(),\n url='https:\/\/github.com\/blancltd\/blanc-contentfiles',\n maintainer='Blanc Ltd',\n maintainer_email='studio@blanc.ltd.uk',\n platforms=['any'],\n packages=find_packages(),\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n ],\n license='BSD',\n)\n","subject":"Update contact details to the company","message":"Update contact details to the company\n","lang":"Python","license":"bsd-3-clause","repos":"blancltd\/blanc-contentfiles"} {"commit":"e059af57acec9c077ddb348ac6dd84ff58d312fe","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nfrom setuptools import setup, find_packages\n\nsetup(\n name='blanc-basic-pages',\n version='0.2.1',\n description='Blanc Basic Pages for Django',\n long_description=open('README.rst').read(),\n url='https:\/\/github.com\/blancltd\/blanc-basic-pages',\n maintainer='Alex Tomkins',\n maintainer_email='alex@blanc.ltd.uk',\n platforms=['any'],\n install_requires=[\n 'django-mptt>=0.6.0',\n 'django-mptt-admin==0.1.8',\n ],\n packages=find_packages(),\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n ],\n license='BSD',\n)\n","new_contents":"#!\/usr\/bin\/env python\nfrom setuptools import setup, find_packages\n\nsetup(\n name='blanc-basic-pages',\n version='0.2.1',\n description='Blanc Basic Pages for Django',\n long_description=open('README.rst').read(),\n url='https:\/\/github.com\/blancltd\/blanc-basic-pages',\n maintainer='Alex Tomkins',\n maintainer_email='alex@blanc.ltd.uk',\n platforms=['any'],\n install_requires=[\n 'django-mptt>=0.6.1',\n 'django-mptt-admin>=0.1.8',\n ],\n packages=find_packages(),\n classifiers=[\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n ],\n license='BSD',\n)\n","subject":"Fix dependencies for Django 1.7","message":"Fix dependencies for Django 1.7\n\nOlder versions of django-mptt will generate warnings\n","lang":"Python","license":"bsd-3-clause","repos":"blancltd\/blanc-basic-pages"} {"commit":"5865599eee7014a02e07413dea71a79e878ac6b1","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nimport ptwit\n\n\nrequires = ['python-twitter>=1.0']\n\n\ndef readme():\n with open('README.rst') as f:\n return f.read()\n\n\nsetup(name='ptwit',\n version=ptwit.__version__,\n description='A simple twitter command line client',\n long_description=readme(),\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 2.7',\n 'Environment :: Console',\n 'Intended Audience :: End Users\/Desktop',\n 'Topic :: Utilities'],\n url='http:\/\/github.com\/ptpt\/ptwit',\n author=ptwit.__author__,\n author_email='ptpttt+ptwit@gmail.com',\n keywords='twitter, command-line, client',\n license=ptwit.__license__,\n py_modules=['ptwit'],\n install_requires=requires,\n entry_points={\n 'console_scripts': ['ptwit=ptwit:main']},\n zip_safe=False)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nimport ptwit\n\n\nrequires = ['python-twitter>=1.0']\n\n\ndef readme():\n with open('README.rst') as f:\n return f.read()\n\n\nsetup(name='ptwit',\n version=ptwit.__version__,\n description='A simple twitter command line client',\n long_description=readme(),\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 2.7',\n 'Environment :: Console',\n 'Intended Audience :: End Users\/Desktop',\n 'Topic :: Utilities'],\n url='http:\/\/github.com\/ptpt\/ptwit',\n author=ptwit.__author__,\n author_email='ptpttt+ptwit@gmail.com',\n keywords='twitter, command-line, client',\n license=ptwit.__license__,\n py_modules=['ptwit'],\n install_requires=requires,\n entry_points={\n 'console_scripts': ['ptwit=ptwit:cmd']},\n zip_safe=False)\n","subject":"Change the entry point to cmd","message":"Change the entry point to cmd\n","lang":"Python","license":"mit","repos":"ptpt\/ptwit"} {"commit":"1825d4b8250699c664f4fd0150691969f91fba54","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\ntry:\n from setuptools import setup, find_packages\nexcept ImportError:\n from ez_setup import use_setuptools\n use_setuptools()\n from setuptools import setup, find_packages\n\n\nsetup(\n name='django-nested-admin',\n version=\"2.1.3\",\n install_requires=[\n 'six>=1.7.0',\n ],\n description=\"Django admin classes that allow for nested inlines\",\n author='The Atlantic',\n author_email='programmers@theatlantic.com',\n url='https:\/\/github.com\/theatlantic\/django-nested-admin',\n packages=find_packages(),\n classifiers=[\n 'Environment :: Web Environment',\n 'Intended Audience :: Developers',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Framework :: Django',\n ],\n include_package_data=True,\n zip_safe=False,\n long_description=''.join(list(open('README.rst'))[3:]))\n","new_contents":"#!\/usr\/bin\/env python\n\ntry:\n from setuptools import setup, find_packages\nexcept ImportError:\n from ez_setup import use_setuptools\n use_setuptools()\n from setuptools import setup, find_packages\n\n\nsetup(\n name='django-nested-admin',\n version=\"2.1.4\",\n install_requires=[\n 'six>=1.7.0',\n ],\n description=\"Django admin classes that allow for nested inlines\",\n author='The Atlantic',\n author_email='programmers@theatlantic.com',\n url='https:\/\/github.com\/theatlantic\/django-nested-admin',\n packages=find_packages(),\n classifiers=[\n 'Environment :: Web Environment',\n 'Intended Audience :: Developers',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Framework :: Django',\n ],\n include_package_data=True,\n zip_safe=False,\n long_description=''.join(list(open('README.rst'))[3:]))\n","subject":"Bump version to v2.1.4 [ci skip]","message":"Bump version to v2.1.4 [ci skip]\n","lang":"Python","license":"bsd-2-clause","repos":"sbussetti\/django-nested-admin,sbussetti\/django-nested-admin,sbussetti\/django-nested-admin"} {"commit":"669280351b04d61df1de5ff03c4c7a258b37ad32","old_file":"sell\/views.py","new_file":"sell\/views.py","old_contents":"from decimal import Decimal\n\nfrom django.shortcuts import render\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom books.models import BookType, Book\nfrom common.bookchooserwizard import BookChooserWizard\n\n\nclass SellWizard(BookChooserWizard):\n @property\n def page_title(self):\n return _(\"Sell books\")\n\n @property\n def url_namespace(self):\n return \"sell\"\n\n @property\n def session_var_name(self):\n return \"sell_chosen_books\"\n\n @property\n def feature_add_new(self):\n return True\n\n def process_books_summary(self, session, user, book_list):\n for book in book_list:\n amount = book['amount']\n del book['amount']\n\n user.save()\n dbbook = Book(owner=user, accepted=False, sold=False)\n if 'pk' in book:\n dbbook.book_type_id = book['pk']\n else:\n book['price'] = Decimal(book['price'])\n if book['publication_year'] == \"\":\n book['publication_year'] = 1970\n\n book_type = BookType(**book)\n book_type.save()\n dbbook.book_type = book_type\n\n for i in range(0, amount):\n dbbook.pk = None\n dbbook.save()\n\n return True, None\n\n def success(self, request):\n return render(request, 'sell\/success.html')","new_contents":"from decimal import Decimal\nimport re\n\nfrom django.shortcuts import render\nfrom django.utils.translation import ugettext_lazy as _\n\nfrom books.models import BookType, Book\nfrom common.bookchooserwizard import BookChooserWizard\n\n\nclass SellWizard(BookChooserWizard):\n @property\n def page_title(self):\n return _(\"Sell books\")\n\n @property\n def url_namespace(self):\n return \"sell\"\n\n @property\n def session_var_name(self):\n return \"sell_chosen_books\"\n\n @property\n def feature_add_new(self):\n return True\n\n def process_books_summary(self, session, user, book_list):\n for book in book_list:\n amount = book['amount']\n del book['amount']\n\n user.save()\n dbbook = Book(owner=user, accepted=False, sold=False)\n if 'pk' in book:\n dbbook.book_type_id = book['pk']\n else:\n book['isbn'] = re.sub(r'[^\\d.]+', '', book['isbn'])\n book['price'] = Decimal(book['price'])\n if book['publication_year'] == \"\":\n book['publication_year'] = 1970\n\n book_type = BookType(**book)\n book_type.save()\n dbbook.book_type = book_type\n\n for i in range(0, amount):\n dbbook.pk = None\n dbbook.save()\n\n return True, None\n\n def success(self, request):\n return render(request, 'sell\/success.html')","subject":"Delete non-digit characters in ISBN in server side","message":"Delete non-digit characters in ISBN in server side\n","lang":"Python","license":"agpl-3.0","repos":"m4tx\/egielda,m4tx\/egielda,m4tx\/egielda"} {"commit":"aef6566cc6ec2956d7b7f25e9788b58fb84a220c","old_file":"setup.py","new_file":"setup.py","old_contents":"# coding=utf-8\nimport os\nimport sys\n\nfrom setuptools import setup, Command, find_packages\n\n\nsetup(\n name=\"django-celery-transactions\",\n version=\"0.3.1\",\n description=\"Django transaction support for Celery tasks.\",\n long_description=\"See https:\/\/github.com\/fellowshipofone\/django-celery-transactions\",\n author=\"Nicolas Grasset\",\n author_email=\"nicolas.grasset@gmail.com\",\n url=\"https:\/\/github.com\/fellowshipofone\/django-celery-transactions\",\n license=\"Simplified BSD\",\n packages=[\"djcelery_transactions\"],\n install_requires=[\n \"celery>=2.2.7\",\n \"Django>=1.2.4\",\n ],\n classifiers=[\n \"Framework :: Django\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: BSD License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Topic :: Database\",\n ],\n)\n","new_contents":"# coding=utf-8\nimport os\nimport sys\n\nfrom setuptools import setup, Command, find_packages\n\n\nsetup(\n name=\"django-celery-transactions\",\n version=\"0.3.2\",\n description=\"Django transaction support for Celery tasks.\",\n long_description=\"See https:\/\/github.com\/fellowshipofone\/django-celery-transactions\",\n author=\"Nicolas Grasset\",\n author_email=\"nicolas.grasset@gmail.com\",\n url=\"https:\/\/github.com\/fellowshipofone\/django-celery-transactions\",\n license=\"Simplified BSD\",\n packages=[\"djcelery_transactions\"],\n install_requires=[\n \"celery>=2.2.7\",\n \"Django>=1.2.4\",\n ],\n classifiers=[\n \"Framework :: Django\",\n \"Intended Audience :: Developers\",\n \"License :: OSI Approved :: BSD License\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"Topic :: Database\",\n ],\n)\n","subject":"Bump to version 0.3.2 !minor","message":"pkg: Bump to version 0.3.2 !minor\n","lang":"Python","license":"bsd-2-clause","repos":"stored\/django-celery-transactions,fellowshipofone\/django-celery-transactions"} {"commit":"eb5dab3b3231688966254a1797ced7eec67b6e8a","old_file":"setup.py","new_file":"setup.py","old_contents":"import multiprocessing\nfrom setuptools import setup, find_packages\n\nsetup(\n name='sow-generator',\n version='0.1',\n description='Create a scope of work from templates and version controlled documentation.',\n long_description = open('README.rst', 'r').read() + open('CHANGELOG.rst', 'r').read() + open('AUTHORS.rst', 'r').read(),\n author='Hedley Roos',\n author_email='hedley@praekelt.com',\n license='BSD',\n url='http:\/\/github.com\/praekelt\/sow-generator',\n packages = find_packages(),\n install_requires = [\n 'Django<1.7',\n 'South',\n 'celery',\n 'django-celery',\n 'raven',\n 'PyYAML',\n 'requests',\n 'github3.py',\n 'pyandoc',\n 'django-object-tools',\n 'django-adminplus'\n ],\n include_package_data=True,\n tests_require=[\n 'django-setuptest>=0.1.4',\n ],\n test_suite=\"setuptest.setuptest.SetupTestSuite\",\n classifiers=[\n \"Programming Language :: Python\",\n \"License :: OSI Approved :: BSD License\",\n \"Development Status :: 4 - Beta\",\n \"Operating System :: OS Independent\",\n \"Framework :: Django\",\n \"Intended Audience :: Developers\",\n \"Topic :: Internet :: WWW\/HTTP :: Dynamic Content\",\n ],\n zip_safe=False,\n)\n","new_contents":"import multiprocessing\nfrom setuptools import setup, find_packages\n\nsetup(\n name='sow-generator',\n version='0.1.1',\n description='Create a scope of work from templates and version controlled documentation.',\n long_description = open('README.rst', 'r').read() + open('CHANGELOG.rst', 'r').read() + open('AUTHORS.rst', 'r').read(),\n author='Hedley Roos',\n author_email='hedley@praekelt.com',\n license='BSD',\n url='http:\/\/github.com\/praekelt\/sow-generator',\n packages = find_packages(),\n install_requires = [\n 'Django<1.7',\n 'South',\n 'celery',\n 'django-celery',\n 'raven',\n 'PyYAML',\n 'requests',\n 'github3.py',\n 'pyandoc',\n 'django-object-tools',\n 'django-adminplus'\n ],\n include_package_data=True,\n tests_require=[\n 'django-setuptest>=0.1.4',\n ],\n test_suite=\"setuptest.setuptest.SetupTestSuite\",\n classifiers=[\n \"Programming Language :: Python\",\n \"License :: OSI Approved :: BSD License\",\n \"Development Status :: 4 - Beta\",\n \"Operating System :: OS Independent\",\n \"Framework :: Django\",\n \"Intended Audience :: Developers\",\n \"Topic :: Internet :: WWW\/HTTP :: Dynamic Content\",\n ],\n zip_safe=False,\n)\n","subject":"Bump version to see if pip installs it","message":"Bump version to see if pip installs it\n","lang":"Python","license":"bsd-3-clause","repos":"praekelt\/sow-generator"} {"commit":"79f9a028215f6765898ae43faf2ba7e4aaf98f0c","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nfrom setuptools import setup, find_packages\n\ndef read(fname):\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n\nsetup(\n name='django-pgallery',\n version=__import__('pgallery').__version__,\n description='Photo gallery app for PostgreSQL and Django.',\n long_description=read('README.rst'),\n author='Zbigniew Siciarz',\n author_email='zbigniew@siciarz.net',\n url='http:\/\/github.com\/zsiciarz\/django-pgallery',\n download_url='http:\/\/pypi.python.org\/pypi\/django-pgallery',\n license='MIT',\n install_requires=[\n 'Django>=1.4',\n 'Pillow',\n 'psycopg2>=2.4',\n 'django-markitup>=1.0',\n 'django-model-utils>=1.1',\n 'djorm-ext-core>=0.4.2',\n 'djorm-ext-expressions>=0.4.4',\n 'djorm-ext-hstore>=0.4.2',\n 'djorm-ext-pgarray',\n 'sorl-thumbnail>=11',\n ],\n packages=find_packages(),\n include_package_data=True,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Utilities',\n ],\n)\n","new_contents":"import os\nfrom setuptools import setup, find_packages\n\ndef read(fname):\n return open(os.path.join(os.path.dirname(__file__), fname)).read()\n\nsetup(\n name='django-pgallery',\n version=__import__('pgallery').__version__,\n description='Photo gallery app for PostgreSQL and Django.',\n long_description=read('README.rst'),\n author='Zbigniew Siciarz',\n author_email='zbigniew@siciarz.net',\n url='http:\/\/github.com\/zsiciarz\/django-pgallery',\n download_url='http:\/\/pypi.python.org\/pypi\/django-pgallery',\n license='MIT',\n install_requires=[\n 'Django>=1.4',\n 'Pillow',\n 'psycopg2>=2.4',\n 'django-markitup>=1.0',\n 'django-model-utils>=1.1',\n 'djorm-ext-core>=0.4.2',\n 'djorm-ext-expressions>=0.4.4',\n 'djorm-ext-hstore>=0.4.2',\n 'djorm-ext-pgarray',\n 'sorl-thumbnail>=11',\n ],\n packages=find_packages(exclude=['tests']),\n include_package_data=True,\n classifiers=[\n 'Development Status :: 3 - Alpha',\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Utilities',\n ],\n)\n","subject":"Exclude tests package from installing.","message":"Exclude tests package from installing.\n","lang":"Python","license":"mit","repos":"zsiciarz\/django-pgallery,zsiciarz\/django-pgallery"} {"commit":"eeecf68d2d59bc2233478b01748cbf88bab85722","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom distutils.core import setup\n\nexecfile('facebook\/version.py')\n\nsetup(\n name = 'Facebook',\n version = __version__,\n description = 'Facebook makes it even easier to interact with Facebook\\'s Graph API',\n long_description = open('README.rst').read() + '\\n\\n' + open('HISTORY.rst').read(),\n author = 'Johannes Gorset',\n author_email = 'jgorset@gmail.com',\n url = 'http:\/\/github.com\/jgorset\/facebook',\n packages = [\n 'facebook'\n ]\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom distutils.core import setup\n\nexecfile('facebook\/version.py')\n\nsetup(\n name='Facebook',\n version=__version__,\n description='Facebook makes it even easier to interact \"+\\\n \"with Facebook\\'s Graph API',\n long_description=open('README.rst').read() + '\\n\\n' +\n open('HISTORY.rst').read(),\n author='Johannes Gorset',\n author_email='jgorset@gmail.com',\n url='http:\/\/github.com\/jgorset\/facebook',\n requires=['facepy'],\n packages=[\n 'facebook'\n ]\n)\n","subject":"Add missing requires and PEP8ize.","message":"Add missing requires and PEP8ize.\n","lang":"Python","license":"mit","repos":"jgorset\/facebook,vyyvyyv\/facebook,jgorset\/facebook,vyyvyyv\/facebook"} {"commit":"932182858efcdf3d76d3a19cba09967c680e907f","old_file":"setup.py","new_file":"setup.py","old_contents":"# -*- coding: utf-8; mode: python; -*-\n\"\"\"\nA package that implements offline messages for Django\nWeb Framework.\n\n(C) 2011 oDesk www.oDesk.com w\/revisions by Zapier.com\n\"\"\"\n\nfrom setuptools import setup\n\nsetup(\n name='django-offline-messages',\n version='0.3.1',\n description='A package that implements offline messages for Django plus more',\n long_description='A package that implements offline messages for Django Web Framework',\n license='BSD',\n keywords='django offline messages',\n url='https:\/\/github.com\/zapier\/django-offline-messages',\n author='oDesk, www.odesk.com',\n author_email='developers@odesk.com',\n maintainer='Bryan Helmig',\n maintainer_email='bryan@zapier.com',\n packages=['offline_messages', 'offline_messages.migrations'],\n classifiers=['Development Status :: 3 - Alpha',\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n test_suite='tests.runtests.runtests'\n )\n","new_contents":"# -*- coding: utf-8; mode: python; -*-\n\"\"\"\nA package that implements offline messages for Django\nWeb Framework.\n\n(C) 2011 oDesk www.oDesk.com w\/revisions by Zapier.com\n\"\"\"\n\nfrom setuptools import setup\n\nsetup(\n name='django-offline-messages',\n version='0.3.1',\n description='A package that implements offline messages for Django plus more',\n long_description='A package that implements offline messages for Django Web Framework',\n license='BSD',\n keywords='django offline messages',\n url='https:\/\/github.com\/zapier\/django-offline-messages',\n author='oDesk, www.odesk.com',\n author_email='developers@odesk.com',\n maintainer='Bryan Helmig',\n maintainer_email='bryan@zapier.com',\n packages=['offline_messages', 'offline_messages.migrations'],\n classifiers=['Development Status :: 3 - Alpha',\n 'Environment :: Web Environment',\n 'Framework :: Django',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: OS Independent',\n 'Programming Language :: Python',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n ],\n test_suite='tests.runtests.runtests',\n install_requires=['django-jsonfield']\n)\n","subject":"Add django-jsonfield as a requirement","message":"Add django-jsonfield as a requirement\n","lang":"Python","license":"bsd-3-clause","repos":"dym\/django-offline-messages"} {"commit":"418ffb8439f6c00d1704944aee28cda756ca6455","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\nimport os\n\nversion = '0.6'\n\n\ndef read(filename):\n return open(os.path.join(os.path.dirname(__file__), filename)).read()\n\nsetup(\n name='imbox',\n version=version,\n description=\"Python IMAP for Human beings\",\n long_description=read('README.md'),\n keywords='email, IMAP, parsing emails',\n author='Martin Rusev',\n author_email='martinrusev@live.com',\n url='https:\/\/github.com\/martinrusev\/imbox',\n license='MIT',\n packages=['imbox'],\n package_dir={'imbox': 'imbox'},\n zip_safe=False,\n install_requires=['six'],\n classifiers=(\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n ),\n)\n","new_contents":"from setuptools import setup\nimport os\n\nversion = '0.7'\n\n\ndef read(filename):\n return open(os.path.join(os.path.dirname(__file__), filename)).read()\n\nsetup(\n name='imbox',\n version=version,\n description=\"Python IMAP for Human beings\",\n long_description=read('README.md'),\n keywords='email, IMAP, parsing emails',\n author='Martin Rusev',\n author_email='martinrusev@live.com',\n url='https:\/\/github.com\/martinrusev\/imbox',\n license='MIT',\n packages=['imbox'],\n package_dir={'imbox': 'imbox'},\n zip_safe=False,\n install_requires=['six'],\n classifiers=(\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.3',\n 'Programming Language :: Python :: 3.4',\n ),\n)\n","subject":"Update version and push to pypi","message":"Update version and push to pypi\n","lang":"Python","license":"mit","repos":"martinrusev\/imbox"} {"commit":"e10d5b8a975f1d9482f845c1476edcc4598bb84a","old_file":"setup.py","new_file":"setup.py","old_contents":"# -*- coding: utf-8 -*-\nimport os\nfrom setuptools import find_packages, setup\nfrom pyglins import __version__, __description__\n\n\ndef read_readme():\n with open(os.path.join(os.path.dirname(__file__), 'README.md')) as file:\n return file.read()\n\n\nsetup(name='pyglins',\n version=__version__,\n description=__description__,\n long_description=read_readme(),\n classifiers=[\n 'Development Status :: 1 - Planning',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n 'Topic :: Utilities',\n 'Operating System :: OS Independent'],\n keywords='plugin manager',\n author='Javier Caballero',\n author_email='paxet83@gmail.com',\n url='https:\/\/github.com\/paxet\/pyglins',\n license='MIT',\n packages=find_packages(exclude=['tests']),\n )\n\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom setuptools import find_packages, setup\nfrom pyglins import __version__, __description__\n\n\ndef read_readme():\n try:\n import pypandoc\n description = pypandoc.convert('README.md', 'rst')\n except (IOError, ImportError):\n with open('README.md') as file:\n description = file.read()\n return description\n\n\nsetup(name='pyglins',\n version=__version__,\n description=__description__,\n long_description=read_readme(),\n classifiers=[\n 'Development Status :: 1 - Planning',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Topic :: Software Development :: Libraries :: Python Modules',\n 'Topic :: Utilities',\n 'Operating System :: OS Independent'],\n keywords='plugin manager',\n author='Javier Caballero',\n author_email='paxet83@gmail.com',\n url='https:\/\/github.com\/paxet\/pyglins',\n license='MIT',\n packages=find_packages(exclude=['tests']),\n )\n\n","subject":"Convert README.md to reStructuredText with pypandoc","message":"Convert README.md to reStructuredText with pypandoc\n","lang":"Python","license":"mit","repos":"paxet\/pyglins"} {"commit":"1b4ca927245e424a340f5caa2ed9bda615dd8a5c","old_file":"setup.py","new_file":"setup.py","old_contents":"\"\"\"setup.py\n\n..codeauthor:: John Lane <jlane@fanthreesixty.com>\n\n\"\"\"\n\nfrom setuptools import setup, find_packages\nfrom sda import __author__, __email__, __license__, __version__\n\n\nsetup(\n name='sda',\n version=__version__,\n packages=find_packages(),\n scripts=[],\n description='A wrapper for Selenium. This library uses custom data attributes to accelerate '\n 'testing through the Selenium framework',\n author=__author__,\n author_email=__email__,\n url='https:\/\/github.com\/jlane9\/selenium-data-attributes',\n download_url='https:\/\/github.com\/jlane9\/selenium-data-attributes\/tarball\/{}'.format(__version__),\n keywords='testing selenium qa web automation',\n install_requires=['lxml', 'cssselect'],\n license=__license__,\n classifiers=['Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'License :: OSI Approved :: MIT License',\n 'Topic :: Software Development :: Quality Assurance',\n 'Topic :: Software Development :: Testing'])\n","new_contents":"\"\"\"setup.py\n\n..codeauthor:: John Lane <jlane@fanthreesixty.com>\n\n\"\"\"\n\nfrom setuptools import setup, find_packages\nfrom sda import __author__, __email__, __license__, __version__\n\n\nsetup(\n name='sda',\n version=__version__,\n packages=find_packages(),\n scripts=[],\n description='A wrapper for Selenium. This library uses custom data attributes to accelerate '\n 'testing through the Selenium framework',\n author=__author__,\n author_email=__email__,\n url='https:\/\/github.com\/jlane9\/selenium-data-attributes',\n download_url='https:\/\/github.com\/jlane9\/selenium-data-attributes\/tarball\/{}'.format(__version__),\n keywords='testing selenium qa web automation',\n install_requires=['lxml', 'cssselect'],\n license=__license__,\n classifiers=['Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 2.6',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.2',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'License :: OSI Approved :: MIT License',\n 'Topic :: Software Development :: Quality Assurance',\n 'Topic :: Software Development :: Testing'])\n","subject":"Upgrade project to production ready","message":"Upgrade project to production ready\n\n","lang":"Python","license":"mit","repos":"jlane9\/selenium_data_attributes,jlane9\/selenium_data_attributes"} {"commit":"9a424163fa97bbb5ab7b19ecb9707fa05cb1bef7","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup\n\n# Version info -- read without importing\n_locals = {}\nwith open('releases\/_version.py') as fp:\n exec(fp.read(), None, _locals)\nversion = _locals['__version__']\n\nsetup(\n name='releases',\n version=version,\n description='A Sphinx extension for changelog manipulation',\n long_description=open(\"README.rst\").read(),\n author='Jeff Forcier',\n author_email='jeff@bitprophet.org',\n url='https:\/\/github.com\/bitprophet\/releases',\n packages=['releases'],\n install_requires=[\n 'semantic_version<2.7',\n 'sphinx>=1.3,<1.8',\n ],\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: Unix',\n 'Operating System :: POSIX',\n 'Operating System :: Microsoft :: Windows',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Documentation',\n 'Topic :: Documentation',\n 'Topic :: Documentation :: Sphinx',\n ],\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup\n\n# Version info -- read without importing\n_locals = {}\nwith open('releases\/_version.py') as fp:\n exec(fp.read(), None, _locals)\nversion = _locals['__version__']\n\nsetup(\n name='releases',\n version=version,\n description='A Sphinx extension for changelog manipulation',\n long_description=open(\"README.rst\").read(),\n author='Jeff Forcier',\n author_email='jeff@bitprophet.org',\n url='https:\/\/github.com\/bitprophet\/releases',\n packages=['releases'],\n install_requires=[\n 'semantic_version<2.7',\n 'sphinx>=1.3',\n ],\n classifiers=[\n 'Development Status :: 5 - Production\/Stable',\n 'Intended Audience :: Developers',\n 'License :: OSI Approved :: BSD License',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: Unix',\n 'Operating System :: POSIX',\n 'Operating System :: Microsoft :: Windows',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2.7',\n 'Programming Language :: Python :: 3.4',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Topic :: Software Development',\n 'Topic :: Software Development :: Documentation',\n 'Topic :: Documentation',\n 'Topic :: Documentation :: Sphinx',\n ],\n)\n","subject":"Remove upper limit for Sphinx version","message":"Remove upper limit for Sphinx version\n","lang":"Python","license":"bsd-2-clause","repos":"bitprophet\/releases"} {"commit":"b389a7c0a3129573d28308cd175002b612b016f6","old_file":"setup.py","new_file":"setup.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"Setup for the DeepOBS package\"\"\"\n\nimport setuptools\n\n\ndef readme():\n with open('README.md') as f:\n return f.read()\n\nsetuptools.setup(\n name='deepobs',\n version='1.1.1',\n description='Deep Learning Optimizer Benchmark Suite',\n long_description=readme(),\n author='Frank Schneider, Lukas Balles and Philipp Hennig,',\n author_email='frank.schneider@tue.mpg.de',\n license='MIT',\n packages=setuptools.find_packages(),\n classifiers=[\n \"Programming Language :: Python :: 3.6\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: Science\/Research\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Topic :: Scientific\/Engineering :: Artificial Intelligence\",\n ],\n install_requires=[\n 'argparse', 'numpy', 'pandas', 'matplotlib', 'matplotlib2tikz',\n 'seaborn'\n ],\n scripts=[\n 'deepobs\/scripts\/deepobs_prepare_data.sh',\n 'deepobs\/scripts\/deepobs_get_baselines.sh',\n 'deepobs\/scripts\/deepobs_plot_results.py',\n 'deepobs\/scripts\/deepobs_estimate_runtime.py'\n ],\n zip_safe=False)\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"Setup for the DeepOBS package\"\"\"\n\nimport setuptools\n\n\ndef readme():\n with open(\"README.md\") as f:\n return f.read()\n\n\nsetuptools.setup(\n name=\"deepobs\",\n version=\"1.1.2\",\n description=\"Deep Learning Optimizer Benchmark Suite\",\n long_description=readme(),\n author=\"Frank Schneider, Lukas Balles and Philipp Hennig,\",\n author_email=\"frank.schneider@tue.mpg.de\",\n license=\"MIT\",\n packages=setuptools.find_packages(),\n classifiers=[\n \"Programming Language :: Python :: 3.6\",\n \"Intended Audience :: Developers\",\n \"Intended Audience :: Science\/Research\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Topic :: Scientific\/Engineering :: Artificial Intelligence\",\n ],\n install_requires=[\n \"argparse\",\n \"numpy\",\n \"pandas\",\n \"matplotlib\",\n \"matplotlib2tikz==0.6.18\",\n \"seaborn\",\n ],\n scripts=[\n \"deepobs\/scripts\/deepobs_prepare_data.sh\",\n \"deepobs\/scripts\/deepobs_get_baselines.sh\",\n \"deepobs\/scripts\/deepobs_plot_results.py\",\n \"deepobs\/scripts\/deepobs_estimate_runtime.py\",\n ],\n zip_safe=False,\n)\n","subject":"Install specific version of matplotlib2tikz","message":"Install specific version of matplotlib2tikz\n","lang":"Python","license":"mit","repos":"fsschneider\/DeepOBS,fsschneider\/DeepOBS"} {"commit":"51d0498f1c444f00ce982a93d8c9fdfb72a196b4","old_file":"setup.py","new_file":"setup.py","old_contents":"#! \/usr\/bin\/env python\n'''\nThis file is part of targetcli.\nCopyright (c) 2011-2013 by Datera, Inc\n\nLicensed under the Apache License, Version 2.0 (the \"License\"); you may\nnot use this file except in compliance with the License. You may obtain\na copy of the License at\n\n http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\nWARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\nLicense for the specific language governing permissions and limitations\nunder the License.\n'''\n\nfrom distutils.core import setup\n\n__version__ = ''\nexecfile('targetcli\/version.py')\n\nsetup(\n name = 'targetcli-fb',\n version = __version__,\n description = 'An administration shell for RTS storage targets.',\n license = 'Apache 2.0',\n maintainer = 'Andy Grover',\n maintainer_email = 'agrover@redhat.com',\n url = 'http:\/\/github.com\/agrover\/targetcli-fb',\n packages = ['targetcli'],\n scripts = ['scripts\/targetcli'],\n )\n","new_contents":"#! \/usr\/bin\/env python\n'''\nThis file is part of targetcli.\nCopyright (c) 2011-2013 by Datera, Inc\n\nLicensed under the Apache License, Version 2.0 (the \"License\"); you may\nnot use this file except in compliance with the License. You may obtain\na copy of the License at\n\n http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n\nUnless required by applicable law or agreed to in writing, software\ndistributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\nWARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\nLicense for the specific language governing permissions and limitations\nunder the License.\n'''\n\nfrom distutils.core import setup\n\n__version__ = ''\nexec(open('targetcli\/version.py').read())\n\nsetup(\n name = 'targetcli-fb',\n version = __version__,\n description = 'An administration shell for RTS storage targets.',\n license = 'Apache 2.0',\n maintainer = 'Andy Grover',\n maintainer_email = 'agrover@redhat.com',\n url = 'http:\/\/github.com\/agrover\/targetcli-fb',\n packages = ['targetcli'],\n scripts = ['scripts\/targetcli'],\n )\n","subject":"Replace execfile() with exec() since it does not work with Python 3","message":"Replace execfile() with exec() since it does not work with Python 3\n\nSigned-off-by: Christophe Vu-Brugier <1930e27f67e1e10d51770b88cb06d386f1aa46ae@yahoo.fr>\n","lang":"Python","license":"apache-2.0","repos":"agrover\/targetcli-fb,cloud4life\/targetcli-fb,cvubrugier\/targetcli-fb"} {"commit":"99b79326fa18f46fe449e11fd0bfa17814d7a148","old_file":"setup.py","new_file":"setup.py","old_contents":"from distutils.core import setup\n\nsetup(\n name='resync',\n version='0.6.1',\n packages=['resync'],\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n \"Environment :: Web Environment\"],\n author='Simeon Warner',\n author_email='simeon.warner@cornell.edu',\n long_description=open('README.md').read(),\n url='http:\/\/github.com\/resync\/resync',\n)\n","new_contents":"from distutils.core import setup\n\nsetup(\n name='resync',\n version='0.6.1',\n packages=['resync'],\n scripts=['bin\/resync'],\n classifiers=[\"Development Status :: 3 - Alpha\",\n \"Programming Language :: Python\",\n \"Topic :: Software Development :: Libraries :: Python Modules\",\n \"Environment :: Web Environment\"],\n author='Simeon Warner',\n author_email='simeon.warner@cornell.edu',\n long_description=open('README.md').read(),\n url='http:\/\/github.com\/resync\/resync',\n)\n","subject":"Add config to install resync script","message":"Add config to install resync script\n","lang":"Python","license":"apache-2.0","repos":"resync\/resync,dans-er\/resync,lindareijnhoudt\/resync,lindareijnhoudt\/resync,dans-er\/resync"} {"commit":"1594644990fe6c621d309b587d844669ec273dec","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nimport os\nimport sys\nfrom setuptools import setup, find_packages\nfrom pkg_resources import resource_filename\n\n# depending on your execution context the version file\n# may be located in a different place!\nvsn_path = resource_filename(__name__, 'hvac\/version')\nif not os.path.exists(vsn_path):\n vsn_path = resource_filename(__name__, 'version')\n if not os.path.exists(vsn_path):\n print(\"%s is missing\" % vsn_path)\n sys.exit(1)\n\nsetup(\n name='hvac',\n version=open(vsn_path, 'r').read(),\n description='HashiCorp Vault API client',\n author='Ian Unruh',\n author_email='ianunruh@gmail.com',\n url='https:\/\/github.com\/ianunruh\/hvac',\n keywords=['hashicorp', 'vault'],\n classifiers=['License :: OSI Approved :: Apache Software License'],\n packages=find_packages(),\n install_requires=[\n 'requests>=2.7.0',\n ],\n include_package_data=True,\n package_data={'hvac': ['version']},\n extras_require={\n 'parser': ['pyhcl>=0.2.1,<0.3.0']\n }\n)\n","new_contents":"#!\/usr\/bin\/env python\nimport os\nimport sys\nfrom setuptools import setup, find_packages\nfrom pkg_resources import resource_filename\n\nwith open(\"README.md\", \"r\") as fh:\n long_description = fh.read()\n\n# depending on your execution context the version file\n# may be located in a different place!\nvsn_path = resource_filename(__name__, 'hvac\/version')\nif not os.path.exists(vsn_path):\n vsn_path = resource_filename(__name__, 'version')\n if not os.path.exists(vsn_path):\n print(\"%s is missing\" % vsn_path)\n sys.exit(1)\n\nsetup(\n name='hvac',\n version=open(vsn_path, 'r').read(),\n description='HashiCorp Vault API client',\n long_description=long_description,\n long_description_content_type=\"text\/markdown\",\n author='Ian Unruh',\n author_email='ianunruh@gmail.com',\n url='https:\/\/github.com\/ianunruh\/hvac',\n keywords=['hashicorp', 'vault'],\n classifiers=['License :: OSI Approved :: Apache Software License'],\n packages=find_packages(),\n install_requires=[\n 'requests>=2.7.0',\n ],\n include_package_data=True,\n package_data={'hvac': ['version']},\n extras_require={\n 'parser': ['pyhcl>=0.2.1,<0.3.0']\n }\n)\n","subject":"Include README.md in package metadata","message":"Include README.md in package metadata\n","lang":"Python","license":"apache-2.0","repos":"ianunruh\/hvac,ianunruh\/hvac"} {"commit":"769cf7b47fde883e4b44cec3adf8944bd62f7363","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom setuptools import setup\n\nimport inflection\n\nsetup(\n name='inflection',\n version=inflection.__version__,\n description=\"A port of Ruby on Rails inflector to Python\",\n long_description=open('README.rst').read(),\n author='Janne Vanhala',\n author_email='janne.vanhala@gmail.com',\n url='http:\/\/github.com\/jpvanhal\/inflection',\n license='MIT',\n py_modules=['inflection'],\n zip_safe=False,\n python_requires='>=3.5',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n ],\n)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nfrom setuptools import setup\n\nimport inflection\n\nsetup(\n name='inflection',\n version=inflection.__version__,\n description=\"A port of Ruby on Rails inflector to Python\",\n long_description=open('README.rst').read(),\n author='Janne Vanhala',\n author_email='janne.vanhala@gmail.com',\n url='https:\/\/github.com\/jpvanhal\/inflection',\n license='MIT',\n py_modules=['inflection'],\n zip_safe=False,\n python_requires='>=3.5',\n classifiers=[\n 'Development Status :: 4 - Beta',\n 'Intended Audience :: Developers',\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 3',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Programming Language :: Python :: 3.7',\n 'Programming Language :: Python :: 3.8',\n 'Programming Language :: Python :: Implementation :: CPython',\n 'Programming Language :: Python :: Implementation :: PyPy',\n ],\n)\n","subject":"Use HTTPS for the project url","message":"Use HTTPS for the project url\n","lang":"Python","license":"mit","repos":"jpvanhal\/inflection"} {"commit":"d96027040f96de15c2bddf8ba5b94711b5af6e1f","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n# coding=utf-8\n__author__ = 'kulakov.ilya@gmail.com'\n\nfrom setuptools import setup\nfrom sys import platform\n\n\nREQUIREMENTS = []\n\n\nif platform.startswith('darwin'):\n REQUIREMENTS.append('pyobjc >= 2.5')\n\n\nsetup(\n name=\"Power\",\n version=\"1.1\",\n description=\"Cross-platform system power status information.\",\n author=\"Ilya Kulakov\",\n author_email=\"kulakov.ilya@gmail.com\",\n url=\"https:\/\/github.com\/Kentzo\/Power\",\n platforms=[\"Mac OS X 10.6+\", \"Windows XP+\", \"Linux 2.6+\"],\n packages=['power'],\n classifiers=[\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: Microsoft :: Windows',\n 'Operating System :: POSIX :: Linux',\n 'Topic :: System :: Power (UPS)',\n 'Topic :: System :: Hardware',\n 'Topic :: System :: Monitoring'\n ],\n install_requires=REQUIREMENTS\n)\n","new_contents":"#!\/usr\/bin\/env python\n# coding=utf-8\n__author__ = 'kulakov.ilya@gmail.com'\n\nfrom setuptools import setup\nfrom sys import platform\n\n\nREQUIREMENTS = []\n\n\n# Not avaialable at PyPi yet\n# if platform.startswith('darwin'):\n # REQUIREMENTS.append('pyobjc >= 2.5')\n\n\nsetup(\n name=\"Power\",\n version=\"1.1\",\n description=\"Cross-platform system power status information.\",\n author=\"Ilya Kulakov\",\n author_email=\"kulakov.ilya@gmail.com\",\n url=\"https:\/\/github.com\/Kentzo\/Power\",\n platforms=[\"Mac OS X 10.6+\", \"Windows XP+\", \"Linux 2.6+\"],\n packages=['power'],\n classifiers=[\n 'Natural Language :: English',\n 'License :: OSI Approved :: MIT License',\n 'Programming Language :: Python',\n 'Programming Language :: Python :: 2',\n 'Operating System :: MacOS :: MacOS X',\n 'Operating System :: Microsoft :: Windows',\n 'Operating System :: POSIX :: Linux',\n 'Topic :: System :: Power (UPS)',\n 'Topic :: System :: Hardware',\n 'Topic :: System :: Monitoring'\n ],\n install_requires=REQUIREMENTS\n)\n","subject":"Remove PyObjC requirement since it's not avialable at PyPi yet.","message":"Remove PyObjC requirement since it's not avialable at PyPi yet.\n","lang":"Python","license":"mit","repos":"Kentzo\/Power"} {"commit":"7429bfd2fcca23922bd9fe3a928cb336b0bb6b23","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom distutils.core import setup\n\nversion = \"0.12\"\n\n\"\"\"Setup script for pexif\"\"\"\n\nsetup (\n name = \"pexif\",\n version = version,\n description = \"A module for editing JPEG EXIF data\",\n long_description = \"This module allows you to parse and edit the EXIF data tags in a JPEG image.\",\n author = \"Ben Leslie\",\n author_email = \"benno@benno.id.au\",\n url = \"http:\/\/www.benno.id.au\/code\/pexif\/\",\n download_url = \"http:\/\/www.benno.id.au\/code\/pexif\/pexif-%s.tar.gz\" % version,\n license = \"http:\/\/www.opensource.org\/licenses\/mit-license.php\",\n py_modules = [\"pexif\"],\n scripts = [\"scripts\/dump_exif.py\", \"scripts\/setgps.py\", \"scripts\/getgps.py\", \"scripts\/noop.py\",\n \"scripts\/timezone.py\"],\n platforms = [\"any\"],\n classifiers = [\"Development Status :: 4 - Beta\",\n \"Intended Audience :: Developers\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"License :: OSI Approved :: Python Software Foundation License\",\n \"Topic :: Multimedia :: Graphics\"]\n )\n \n \n","new_contents":"#!\/usr\/bin\/env python\n\nfrom distutils.core import setup\n\nversion = \"0.13\"\n\n\"\"\"Setup script for pexif\"\"\"\n\nsetup (\n name = \"pexif\",\n version = version,\n description = \"A module for editing JPEG EXIF data\",\n long_description = \"This module allows you to parse and edit the EXIF data tags in a JPEG image.\",\n author = \"Ben Leslie\",\n author_email = \"benno@benno.id.au\",\n url = \"http:\/\/www.benno.id.au\/code\/pexif\/\",\n download_url = \"http:\/\/www.benno.id.au\/code\/pexif\/pexif-%s.tar.gz\" % version,\n license = \"http:\/\/www.opensource.org\/licenses\/mit-license.php\",\n py_modules = [\"pexif\"],\n scripts = [\"scripts\/dump_exif.py\", \"scripts\/setgps.py\", \"scripts\/getgps.py\", \"scripts\/noop.py\",\n \"scripts\/timezone.py\"],\n platforms = [\"any\"],\n classifiers = [\"Development Status :: 4 - Beta\",\n \"Intended Audience :: Developers\",\n \"Operating System :: OS Independent\",\n \"Programming Language :: Python\",\n \"License :: OSI Approved :: Python Software Foundation License\",\n \"Topic :: Multimedia :: Graphics\"]\n )\n \n \n","subject":"Bump version number to 0.13","message":"Bump version number to 0.13\n\ngit-svn-id: 2223166f32dfc0252f77f60c9446e3f27730b4ad@17 8ba0a3ab-1543-0410-831c-f9a674249a0d\n","lang":"Python","license":"mit","repos":"bennoleslie\/pexif,untereiner\/pexif,ebrelsford\/pexif"} {"commit":"323498419bfe080c9807586c2e5cc2678b7d2fd4","old_file":"setup.py","new_file":"setup.py","old_contents":"from setuptools import setup\nfrom os import path\n\nreadme = open(path.join(path.abspath(path.dirname(__file__)), 'README.md')).read()\n\nsetup(\n name='bottle-websocket',\n version='0.2',\n author='Zach Kelling',\n author_email='zeekayy@gmail.com',\n packages=['bottle_websocket',],\n description='WebSockets for bottle',\n long_description=readme,\n install_requires=['bottle', 'gevent-websocket'],\n)\n","new_contents":"from setuptools import setup\n\nsetup(\n name='bottle-websocket',\n version='0.2.5',\n author='Zach Kelling',\n author_email='zeekayy@gmail.com',\n packages=['bottle_websocket',],\n package_data={'': ['README.md']},\n description='WebSockets for bottle',\n long_description=open('README.md'),\n install_requires=['bottle', 'gevent-websocket'],\n)\n","subject":"Use package_data, and bump version.","message":"Use package_data, and bump version.\n","lang":"Python","license":"mit","repos":"xujun10110\/bottle-websocket,zeekay\/bottle-websocket"} {"commit":"7b825d860c84eae8b5a74ef16ae8e1da08dde888","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\nfrom setuptools import setup, find_packages\n\n\nsetup(\n name='synapse',\n version='0.1.0',\n description='Synapse Distributed Key-Value Hypergraph Analysis Framework',\n author='Invisigoth Kenshoto',\n author_email='invisigoth.kenshoto@gmail.com',\n url='https:\/\/github.com\/vertexproject\/synapse',\n license='Apache License 2.0',\n\n packages=find_packages(exclude=['scripts',\n ]),\n\n include_package_data=True,\n\n install_requires=[\n 'tornado>=3.2.2,<5.0.0',\n 'pyOpenSSL>=16.2.0,<18.0.0',\n 'msgpack==0.5.1',\n 'xxhash>=1.0.1,<2.0.0',\n 'lmdb>=0.94,<1.0.0',\n 'regex>=2017.9.23',\n 'PyYAML>=3.12,<4.0',\n ],\n\n classifiers=[\n 'Development Status :: 4 - Beta',\n\n 'License :: OSI Approved :: Apache Software License',\n\n 'Topic :: System :: Clustering',\n 'Topic :: System :: Distributed Computing',\n 'Topic :: System :: Software Distribution',\n\n 'Programming Language :: Python :: 3.6',\n ],\n)\n","new_contents":"#!\/usr\/bin\/env python\nfrom setuptools import setup, find_packages\n\n\nsetup(\n name='synapse',\n version='0.1.0-alpha',\n description='Synapse Distributed Key-Value Hypergraph Analysis Framework',\n author='Invisigoth Kenshoto',\n author_email='invisigoth.kenshoto@gmail.com',\n url='https:\/\/github.com\/vertexproject\/synapse',\n license='Apache License 2.0',\n\n packages=find_packages(exclude=['scripts',\n ]),\n\n include_package_data=True,\n\n install_requires=[\n 'tornado>=3.2.2,<5.0.0',\n 'pyOpenSSL>=16.2.0,<18.0.0',\n 'msgpack==0.5.1',\n 'xxhash>=1.0.1,<2.0.0',\n 'lmdb>=0.94,<1.0.0',\n 'regex>=2017.9.23',\n 'PyYAML>=3.12,<4.0',\n ],\n\n classifiers=[\n 'Development Status :: 4 - Beta',\n\n 'License :: OSI Approved :: Apache Software License',\n\n 'Topic :: System :: Clustering',\n 'Topic :: System :: Distributed Computing',\n 'Topic :: System :: Software Distribution',\n\n 'Programming Language :: Python :: 3.6',\n ],\n)\n","subject":"Add -alpha prerelease version to version.","message":"Add -alpha prerelease version to version.\n","lang":"Python","license":"apache-2.0","repos":"vertexproject\/synapse,vertexproject\/synapse,vertexproject\/synapse"} {"commit":"8a97210aa5d83f6eac266a19fbad9a2159b14328","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport sys\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nif sys.argv[-1] == 'publish':\n os.system('python setup.py sdist bdist_wininst upload -r pypi')\n sys.exit()\n\nwith open('README.rst') as f:\n readme = f.read()\nwith open('LICENSE') as f:\n license = f.read()\n\nsetup(\n name='django-rest-surveys',\n version='0.1.0',\n description='A RESTful backend for giving surveys.',\n long_description=readme,\n author='Designlab',\n author_email='hello@trydesignlab.com',\n url='https:\/\/github.com\/danxshap\/django-rest-surveys',\n packages=['rest_surveys'],\n package_data={'': ['LICENSE']},\n package_dir={'rest_surveys': 'rest_surveys'},\n install_requires=['Django', 'django-inline-ordering'],\n license=license,\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport sys\n\ntry:\n from setuptools import setup\nexcept ImportError:\n from distutils.core import setup\n\nif sys.argv[-1] == 'publish':\n os.system('python setup.py sdist bdist_wininst upload -r pypi')\n sys.exit()\n\nwith open('README.rst') as f:\n readme = f.read()\nwith open('LICENSE') as f:\n license = f.read()\n\nsetup(\n name='django-rest-surveys',\n version='0.1.0',\n description='A RESTful backend for giving surveys.',\n long_description=readme,\n author='Designlab',\n author_email='hello@trydesignlab.com',\n url='https:\/\/github.com\/danxshap\/django-rest-surveys',\n packages=['rest_surveys'],\n package_data={'': ['LICENSE']},\n package_dir={'rest_surveys': 'rest_surveys'},\n install_requires=['Django>=1.7', 'djangorestframework>=3.0', 'django-inline-ordering'],\n license=license,\n)\n","subject":"Set the required versions of required packages","message":"Set the required versions of required packages\n","lang":"Python","license":"mit","repos":"danxshap\/django-rest-surveys"} {"commit":"eb339f661aec7c01a727a01fd8b9e2880d99cc9c","old_file":"setup.py","new_file":"setup.py","old_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup, find_packages\n\n# Get version string\nwith open('gdx2py\/version.py') as f: exec(f.read())\n\nsetup(name='GDX2py',\n version=__version__, # pylint: disable=undefined-variable\n author='Erkka Rinne',\n author_email='erkka.rinne@vtt.fi',\n description='Read and write GAMS Data eXchange (GDX) files using Python',\n python_requires='>=3.6',\n install_requires=[\n 'gdxcc>=7',\n ],\n setup_requires=['pytest-runner'],\n tests_require=['pytest', 'pytest-datadir'],\n url='https:\/\/github.com\/ererkka\/GDX2py',\n packages=find_packages(exclude=['tests']),\n)\n","new_contents":"#!\/usr\/bin\/env python\n\nfrom setuptools import setup, find_packages\n\n# Get version string\nwith open('gdx2py\/version.py') as f: exec(f.read())\n\nsetup(name='GDX2py',\n version=__version__, # pylint: disable=undefined-variable\n author='Erkka Rinne',\n author_email='erkka.rinne@vtt.fi',\n description='Read and write GAMS Data eXchange (GDX) files using Python',\n python_requires='>=3.6',\n install_requires=[\n 'gdxcc>=7',\n ],\n setup_requires=['pytest-runner'],\n tests_require=['pytest', 'pytest-datadir'],\n url='https:\/\/github.com\/ererkka\/GDX2py',\n packages=find_packages(exclude=['tests']),\n classifiers=[\n \"Programming Language :: Python :: 3\",\n \"License :: OSI Approved :: MIT License\",\n \"Operating System :: OS Independent\",\n \"Topic :: Scientific\/Engineering\",\n \"Intended Audience :: Developers\",\n \"Development Status :: 4 - Beta\"\n ],\n)\n","subject":"Add classifiers to package info","message":"Add classifiers to package info\n","lang":"Python","license":"mit","repos":"ererkka\/GDX2py"} {"commit":"30eaf18482ae1fd83eee9e71de78c9dae85b139d","old_file":"setup.py","new_file":"setup.py","old_contents":"\"\"\"Config for PyPI.\"\"\"\n\nfrom setuptools import find_packages\nfrom setuptools import setup\n\n\nsetup(\n author='Kyle P. Johnson',\n author_email='kyle@kyle-p-johnson.com',\n classifiers=[\n 'Intended Audience :: Education',\n 'Intended Audience :: Science\/Research',\n 'License :: OSI Approved :: MIT License',\n 'Natural Language :: Chinese (Traditional)',\n 'Natural Language :: English',\n 'Natural Language :: Greek',\n 'Natural Language :: Latin',\n 'Operating System :: POSIX',\n 'Programming Language :: Python :: 3.7',\n 'Topic :: Scientific\/Engineering :: Artificial Intelligence',\n 'Topic :: Text Processing',\n 'Topic :: Text Processing :: General',\n 'Topic :: Text Processing :: Linguistic',\n ],\n description='NLP for the ancient world',\n install_requires=['gitpython',\n 'nltk',\n 'python-crfsuite',\n 'pyuca',\n 'pyyaml',\n 'regex',\n 'whoosh'],\n keywords=['nlp', 'nltk', 'greek', 'latin', 'chinese', 'sanskrit', 'pali', 'tibetan', 'arabic', \"germanic\"],\n license='MIT',\n long_description='The Classical Language Toolkit (CLTK) is a framework for natural language processing for Classical languages.', # pylint: disable=C0301,\n name='cltk',\n packages=find_packages(),\n url='https:\/\/github.com\/cltk\/cltk',\n version='0.1.117',\n zip_safe=True,\n test_suite='cltk.tests.test_cltk',\n)\n","new_contents":"\"\"\"Config for PyPI.\"\"\"\n\nfrom setuptools import find_packages\nfrom setuptools import setup\n\n\nsetup(\n author='Kyle P. Johnson',\n author_email='kyle@kyle-p-johnson.com',\n classifiers=[\n 'Intended Audience :: Education',\n 'Intended Audience :: Science\/Research',\n 'License :: OSI Approved :: MIT License',\n 'Natural Language :: Chinese (Traditional)',\n 'Natural Language :: English',\n 'Natural Language :: Greek',\n 'Natural Language :: Latin',\n 'Operating System :: POSIX',\n 'Programming Language :: Python :: 3.7',\n 'Topic :: Scientific\/Engineering :: Artificial Intelligence',\n 'Topic :: Text Processing',\n 'Topic :: Text Processing :: General',\n 'Topic :: Text Processing :: Linguistic',\n ],\n description='NLP for the ancient world',\n install_requires=['gitpython',\n 'nltk',\n 'python-crfsuite',\n 'pyuca',\n 'pyyaml',\n 'regex',\n 'whoosh'],\n keywords=['nlp', 'nltk', 'greek', 'latin', 'chinese', 'sanskrit', 'pali', 'tibetan', 'arabic', \"germanic\"],\n license='MIT',\n long_description='The Classical Language Toolkit (CLTK) is a framework for natural language processing for Classical languages.', # pylint: disable=C0301,\n name='cltk',\n packages=find_packages(),\n url='https:\/\/github.com\/cltk\/cltk',\n version='0.1.118',\n zip_safe=True,\n test_suite='cltk.tests.test_cltk',\n)\n","subject":"Bump vers for Ensemble lemmatizer","message":"Bump vers for Ensemble lemmatizer\n\nFrom Patrick Burns @diyclassics https:\/\/github.com\/cltk\/cltk\/pull\/981","lang":"Python","license":"mit","repos":"kylepjohnson\/cltk,diyclassics\/cltk,D-K-E\/cltk,cltk\/cltk"} {"commit":"170317364ae1ee97bad94fb9d53187ce674b5ebb","old_file":"setup.py","new_file":"setup.py","old_contents":"from os.path import isdir, isfile, join\nfrom distutils import log\n\nfrom setuptools import setup, find_packages\n\nINSTALL_REQUIRES = [\n 'BTrees',\n 'zope.component',\n 'zodbpickle',\n 'ZODB',\n 'zope.index',\n 'zerodbext.catalog',\n 'cachetools',\n 'zc.zlibstorage',\n 'pycryptodome',\n 'flask-cors',\n 'flask',\n 'requests',\n 'jsonpickle',\n 'pyelliptic',\n 'ecdsa',\n 'zope.event',\n 'zope.lifecycleevent',\n 'ZEO',\n 'six'\n]\n\n\n# This is to avoid build errors on brand new Amazon Ubuntu instances\n# which may not have libffi-dev installed.\n#\n# If we have neither cffi nor ffi.h we fall back to pycryptodome.\n# Note that the warning is only visible if pip is run with -v.\n\ndef have_cffi():\n try:\n import cffi\n except ImportError:\n return False\n else:\n return True\n\ndef have_ffi_h():\n include_dirs = [\"\/usr\/include\", \"\/usr\/local\/include\"]\n for dir in include_dirs:\n if isdir(dir):\n if isfile(join(dir, \"ffi.h\")) or isfile(join(dir, \"ffi\", \"ffi.h\")):\n return True\n return False\n\nif have_cffi() or have_ffi_h():\n INSTALL_REQUIRES.append(\"aes256gcm-nacl\")\nelse:\n log.warn(\"warning: *** ffi.h not found - aes256gcm-nacl optimization disabled ***\")\n INSTALL_REQUIRES.append(\"pycryptodome\")\n\n\nsetup(\n name=\"zerodb\",\n version=\"0.97.2.1\",\n description=\"End-to-end encrypted database\",\n author=\"ZeroDB Inc.\",\n author_email=\"michael@zerodb.io\",\n license=\"AGPLv3\",\n url=\"http:\/\/zerodb.io\",\n packages=find_packages(),\n install_requires=INSTALL_REQUIRES,\n)\n","new_contents":"from setuptools import setup, find_packages\n\nINSTALL_REQUIRES = [\n 'BTrees',\n 'zope.component',\n 'zodbpickle',\n 'ZODB',\n 'zope.index',\n 'zerodbext.catalog',\n 'cachetools',\n 'zc.zlibstorage',\n 'pycryptodome',\n 'flask-cors',\n 'flask',\n 'requests',\n 'jsonpickle',\n 'pyelliptic',\n 'ecdsa',\n 'zope.event',\n 'zope.lifecycleevent',\n 'aes256gcm-nacl',\n 'ZEO',\n 'six'\n]\n\nsetup(\n name=\"zerodb\",\n version=\"0.97.2.1\",\n description=\"End-to-end encrypted database\",\n author=\"ZeroDB Inc.\",\n author_email=\"michael@zerodb.io\",\n license=\"AGPLv3\",\n url=\"http:\/\/zerodb.io\",\n packages=find_packages(),\n install_requires=INSTALL_REQUIRES,\n)\n","subject":"Revert \"Fall back to pycryptodome if aes256gcm-nacl cannot be built.\"","message":"Revert \"Fall back to pycryptodome if aes256gcm-nacl cannot be built.\"\n\nThis reverts commit f457b06dd1dd9d3b1d3577b5a8415868ff0cb3d8.\n","lang":"Python","license":"agpl-3.0","repos":"zerodb\/zerodb,zero-db\/zerodb,zerodb\/zerodb,zero-db\/zerodb"} {"commit":"d9771c3c0f78b55de4ae00be4553975c2f934c12","old_file":"setup.py","new_file":"setup.py","old_contents":"import os\nimport re\nimport sys\nimport platform\nimport subprocess\nimport multiprocessing\n\nfrom setuptools import setup, Extension\nfrom setuptools.command.build_ext import build_ext\nfrom distutils.version import LooseVersion\n\ncores = multiprocessing.cpu_count()*1.25\n\nthreads=\"-j\" + str(int(cores))\n\n\n\nclass CMakeExtension(Extension):\n def __init__(self, name, sourcedir=''):\n Extension.__init__(self, name, sources=[])\n self.sourcedir = os.path.abspath(sourcedir)\n\ndef readme():\n with open(\"PYTHONREADME.md\", \"r\") as fh:\n return fh.read()\n\ndef operatingsystem():\n if (platform.platform().find(\"Darwin\") >= 0):\n return \"Operating System :: MacOS\"\n else: \n return \"Operating System :: POSIX :: Linux\"\n\nsetup(\n name='sharkbite',\n version='1.2.0.0',\n author='Marc Parisi',\n author_email='phrocker@apache.org',\n url='https:\/\/docs.sharkbite.io\/',\n description='Apache Accumulo and Apache HDFS Python Connector',\n long_description=readme(),\n long_description_content_type='text\/markdown',\n ext_modules=[CMakeExtension('sharkbite.pysharkbite')],\n zip_safe=False,\n classifiers=[\n \"Programming Language :: C++\",\n \"License :: OSI Approved :: Apache Software License\",\n operatingsystem(),\n ],\n python_requires='>=3.6',\n packages=['sharkbite']\n)\n","new_contents":"import os\nimport re\nimport sys\nimport platform\nimport subprocess\nimport multiprocessing\n\nfrom setuptools import setup, Extension\nfrom setuptools.command.build_ext import build_ext\nfrom distutils.version import LooseVersion\n\ncores = multiprocessing.cpu_count()*1.25\n\nthreads=\"-j\" + str(int(cores))\n\n\n\nclass CMakeExtension(Extension):\n def __init__(self, name, sourcedir=''):\n Extension.__init__(self, name, sources=[])\n self.sourcedir = os.path.abspath(sourcedir)\n\ndef readme():\n with open(\"PYTHONREADME.md\", \"r\") as fh:\n return fh.read()\n\ndef operatingsystem():\n if (platform.platform().find(\"Darwin\") >= 0):\n return \"Operating System :: MacOS\"\n else: \n return \"Operating System :: POSIX :: Linux\"\n\nsetup(\n name='sharkbite',\n version='1.2.0.1',\n author='Marc Parisi',\n author_email='phrocker@apache.org',\n url='https:\/\/docs.sharkbite.io\/',\n description='Apache Accumulo and Apache HDFS Python Connector',\n long_description=readme(),\n long_description_content_type='text\/markdown',\n ext_modules=[CMakeExtension('sharkbite.pysharkbite')],\n zip_safe=False,\n classifiers=[\n \"Programming Language :: C++\",\n \"License :: OSI Approved :: Apache Software License\",\n operatingsystem(),\n ],\n python_requires='>=3.6',\n packages=['sharkbite']\n)\n","subject":"Update error in 3.7 release","message":"Update error in 3.7 release\n","lang":"Python","license":"apache-2.0","repos":"phrocker\/sharkbite,phrocker\/sharkbite,phrocker\/sharkbite,phrocker\/sharkbite,phrocker\/sharkbite"} {"commit":"bc308c5b7893b1853aa1de8397989fa81f9d716d","old_file":"setup.py","new_file":"setup.py","old_contents":"# Copyright (C) 2016-present the ayncpg authors and contributors\n# <see AUTHORS file>\n#\n# This module is part of asyncpg and is released under\n# the Apache 2.0 License: http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n\n\nimport setuptools\nimport sys\n\n\nif sys.version_info < (3, 5):\n raise RuntimeError('asyncpg requires Python 3.5 or greater')\n\n\nsetuptools.setup(\n name='asyncpg',\n version='0.5.1',\n description='An asyncio PosgtreSQL driver',\n classifiers=[\n 'License :: OSI Approved :: Apache Software License',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 3 :: Only',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Operating System :: POSIX',\n 'Operating System :: MacOS :: MacOS X',\n 'Development Status :: 4 - Beta',\n ],\n platforms=['POSIX'],\n author='MagicStack Inc',\n author_email='hello@magic.io',\n license='Apache License, Version 2.0',\n packages=['asyncpg'],\n provides=['asyncpg'],\n include_package_data=True,\n ext_modules=[\n setuptools.Extension(\n \"asyncpg.protocol.protocol\",\n [\"asyncpg\/protocol\/record\/recordobj.c\",\n \"asyncpg\/protocol\/protocol.c\"],\n extra_compile_args=['-O2'])\n ]\n)\n","new_contents":"# Copyright (C) 2016-present the ayncpg authors and contributors\n# <see AUTHORS file>\n#\n# This module is part of asyncpg and is released under\n# the Apache 2.0 License: http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n\n\nimport setuptools\nimport sys\n\n\nif sys.version_info < (3, 5):\n raise RuntimeError('asyncpg requires Python 3.5 or greater')\n\n\nsetuptools.setup(\n name='asyncpg',\n version='0.5.2',\n description='An asyncio PosgtreSQL driver',\n classifiers=[\n 'License :: OSI Approved :: Apache Software License',\n 'Intended Audience :: Developers',\n 'Programming Language :: Python :: 3 :: Only',\n 'Programming Language :: Python :: 3.5',\n 'Programming Language :: Python :: 3.6',\n 'Operating System :: POSIX',\n 'Operating System :: MacOS :: MacOS X',\n 'Development Status :: 4 - Beta',\n ],\n platforms=['POSIX'],\n author='MagicStack Inc',\n author_email='hello@magic.io',\n license='Apache License, Version 2.0',\n packages=['asyncpg'],\n provides=['asyncpg'],\n include_package_data=True,\n ext_modules=[\n setuptools.Extension(\n \"asyncpg.protocol.protocol\",\n [\"asyncpg\/protocol\/record\/recordobj.c\",\n \"asyncpg\/protocol\/protocol.c\"],\n extra_compile_args=['-O2'])\n ]\n)\n","subject":"Bump the version to v0.5.2","message":"Bump the version to v0.5.2\n","lang":"Python","license":"apache-2.0","repos":"MagicStack\/asyncpg,MagicStack\/asyncpg"} {"commit":"96eaeffede19d337ccd56080f9fb56b59aaf3bbb","old_file":"setup.py","new_file":"setup.py","old_contents":"import subprocess\nimport sys\nfrom distutils.core import setup, Command\n\n\nclass TestCommand(Command):\n user_options = []\n\n def initialize_options(self):\n pass\n\n def finalize_options(self):\n pass\n\n def run(self):\n errno = subprocess.call([sys.executable, 'test_facebook.py'])\n raise SystemExit(errno)\n\nsetup(\n name='facebook-ads-api',\n version='0.1.42',\n author='Chee-Hyung Yoon',\n author_email='yoonchee@gmail.com',\n py_modules=['facebook', ],\n url='http:\/\/github.com\/narrowcast\/facebook-ads-api',\n license='LICENSE',\n description='Python client for the Facebook Ads API',\n long_description=open('README.md').read(),\n cmdclass={'test': TestCommand},\n)\n","new_contents":"import subprocess\nimport sys\nfrom distutils.core import setup, Command\n\n\nclass TestCommand(Command):\n user_options = []\n\n def initialize_options(self):\n pass\n\n def finalize_options(self):\n pass\n\n def run(self):\n errno = subprocess.call([sys.executable, 'test_facebook.py'])\n raise SystemExit(errno)\n\nsetup(\n name='facebook-ads-api',\n version='0.1.43',\n author='Chee-Hyung Yoon',\n author_email='yoonchee@gmail.com',\n py_modules=['facebook', ],\n url='http:\/\/github.com\/narrowcast\/facebook-ads-api',\n license='LICENSE',\n description='Python client for the Facebook Ads API',\n long_description=open('README.md').read(),\n cmdclass={'test': TestCommand},\n)\n","subject":"Bump up version after merging a pull request.","message":"Bump up version after merging a pull request.\n","lang":"Python","license":"mit","repos":"taenyon\/facebook-ads-api,GallopLabs\/facebook-ads-api,narrowcast\/facebook-ads-api"} {"commit":"13ef6879aeca9881483bd9f575d66377f1dde0c1","old_file":"tests\/test_io.py","new_file":"tests\/test_io.py","old_contents":"import numpy as np\nfrom tempfile import NamedTemporaryFile\nfrom microscopium import io as mio\n\ndef test_imsave_tif_compress():\n im = np.random.randint(0, 256, size=(1024, 1024, 3)).astype(np.uint8)\n with NamedTemporaryFile(suffix='.tif') as fout:\n fname = fout.name\n fout.close()\n mio.imsave(im, fname, compress=2)\n imin = mio.imread(fname)\n np.testing.assert_array_equal(im, imin)\n","new_contents":"import os\nimport numpy as np\nfrom tempfile import NamedTemporaryFile\nfrom microscopium import io as mio\nfrom microscopium import pathutils as pth\n\n\ndef test_recursive_glob():\n abspath = os.path.dirname(__file__)\n tiffs0 = pth.all_matching_files(abspath, '*.tif')\n assert len(tiffs0) == 8\n assert tiffs0[0].startswith('\/')\n tiffs1 = pth.all_matching_files(abspath, '*.TIF')\n assert len(tiffs1) == 0\n tiffs2 = pth.all_matching_files(abspath, '*.TIF', case_sensitive=False,\n full=False)\n assert len(tiffs2) == 8\n assert tiffs2[0].startswith('MYORES')\n\n\ndef test_imsave_tif_compress():\n im = np.random.randint(0, 256, size=(1024, 1024, 3)).astype(np.uint8)\n with NamedTemporaryFile(suffix='.tif') as fout:\n fname = fout.name\n fout.close()\n mio.imsave(im, fname, compress=2)\n imin = mio.imread(fname)\n np.testing.assert_array_equal(im, imin)\n","subject":"Improve test coverage by testing recursive glob","message":"Improve test coverage by testing recursive glob\n","lang":"Python","license":"bsd-3-clause","repos":"jni\/microscopium,microscopium\/microscopium,Don86\/microscopium,microscopium\/microscopium,jni\/microscopium,Don86\/microscopium"} {"commit":"cddb15f25df404d849601e7c8db2df15d2443958","old_file":"intellipaste.py","new_file":"intellipaste.py","old_contents":"#!\/usr\/bin\/env python3\n\nimport os\nimport time\nimport json\nimport requests\nimport pyperclip\n\nAPI_KEY = os.environ.get('API_KEY')\n# API_KEY = \"\"\n\n\ndef google_url_shorten(url):\n req_url = \"https:\/\/www.googleapis.com\/urlshortener\/v1\/url?\" + API_KEY\n payload = {'longUrl': url}\n headers = {'content-type': 'application\/json'}\n r = requests.post(req_url, data=json.dumps(payload), headers=headers)\n resp = json.loads(r.text)\n return resp[\"id\"]\n\n\nrecent_value = \"\"\nwhile True:\n tmp_value = pyperclip.paste()\n if (tmp_value != recent_value and not tmp_value.startswith(\"https:\/\/goo.gl\") and not tmp_value.startswith(\"https:\/\/git\")):\n recent_value = tmp_value\n url = str(recent_value)\n if url.startswith(\"http:\/\/\") or url.startswith(\"https:\/\/\") or url.startswith(\"www.\"):\n pyperclip.copy(google_url_shorten(url))\n time.sleep(0.5)\n","new_contents":"#!\/usr\/bin\/env python3\n\nimport os\nimport time\nimport json\nimport requests\nimport pyperclip\n\nAPI_KEY = os.environ.get('API_KEY')\n# API_KEY = \"\"\n\n\ndef google_url_shorten(url):\n req_url = \"https:\/\/www.googleapis.com\/urlshortener\/v1\/url?\" + str(API_KEY)\n payload = {'longUrl': url}\n headers = {'content-type': 'application\/json'}\n r = requests.post(req_url, data=json.dumps(payload), headers=headers)\n resp = json.loads(r.text)\n return resp[\"id\"]\n\n\nrecent_value = \"\"\nwhile True:\n tmp_value = pyperclip.paste()\n if (tmp_value != recent_value and not tmp_value.startswith(\"https:\/\/goo.gl\") and not tmp_value.startswith(\"https:\/\/git\")):\n recent_value = tmp_value\n url = str(recent_value)\n if url.startswith(\"http:\/\/\") or url.startswith(\"https:\/\/\") or url.startswith(\"www.\"):\n pyperclip.copy(google_url_shorten(url))\n time.sleep(0.5)\n","subject":"Convert environment variable to string","message":"Convert environment variable to string","lang":"Python","license":"apache-2.0","repos":"LucasHMS\/intelliPaste"} {"commit":"11c1f5a9806f4a21abdb5ac7e4310ef3242389b0","old_file":"server_dev.py","new_file":"server_dev.py","old_contents":"from projects_controller import ProjectsController\nfrom redirects_controller import RedirectsController\nfrom flask import Flask, render_template, redirect, abort\n\n\nDATA_DIR = 'data'\n\napp = Flask(__name__)\napp.url_map.strict_slashes = False\n\nprojects_controller = ProjectsController(DATA_DIR)\nredirects_controller = RedirectsController(DATA_DIR)\n\n\n@app.errorhandler(404)\ndef page_not_found(e):\n return render_template('404.html'), 404\n\n@app.route('\/')\ndef index():\n projects = projects_controller.get_current_projects()\n return render_template('index.html', projects=projects)\n\n@app.route('\/start')\ndef start_project():\n return render_template('start_project.html')\n\n@app.route('\/<dynamic>')\ndef project(dynamic):\n\n # First, test if if it's a project\n projects = projects_controller.get_all_projects()\n redirects = redirects_controller.get_redirects()\n\n if dynamic in projects:\n project_data = projects[dynamic]\n if 'conclusion_post' in project_data:\n # The project is over, we should redirect to the post\n return redirect(project_data['conclusion_post'])\n else:\n return render_template('project.html', project_data=project_data)\n\n # Next, check if it's a redirect\n elif dynamic in redirects:\n return redirect(redirects[dynamic])\n\n else:\n abort(404)\n\n\nif __name__ == '__main__':\n app.run(debug=True)\n","new_contents":"from projects_controller import ProjectsController\nfrom redirects_controller import RedirectsController\nfrom flask import Flask, render_template, redirect, abort\n\n\nDATA_DIR = 'data'\n\napp = Flask(__name__)\napp.url_map.strict_slashes = False\n\nprojects_controller = ProjectsController(DATA_DIR)\nredirects_controller = RedirectsController(DATA_DIR)\n\n\n@app.errorhandler(404)\ndef page_not_found(e):\n return render_template('404.html'), 404\n\n@app.route('\/')\ndef index():\n projects = projects_controller.get_current_projects()\n return render_template('index.html', projects=projects)\n\n@app.route('\/start')\ndef start_project():\n return render_template('start_project.html')\n\n@app.route('\/<dynamic>')\ndef project(dynamic):\n\n projects = projects_controller.get_all_projects()\n redirects = redirects_controller.get_redirects()\n\n # First, test if if it's a project\n if dynamic in projects:\n project_data = projects[dynamic]\n if 'conclusion_post' in project_data:\n # The project is over, we should redirect to the post\n return redirect(project_data['conclusion_post'])\n else:\n return render_template('project.html', project_data=project_data)\n\n # Next, check if it's a redirect\n elif dynamic in redirects:\n return redirect(redirects[dynamic])\n\n else:\n abort(404)\n\n\nif __name__ == '__main__':\n app.run(debug=True)\n","subject":"Move comment to rational location","message":"Move comment to rational location\n","lang":"Python","license":"mit","repos":"teslaworksumn\/teslaworks.net,teslaworksumn\/teslaworks.net"} {"commit":"97edbee5813b8a87606b8fb3d09b4f116cdaf025","old_file":"mordecai\/tests\/conftest.py","new_file":"mordecai\/tests\/conftest.py","old_contents":"from ..geoparse import Geoparser\nimport pytest\n\nimport spacy\nnlp = spacy.load('en_core_web_lg', disable=['parser', 'tagger'])\n\n@pytest.fixture(scope='session', autouse=True)\ndef geo():\n return Geoparser(nlp=nlp, threads=False, models_path = \"\/Users\/ahalterman\/MIT\/Geolocation\/mordecai_new\/mordecai\/mordecai\/models\/\")\n\n@pytest.fixture(scope='session', autouse=True)\ndef geo_thread():\n return Geoparser(nlp=nlp, threads=True, models_path = \"\/Users\/ahalterman\/MIT\/Geolocation\/mordecai_new\/mordecai\/mordecai\/models\/\")\n","new_contents":"from ..geoparse import Geoparser\nimport pytest\n\nimport spacy\nnlp = spacy.load('en_core_web_lg', disable=['parser', 'tagger'])\n\n@pytest.fixture(scope='session', autouse=True)\ndef geo():\n return Geoparser(nlp=nlp, threads=False)\n\n@pytest.fixture(scope='session', autouse=True)\ndef geo_thread():\n return Geoparser(nlp=nlp, threads=True)\n","subject":"Remove hardcoded paths from testing","message":"Remove hardcoded paths from testing\n","lang":"Python","license":"mit","repos":"openeventdata\/mordecai"} {"commit":"37aa1c9f8faeefe7305cca526a7424a349939add","old_file":"tests\/smoke_test.py","new_file":"tests\/smoke_test.py","old_contents":"# -*- coding: utf-8 -*-\n\nimport unittest\nimport sys\nsys.path.insert(0, '..\/mafia')\n\nfrom game import Game\nfrom game import Player\n\nfrom testclient.testmessenger import TestMessenger\n\nclass SmokeTest(unittest.TestCase):\n def setUp(self):\n self.messenger = TestMessenger()\n \n def test_smoke_test(self):\n game = Game('t,c,c,m', self.messenger)\n player_one = Player('one', 'one')\n player_two = Player('two', 'two')\n player_three = Player('three', 'three')\n player_four = Player('four', 'four')\n game.join(player_one)\n game.join(player_two)\n game.join(player_three)\n game.join(player_four)\n game.vote('one', 'three')\n game.vote('three', 'one')\n game.vote('two', 'three')\n game.vote('four', 'three')\n game.target('one', 'two')\n game.target('two', 'one')\n game.target('four', 'one')\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"# -*- coding: utf-8 -*-\n\nimport unittest\nimport sys\nsys.path.insert(0, '..\/')\n\nfrom mafia.game import Game\nfrom mafia.game import Player\n\nfrom testclient.testmessenger import TestMessenger\n\nclass SmokeTest(unittest.TestCase):\n def setUp(self):\n self.messenger = TestMessenger()\n \n def test_smoke_test(self):\n game = Game('t,c,c,m', self.messenger)\n player_one = Player('one', 'one')\n player_two = Player('two', 'two')\n player_three = Player('three', 'three')\n player_four = Player('four', 'four')\n game.join(player_one)\n game.join(player_two)\n game.join(player_three)\n game.join(player_four)\n game.vote('one', 'three')\n game.vote('three', 'one')\n game.vote('two', 'three')\n game.vote('four', 'three')\n game.target('one', 'two')\n game.target('two', 'one')\n game.target('four', 'one')\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Change the smoke test imports to a relative import for consistency.","message":"Change the smoke test imports to a relative import for consistency.\n","lang":"Python","license":"mit","repos":"BitokuOokami\/PloungeMafiaToolkit"} {"commit":"11d19d1756f6227db894aabcf4bd02e327e292c7","old_file":"tests\/test_basic.py","new_file":"tests\/test_basic.py","old_contents":"from hello_world import hello_world\n\nfrom unittest import TestCase\n\n\nclass BasicTest(TestCase):\n def test_basic_hello_world(self):\n \"\"\"\n Test basic hello world messaging\n \"\"\"\n False\n","new_contents":"from hello_world import hello_world\n\nfrom unittest import TestCase\n\n\nclass BasicTest(TestCase):\n def test_basic_hello_world(self):\n \"\"\"\n Test basic hello world messaging\n \"\"\"\n self.assertTrue(callable(hello_world))\n","subject":"Make things a little better","message":"Make things a little better\n","lang":"Python","license":"mit","repos":"jeansaad\/hello_world"} {"commit":"f23d90d441194d270b3bcf7997550f2f8e7e4c1d","old_file":"sample-functions\/BaseFunctions\/python\/handler.py","new_file":"sample-functions\/BaseFunctions\/python\/handler.py","old_contents":"import sys\n\ndef get_stdin():\n buf = \"\"\n for line in sys.stdin:\n buf = buf + line\n return buf\n\nif(__name__ == \"__main__\"):\n st = get_stdin()\n print(st)\n","new_contents":"import sys\n\ndef get_stdin():\n buf = \"\"\n for line in sys.stdin:\n buf = buf + line\n return buf\n\nif __name__ == \"__main__\":\n st = get_stdin()\n print(st)\n","subject":"Remove braces for Python sample","message":"Remove braces for Python sample\n\nSuggested by community member in an un-merged PR.\n\nSigned-off-by: Alex Ellis (VMware) <5f7133baa0f5c7ca63ff11e11f2e2b0d2cf077c8@gmail.com>\n","lang":"Python","license":"mit","repos":"openfaas\/faas,rgee0\/faas,alexellis\/faas,openfaas\/faas,alexellis\/faas,rgee0\/faas,alexellis\/faas,alexellis\/faas,rgee0\/faas,rgee0\/faas,openfaas\/faas,rgee0\/faas,rgee0\/faas,rgee0\/faas,alexellis\/faas,rgee0\/faas,alexellis\/faas,rgee0\/faas,alexellis\/faas,alexellis\/faas,alexellis\/faas,rgee0\/faas"} {"commit":"acd92d6a9e8c710657a4bcf1c46076f9d06f3d46","old_file":"test_results\/plot_all.py","new_file":"test_results\/plot_all.py","old_contents":"import glob\nimport csv\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nfor file in glob.glob(\"*.csv\"):\n data = np.genfromtxt(file, delimiter = ',', names = True)\n\n plt.figure(figsize=(10,20))\n plt.suptitle(file)\n\n num_plots = len(data.dtype.names)\n count = 1\n for col_name in data.dtype.names:\n plt.subplot(num_plots, 1, count)\n plt.plot(data[col_name], label=col_name)\n plt.legend()\n count += 1\n\n ymin, ymax = plt.ylim()\n if ymin < 0 < ymax:\n plt.axhline(0, hold=True, color = 'grey') # plot line through zero\n\n mng = plt.get_current_fig_manager()\n if plt.get_backend() == 'TkAgg':\n mng.window.state('zoomed')\n elif plt.get_backend() == 'wxAgg':\n mng.frame.Maximize(True)\n elif plt.get_backend() == 'QT4Agg':\n mng.window.showMaximized()\n\n plt.savefig(file.rstrip('.csv') + '.pdf')\n plt.show()\n\n\n\n","new_contents":"import glob\nimport numpy as np\nimport matplotlib.pyplot as plt\nimport matplotlib.backends.backend_pdf\npdf = matplotlib.backends.backend_pdf.PdfPages(\"test-results-plots.pdf\")\n\nfor file in glob.glob(\"*.csv\"):\n data = np.genfromtxt(file, delimiter = ',', names = True)\n\n plt.figure(figsize=(10,20))\n plt.suptitle(file)\n\n num_plots = len(data.dtype.names)\n count = 1\n for col_name in data.dtype.names:\n plt.subplot(num_plots, 1, count)\n plt.plot(data[col_name], label=col_name)\n plt.legend()\n count += 1\n\n ymin, ymax = plt.ylim()\n if ymin < 0 < ymax:\n plt.axhline(0, hold=True, color = 'grey') # plot line through zero\n\n pdf.savefig()\n\n mng = plt.get_current_fig_manager()\n if plt.get_backend() == 'TkAgg':\n mng.window.state('zoomed')\n elif plt.get_backend() == 'wxAgg':\n mng.frame.Maximize(True)\n elif plt.get_backend() == 'QT4Agg':\n mng.window.showMaximized()\n\n\n plt.show()\n plt.close()\n\npdf.close()\n\n\n\n\n","subject":"Save all simulation plots to one PDF instead of multiple","message":"Save all simulation plots to one PDF instead of multiple\n","lang":"Python","license":"agpl-3.0","repos":"BrewPi\/firmware,glibersat\/firmware,glibersat\/firmware,BrewPi\/firmware,glibersat\/firmware,etk29321\/firmware,etk29321\/firmware,etk29321\/firmware,etk29321\/firmware,glibersat\/firmware,glibersat\/firmware,BrewPi\/firmware,BrewPi\/firmware,BrewPi\/firmware,BrewPi\/firmware,BrewPi\/firmware,etk29321\/firmware,BrewPi\/firmware,etk29321\/firmware,glibersat\/firmware,glibersat\/firmware"} {"commit":"cc8f5b35d4c227f82b2872d5bfad24bef37209e5","old_file":"overtime_calculator\/__main__.py","new_file":"overtime_calculator\/__main__.py","old_contents":"import hug\n\nfrom overtime_calculator import auth, api\n\n\n@hug.get(\"\/\", output=hug.output_format.html)\ndef base():\n return \"<h1>Hello, world<\/h1>\"\n\n\n@hug.extend_api()\ndef with_other_apis():\n return [\n auth,\n api\n ]\n","new_contents":"import sys\nimport pathlib\n\nimport hug\n\nfrom overtime_calculator import auth, api\n\n\n@hug.get(\"\/\", output=hug.output_format.html)\ndef base():\n return \"<h1>Hello, world<\/h1>\"\n\n\n@hug.extend_api()\ndef with_other_apis():\n return [\n auth,\n api\n ]\n\nif __name__ == '__main__':\n _file = pathlib.Path(sys.argv[0])\n module = _file.parent.name\n print(\n f\"Start {module} with Hug, like so: hug --file {_file}\",\n file=sys.stderr,\n )\n sys.exit(1)\n","subject":"Improve UX for those who use CLI","message":"Enhancement: Improve UX for those who use CLI\n","lang":"Python","license":"mit","repos":"x10an14\/overtime-calculator"} {"commit":"f3c99d8a8a9d485ebc9a18419a142f03d4730fba","old_file":"examples\/guv_simple_http_response.py","new_file":"examples\/guv_simple_http_response.py","old_contents":"# FIXME: pyuv_cffi needs to build the library BEFORE the standard library is patched\nimport pyuv_cffi\n\nprint('pyuv_cffi imported', pyuv_cffi)\nimport guv\n\nguv.monkey_patch()\nimport guv.server\nimport logging\nimport time\n\nfrom util import create_example\nimport logger\n\nif not hasattr(time, 'perf_counter'):\n time.perf_counter = time.clock\n\nlogger.configure()\nlog = logging.getLogger()\n\nresponse_times = []\n\n\ndef get_avg_time():\n global response_times\n times = response_times[-1000:]\n avg = sum(times) \/ len(times)\n\n if len(response_times) > 5000:\n response_times = times\n\n return avg\n\n\ndef handle(sock, addr):\n # client connected\n start_time = time.perf_counter()\n sock.sendall(create_example())\n sock.close()\n\n total_time = time.perf_counter() - start_time\n response_times.append(total_time)\n\n\nif __name__ == '__main__':\n pool = guv.GreenPool()\n\n try:\n log.debug('Start')\n server_sock = guv.listen(('0.0.0.0', 8001))\n server = guv.server.Server(server_sock, handle, pool, 'spawn_n')\n server.start()\n except (SystemExit, KeyboardInterrupt):\n log.debug('average response time: {}'.format(get_avg_time()))\n log.debug('Bye!')\n","new_contents":"import guv\n\nguv.monkey_patch()\nimport guv.server\nimport logging\nimport time\n\nfrom util import create_example\nimport logger\nfrom pympler import tracker\n\ntr = tracker.SummaryTracker()\n\nif not hasattr(time, 'perf_counter'):\n time.perf_counter = time.clock\n\nlogger.configure()\nlog = logging.getLogger()\n\nresponse_times = []\n\n\ndef get_avg_time():\n global response_times\n times = response_times[-1000:]\n avg = sum(times) \/ len(times)\n\n if len(response_times) > 5000:\n response_times = times\n\n return avg\n\n\ndef handle(sock, addr):\n # client connected\n start_time = time.perf_counter()\n sock.sendall(create_example())\n sock.close()\n\n total_time = time.perf_counter() - start_time\n response_times.append(total_time)\n\n\nif __name__ == '__main__':\n pool = guv.GreenPool()\n\n try:\n log.debug('Start')\n server_sock = guv.listen(('0.0.0.0', 8001))\n server = guv.server.Server(server_sock, handle, pool, 'spawn_n')\n server.start()\n except (SystemExit, KeyboardInterrupt):\n tr.print_diff()\n log.debug('Bye!')\n","subject":"Use pympler to look for memory leaks","message":"Use pympler to look for memory leaks\n","lang":"Python","license":"mit","repos":"veegee\/guv,veegee\/guv"} {"commit":"e68d0b269b5c632bb96cdc04f37b622b15a0382e","old_file":"dataproc\/dataproc_e2e_test.py","new_file":"dataproc\/dataproc_e2e_test.py","old_contents":"# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\" Integration tests for Dataproc samples.\n\nCreates a Dataproc cluster, uploads a pyspark file to Google Cloud Storage,\nsubmits a job to Dataproc that runs the pyspark file, then downloads\nthe output logs from Cloud Storage and verifies the expected output.\"\"\"\n\nimport create_cluster_and_submit_job\nfrom gcp.testing.flaky import flaky\n\nCLUSTER_NAME = 'testcluster2'\nZONE = 'us-central1-b'\n\n\n@flaky\ndef test_e2e(cloud_config):\n output = create_cluster_and_submit_job.main(\n cloud_config.project, ZONE, CLUSTER_NAME, cloud_config.storage_bucket)\n assert \"['Hello,', 'dog', 'elephant', 'panther', 'world!']\" in output\n","new_contents":"# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\" Integration tests for Dataproc samples.\n\nCreates a Dataproc cluster, uploads a pyspark file to Google Cloud Storage,\nsubmits a job to Dataproc that runs the pyspark file, then downloads\nthe output logs from Cloud Storage and verifies the expected output.\"\"\"\n\nimport create_cluster_and_submit_job\nfrom gcp.testing.flaky import flaky\n\nCLUSTER_NAME = 'testcluster2'\nZONE = 'us-central1-b'\n\n\n@flaky\ndef test_e2e(cloud_config):\n output = create_cluster_and_submit_job.main(\n cloud_config.project, ZONE, CLUSTER_NAME, cloud_config.storage_bucket)\n assert b\"['Hello,', 'dog', 'elephant', 'panther', 'world!']\" in output\n","subject":"Fix Dataproc e2e for Python 3","message":"Fix Dataproc e2e for Python 3","lang":"Python","license":"apache-2.0","repos":"JavaRabbit\/CS496_capstone,sharbison3\/python-docs-samples,hashems\/Mobile-Cloud-Development-Projects,sharbison3\/python-docs-samples,GoogleCloudPlatform\/python-docs-samples,BrandonY\/python-docs-samples,BrandonY\/python-docs-samples,BrandonY\/python-docs-samples,sharbison3\/python-docs-samples,sharbison3\/python-docs-samples,hashems\/Mobile-Cloud-Development-Projects,JavaRabbit\/CS496_capstone,JavaRabbit\/CS496_capstone,GoogleCloudPlatform\/python-docs-samples,JavaRabbit\/CS496_capstone,canglade\/NLP,canglade\/NLP,GoogleCloudPlatform\/python-docs-samples,BrandonY\/python-docs-samples,hashems\/Mobile-Cloud-Development-Projects,hashems\/Mobile-Cloud-Development-Projects,canglade\/NLP,canglade\/NLP,GoogleCloudPlatform\/python-docs-samples"} {"commit":"b0b232297f55cd38db85bb2ec5b30a6022a3f4d1","old_file":"tweepy\/asynchronous\/__init__.py","new_file":"tweepy\/asynchronous\/__init__.py","old_contents":"# Tweepy\n# Copyright 2009-2021 Joshua Roesslein\n# See LICENSE for details.\n\n\"\"\"\nTweepy.asynchronoous\n\nAsynchronous interfaces with the Twitter API\n\"\"\"\n\ntry:\n import aiohttp\nexcept ModuleNotFoundError:\n from tweepy.errors import TweepyException\n raise TweepyException(\"tweepy.asynchronous requires aiohttp to be installed\")\n\nfrom tweepy.asynchronous.streaming import AsyncStream\n","new_contents":"# Tweepy\n# Copyright 2009-2021 Joshua Roesslein\n# See LICENSE for details.\n\n\"\"\"\nTweepy.asynchronoous\n\nAsynchronous interfaces with the Twitter API\n\"\"\"\n\ntry:\n import aiohttp\n import oauthlib\nexcept ModuleNotFoundError:\n from tweepy.errors import TweepyException\n raise TweepyException(\n \"tweepy.asynchronous requires aiohttp and oauthlib to be installed\"\n )\n\nfrom tweepy.asynchronous.streaming import AsyncStream\n","subject":"Check oauthlib installation when importing asynchronous subpackage","message":"Check oauthlib installation when importing asynchronous subpackage\n\n","lang":"Python","license":"mit","repos":"svven\/tweepy,tweepy\/tweepy"} {"commit":"f5f850e53a889a5afe483ae2ca07e147d4a94c08","old_file":"tests.py","new_file":"tests.py","old_contents":"#!\/usr\/bin\/env python\n# encoding: utf-8\nimport datetime\nimport unittest\n\nimport mock\n\nfrom nose.tools import assert_equal, assert_is_instance\n\nfrom pandas_finance import get_stock\n\nclass PandasFinanceTestCase(unittest.TestCase):\n @mock.patch('pandas_finance.web.DataReader')\n def test_get_stock_called_correctly(self, mock_datareader):\n mock_datareader()\n start = datetime.datetime(1999, 4, 3, 0, 0)\n end = datetime.datetime(2005, 2, 5, 0, 0)\n get_stock('AAPL', start, end)\n mock_datareader.assert_called_with('AAPL', 'yahoo', start, end)\n","new_contents":"#!\/usr\/bin\/env python\n# encoding: utf-8\nimport datetime\nimport unittest\n\nimport mock\n\nfrom nose.tools import assert_equal, assert_is_instance\nfrom pandas_finance import get_stock, get_required_tickers\n\nclass PandasFinanceTestCase(unittest.TestCase):\n @mock.patch('pandas_finance.web.DataReader')\n def test_get_stock_called_correctly(self, mock_datareader):\n mock_datareader()\n start = datetime.datetime(1999, 4, 3, 0, 0)\n end = datetime.datetime(2005, 2, 5, 0, 0)\n get_stock('AAPL', start, end)\n mock_datareader.assert_called_with('AAPL', 'yahoo', start, end)\n\n def test_get_required_tickers_parses_tickers_with_newline(self):\n m = mock.mock_open(read_data='TWTR,FB,AAPL,MSFT\\n')\n textfile = None # only used to provide valid argument\n with mock.patch('pandas_finance.open', m, create=True):\n result = get_required_tickers(textfile)\n assert_equal('TWTR,FB,AAPL,MSFT', result)\n","subject":"Add test for parsing tickers.","message":"Add test for parsing tickers.\n","lang":"Python","license":"agpl-3.0","repos":"scraperwiki\/stock-tool,scraperwiki\/stock-tool"} {"commit":"019f259ae42a95802dce644511399332506ad1cc","old_file":"tracing\/tracing\/metrics\/metric_runner.py","new_file":"tracing\/tracing\/metrics\/metric_runner.py","old_contents":"# Copyright 2016 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\nimport os\n\nfrom perf_insights import map_single_trace\nfrom perf_insights import function_handle\nfrom perf_insights.mre import file_handle\nfrom perf_insights.mre import job as job_module\n\n_METRIC_MAP_FUNCTION_FILENAME = 'metric_map_function.html'\n\n_METRIC_MAP_FUNCTION_NAME = 'metricMapFunction'\n\ndef _GetMetricsDir():\n return os.path.dirname(os.path.abspath(__file__))\n\ndef _GetMetricRunnerHandle(metric):\n assert isinstance(metric, basestring)\n metrics_dir = _GetMetricsDir()\n metric_mapper_path = os.path.join(metrics_dir, _METRIC_MAP_FUNCTION_FILENAME)\n\n modules_to_load = [function_handle.ModuleToLoad(filename=metric_mapper_path)]\n map_function_handle = function_handle.FunctionHandle(\n modules_to_load, _METRIC_MAP_FUNCTION_NAME, {'metric': metric})\n\n return job_module.Job(map_function_handle, None)\n\ndef RunMetric(filename, metric, extra_import_options=None):\n th = file_handle.URLFileHandle(filename, 'file:\/\/' + filename)\n result = map_single_trace.MapSingleTrace(\n th, _GetMetricRunnerHandle(metric), extra_import_options)\n\n return result\n","new_contents":"# Copyright 2016 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\nimport os\n\nfrom perf_insights import map_single_trace\nfrom perf_insights import function_handle\nfrom perf_insights.mre import file_handle\nfrom perf_insights.mre import job as job_module\n\n_METRIC_MAP_FUNCTION_FILENAME = 'metric_map_function.html'\n\n_METRIC_MAP_FUNCTION_NAME = 'metricMapFunction'\n\ndef _GetMetricsDir():\n return os.path.dirname(os.path.abspath(__file__))\n\ndef _GetMetricRunnerHandle(metric):\n assert isinstance(metric, basestring)\n metrics_dir = _GetMetricsDir()\n metric_mapper_path = os.path.join(metrics_dir, _METRIC_MAP_FUNCTION_FILENAME)\n\n modules_to_load = [function_handle.ModuleToLoad(filename=metric_mapper_path)]\n map_function_handle = function_handle.FunctionHandle(\n modules_to_load, _METRIC_MAP_FUNCTION_NAME, {'metric': metric})\n\n return job_module.Job(map_function_handle, None)\n\ndef RunMetric(filename, metric, extra_import_options=None):\n url = 'file:\/\/' + os.path.abspath(filename)\n th = file_handle.URLFileHandle(filename, url)\n result = map_single_trace.MapSingleTrace(\n th, _GetMetricRunnerHandle(metric), extra_import_options)\n\n return result\n","subject":"Support relative paths in bin\/run_metric","message":"Support relative paths in bin\/run_metric\n\nAs a result of this patch, it will be possible to run:\n\n bin\/run_metric MemoryMetric test_data\/memory_dumps.json\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\ninstead of:\n\n bin\/run_metric MemoryMetric $PWD\/test_data\/memory_dumps.json\n ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^\n\nReview URL: https:\/\/codereview.chromium.org\/1836283008\n","lang":"Python","license":"bsd-3-clause","repos":"catapult-project\/catapult,sahiljain\/catapult,benschmaus\/catapult,catapult-project\/catapult-csm,sahiljain\/catapult,benschmaus\/catapult,benschmaus\/catapult,catapult-project\/catapult,catapult-project\/catapult,catapult-project\/catapult,catapult-project\/catapult,catapult-project\/catapult-csm,catapult-project\/catapult-csm,sahiljain\/catapult,benschmaus\/catapult,catapult-project\/catapult-csm,sahiljain\/catapult,benschmaus\/catapult,catapult-project\/catapult-csm,sahiljain\/catapult,catapult-project\/catapult-csm,benschmaus\/catapult,sahiljain\/catapult,catapult-project\/catapult,catapult-project\/catapult-csm,catapult-project\/catapult,benschmaus\/catapult"} {"commit":"611aee2e704ffbad8579e5005ca36232097f96c5","old_file":"bot\/utils.py","new_file":"bot\/utils.py","old_contents":"from enum import IntEnum\nfrom discord import Embed\n\n\nclass OpStatus(IntEnum):\n\tSUCCESS = 0x2ECC71,\n\tFAILURE = 0xc0392B,\n\tWARNING = 0xf39C12,\n\tNONE = None\n\n\ndef build_embed(ctx, desc: str, title: str = '', status: OpStatus = OpStatus.SUCCESS) -> Embed:\n\tname = ctx.message.server.me.nick if ctx.message.server.me.nick is not None else ctx.bot.user.name\n\tembed = Embed(\n\t\ttitle=title,\n\t\tdescription=desc,\n\t\tcolor=status.value if status is not None else OpStatus.WARNING\n\t)\n\tembed.set_author(name=name, icon_url=ctx.bot.user.avatar_url)\n\treturn embed\n","new_contents":"from enum import IntEnum\nfrom discord import Embed\n\n\nclass OpStatus(IntEnum):\n\tSUCCESS = 0x2ECC71,\n\tFAILURE = 0xc0392B,\n\tWARNING = 0xf39C12,\n\tNONE = -1\n\n\ndef build_embed(ctx, desc: str, title: str = '', status: OpStatus = OpStatus.SUCCESS) -> Embed:\n\tname = ctx.message.server.me.nick if ctx.message.server.me.nick is not None else ctx.bot.user.name\n\tembed = Embed(\n\t\ttitle=title,\n\t\tdescription=desc,\n\t\tcolor=status.value if status is not None and status is not -1 else None if status is -1 else OpStatus.WARNING\n\t)\n\tembed.set_author(name=name, icon_url=ctx.bot.user.avatar_url)\n\treturn embed\n","subject":"Fix issue with enum not accepting null","message":"Fix issue with enum not accepting null\n\n","lang":"Python","license":"apache-2.0","repos":"HellPie\/discord-reply-bot"} {"commit":"b3466fc14e9616c620258eea382b644ac2585845","old_file":"rest\/urls.py","new_file":"rest\/urls.py","old_contents":"# Author: Braedy Kuzma\nfrom django.conf.urls import url\nfrom . import views\n\nurlpatterns = [\n url(r'^posts\/$', views.PostsView.as_view(), name='posts'),\n url(r'^posts\/(?P<pid>[0-9a-fA-F\\-]+)\/$', views.PostView.as_view(),\n name='post'),\n url(r'^posts\/(?P<pid>[0-9a-fA-F\\-]+)\/comments\/$',\n views.CommentView.as_view(), name='comments'),\n url(r'^author\/(?P<aid>[0-9a-fA-F\\-]+)\/$', views.AuthorView.as_view(),\n name='author'),\n url(r'^author\/(?P<aid>[0-9a-fA-F\\-]+)\/friends\/$',\n views.AuthorFriendsView.as_view(), name='friends')\n]\n","new_contents":"# Author: Braedy Kuzma\nfrom django.conf.urls import url\nfrom . import views\n\nurlpatterns = [\n url(r'^posts\/$', views.PostsView.as_view(), name='posts'),\n url(r'^posts\/(?P<pid>[0-9a-fA-F\\-]+)\/$', views.PostView.as_view(),\n name='post'),\n url(r'^posts\/(?P<pid>[0-9a-fA-F\\-]+)\/comments\/$',\n views.CommentView.as_view(), name='comments'),\n url(r'^author\/posts\/$', views.PostView.as_view(), name='authorpost'),\n url(r'^author\/(?P<aid>[0-9a-fA-F\\-]+)\/$', views.AuthorView.as_view(),\n name='author'),\n url(r'^author\/(?P<aid>[0-9a-fA-F\\-]+)\/friends\/$',\n views.AuthorFriendsView.as_view(), name='friends')\n]\n","subject":"Add extra endpoint for posts?","message":"Add extra endpoint for posts?\n","lang":"Python","license":"apache-2.0","repos":"CMPUT404W17T06\/CMPUT404-project,CMPUT404W17T06\/CMPUT404-project,CMPUT404W17T06\/CMPUT404-project"} {"commit":"ab7ce76c47ea99080c105feb1a4f8aed39554597","old_file":"doc\/examples\/example_world.py","new_file":"doc\/examples\/example_world.py","old_contents":"\nfrom __future__ import unicode_literals\n\nfrom imaginary.world import ImaginaryWorld\nfrom imaginary.objects import Thing, Container, Exit\nfrom imaginary.garments import createShirt, createPants\nfrom imaginary.iimaginary import IClothing, IClothingWearer\n\nfrom examplegame.squeaky import Squeaker\n\n\ndef world(store):\n def room(name):\n it = Thing(store=store, name=name)\n Container.createFor(it, capacity=1000)\n return it\n world = ImaginaryWorld(store=store,\n origin=room(\"The Beginning\"))\n protagonist = world.create(\"An Example Player\")\n shirt = createShirt(store=store, name=\"shirt\", location=world.origin)\n pants = createPants(store=store, name=\"pants\", location=world.origin)\n middle = room(\"The Middle\")\n wearer = IClothingWearer(protagonist)\n wearer.putOn(IClothing(shirt))\n wearer.putOn(IClothing(pants))\n Exit.link(world.origin, middle, \"north\")\n\n squeakerThing = Thing(name=\"squeaker\", location=middle, store=store)\n Squeaker.createFor(squeakerThing)\n return world\n","new_contents":"\nfrom __future__ import unicode_literals\n\nfrom imaginary.world import ImaginaryWorld\nfrom imaginary.objects import Thing, Container, Exit\nfrom imaginary.garments import createShirt, createPants\nfrom imaginary.iimaginary import IClothing, IClothingWearer\n\nfrom examplegame.squeaky import Squeaker\n\n\ndef world(store):\n def room(name):\n it = Thing(\n store=store,\n name=name,\n proper=True,\n )\n Container.createFor(it, capacity=1000)\n return it\n world = ImaginaryWorld(store=store,\n origin=room(\"The Beginning\"))\n protagonist = world.create(\"An Example Player\")\n shirt = createShirt(store=store, name=\"shirt\", location=world.origin)\n pants = createPants(store=store, name=\"pants\", location=world.origin)\n middle = room(\"The Middle\")\n wearer = IClothingWearer(protagonist)\n wearer.putOn(IClothing(shirt))\n wearer.putOn(IClothing(pants))\n Exit.link(world.origin, middle, \"north\")\n\n squeakerThing = Thing(name=\"squeaker\", location=middle, store=store)\n Squeaker.createFor(squeakerThing)\n return world\n","subject":"Make rooms in the example game (grammatically) \"proper\" (nouns)","message":"Make rooms in the example game (grammatically) \"proper\" (nouns)\n","lang":"Python","license":"mit","repos":"twisted\/imaginary"} {"commit":"94e8b7bf8b24dfa36f240e601cb0894b10cab21a","old_file":"tools\/examples\/geturl.py","new_file":"tools\/examples\/geturl.py","old_contents":"#!\/usr\/bin\/env python2\n#\n# USAGE: geturl.py FILE_OR_DIR1 FILE_OR_DIR2 ...\n#\n# prints out the URL associated with each item\n#\n\nimport sys\nimport svn._wc\nimport svn.util\n\ndef main(pool, files):\n for f in files:\n entry = svn._wc.svn_wc_entry(f, 0, pool)\n print svn._wc.svn_wc_entry_t_url_get(entry)\n\nif __name__ == '__main__':\n svn.util.run_app(main, sys.argv[1:])\n","new_contents":"#!\/usr\/bin\/env python2\n#\n# USAGE: geturl.py FILE_OR_DIR1 FILE_OR_DIR2 ...\n#\n# prints out the URL associated with each item\n#\n\nimport os\nimport sys\n\nimport svn.wc\nimport svn.util\n\ndef main(pool, files):\n for f in files:\n dirpath = fullpath = os.path.abspath(f)\n if not os.path.isdir(dirpath):\n dirpath = os.path.dirname(dirpath)\n adm_baton = svn.wc.svn_wc_adm_open(None, dirpath, 1, 1, pool)\n try:\n entry = svn.wc.svn_wc_entry(fullpath, adm_baton, 0, pool)\n print svn.wc.svn_wc_entry_t_url_get(entry)\n except:\n svn.wc.svn_wc_adm_close(adm_baton)\n\nif __name__ == '__main__':\n svn.util.run_app(main, sys.argv[1:])\n","subject":"Update the example to use the new access baton stuff.","message":"Update the example to use the new access baton stuff.\n\n\ngit-svn-id: f8a4e5e023278da1e04e203c7fe051e3c4285d88@844036 13f79535-47bb-0310-9956-ffa450edef68\n","lang":"Python","license":"apache-2.0","repos":"YueLinHo\/Subversion,YueLinHo\/Subversion,wbond\/subversion,wbond\/subversion,YueLinHo\/Subversion,wbond\/subversion,YueLinHo\/Subversion,wbond\/subversion,YueLinHo\/Subversion,YueLinHo\/Subversion,YueLinHo\/Subversion,wbond\/subversion,wbond\/subversion,YueLinHo\/Subversion,wbond\/subversion"} {"commit":"d76cb1aa296bc800cb24427110910a038a62a311","old_file":"vctk\/__init__.py","new_file":"vctk\/__init__.py","old_contents":"","new_contents":"# coding: utf-8\n\nfrom interface import *\n\n\nclass SpeechParameters(object):\n\n \"\"\"\n Speech parameters\n \"\"\"\n\n def __init__(self, f0, spectrum_envelope, aperiodicity):\n self.f0 = f0\n self.spectrum_envelope = spectrum_envelope\n self.aperiodicity = aperiodicity\n\n\nclass VoiceConverter(object):\n\n \"\"\"\n Voice conversion\n\n This class assumes:\n - *_parameterizer implements `Parameterizer`\n - *_converter implements `Converter`\n - analyzer implements `Analyzer`\n - synthesizer implments `Synthesizer`\n\n analyzer and synthesizer must be specified explicitly.\n\n *_parameterizer and *_converter can be None.\n\n TODO:\n parameterizerは、デフォでTrasparentParameterizer\n (つまり特徴量をそのままパスするだけのparamterizer)にする?\n \"\"\"\n\n def __init__(self,\n f0_parameterizer=None,\n f0_converter=None,\n spectrum_envelope_parameterizer=None,\n spectrum_envelope_converter=None,\n aperiodicity_parameterizer=None,\n aperiodicity_converter=None,\n analyzer=None,\n synthesizer=None\n ):\n self.f0_converter = f0_converter\n self.f0_parameterizer = f0_parameterizer\n self.spectrum_envelope_converter = spectrum_envelope_converter\n self.spectrum_envelope_parameterizer = spectrum_envelope_parameterizer\n self.aperiodicity_converter = aperiodicity_converter\n self.aperiodicity_parameterizer = aperiodicity_parameterizer\n\n if analyzer == None or synthesizer == None:\n raise \"backend must be specified explicitly!\"\n\n self.analyzer = analyzer\n self.synthesizer = synthesizer\n\n # speech paramters will be stored.\n self.params = None\n\n def analyze(self, x):\n \"\"\"\n Decompose speech into parametric representation\n \"\"\"\n self.params = self.analyzer.analyze(x)\n\n def convert(self):\n \"\"\"\n Perform speech parameter conversion\n \"\"\"\n if self.params == None:\n raise \"`analyze` must be called before `convert`\"\n\n if self.f0_converter != None:\n self.params.f0 = self.f0_parameterizer.backward(\n self.f0_converter.convert(\n self.f0_parameterizer.forward(self.params.f0)\n )\n )\n\n if self.spectrum_envelope_converter != None:\n self.params.spectrum_envelop = \\\n self.spectrum_envelope_parameterizer.backward(\n self.spectrum_envelope_converter.convert(\n self.spectrum_envelope_parameterizer.forward(\n self.params.spectrum_envelope\n )\n )\n )\n\n if self.aperiodicity_converter != None:\n self.params.aperiodicity = self.aperiodicity_parameterizer.backward(\n self.aperiodicity_converter.convert(\n self.aperiodicity_parameterizer.forward(\n self.params.aperiodicity)\n )\n )\n\n def synthesis(self):\n \"\"\"\n Synthesize speech waveform\n \"\"\"\n if self.params == None:\n raise \"`analyze` must be called before `synthesis`\"\n\n return self.synthesizer.synthesis(self.params)\n","subject":"Add class `VoiceConverter` that peforms all required processes in statistical voice conversion: speech analysis, feature parameterization, feature conversion and waveform syntheis","message":"Add class `VoiceConverter` that peforms all required processes in statistical voice conversion: speech analysis, feature parameterization, feature conversion and waveform syntheis\n","lang":"Python","license":"mit","repos":"k2kobayashi\/sprocket"} {"commit":"15652b6817ad4548881883ee89981aff49c52c56","old_file":"muranoagent\/version.py","new_file":"muranoagent\/version.py","old_contents":"# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\nimport pbr.version\n\nversion_info = pbr.version.VersionInfo('muranoagent')\n","new_contents":"# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\nimport pbr.version\n\nversion_info = pbr.version.VersionInfo('murano-agent')\n","subject":"Fix murano agent component name","message":"Fix murano agent component name\n\nChange-Id: I8f76630d3f007f89585b3418ff661c48004a2b5f\nCloses-bug: #1373347\n","lang":"Python","license":"apache-2.0","repos":"Bloomie\/murano-agent,Bloomie\/murano-agent,telefonicaid\/murano-agent,openstack\/murano-agent,telefonicaid\/murano-agent,Bloomie\/murano-agent,openstack\/murano-agent,telefonicaid\/murano-agent,openstack\/murano-agent,openstack\/murano-agent,Bloomie\/murano-agent"} {"commit":"d23b82e89c756e8cd51aa2abfacd6cc7b2907fae","old_file":"pygame\/__init__.py","new_file":"pygame\/__init__.py","old_contents":"\"\"\" XXX: fish \"\"\"\n\n__all__ = ['display', 'color', 'surface', 'Color', 'time', 'event',\n 'constants', 'sprite', 'Rect', 'Surface', 'QUIT', 'init',\n 'mouse', 'locals', 'image']\n\nfrom pygame.color import Color\nfrom pygame.rect import Rect\nfrom pygame.surface import Surface\nfrom pygame.constants import QUIT\nfrom pygame import display, color, surface, time, event, constants, sprite,\\\n mouse, locals, images\n# XXX\nfrom pygame.display import init\n","new_contents":"\"\"\" XXX: fish \"\"\"\n\n__all__ = ['display', 'color', 'surface', 'Color', 'time', 'event',\n 'constants', 'sprite', 'Rect', 'Surface', 'QUIT', 'init',\n 'mouse', 'locals', 'image']\n\nfrom pygame.color import Color\nfrom pygame.rect import Rect\nfrom pygame.surface import Surface\nfrom pygame.constants import QUIT\nfrom pygame import display, color, surface, time, event, constants, sprite,\\\n mouse, locals, image\n# XXX\nfrom pygame.display import init\n","subject":"Fix error introduced in rebase","message":"Fix error introduced in rebase\n","lang":"Python","license":"lgpl-2.1","repos":"GertBurger\/pygame_cffi,caseyc37\/pygame_cffi,GertBurger\/pygame_cffi,CTPUG\/pygame_cffi,caseyc37\/pygame_cffi,caseyc37\/pygame_cffi,GertBurger\/pygame_cffi,GertBurger\/pygame_cffi,CTPUG\/pygame_cffi,CTPUG\/pygame_cffi"} {"commit":"dc10cbafe045d55906d627816a88323fb4a8c948","old_file":"exec_proc.py","new_file":"exec_proc.py","old_contents":"#!\/usr\/bin\/env python\n# Copyright 2014 Boundary, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom subprocess import Popen,PIPE\nimport shlex\nimport logging\n\n\n\nclass ExecProc:\n \n def __init__(self):\n self.command = None\n self.debug = False\n \n def setDebug(self,debug):\n self.debug = debug\n \n def setCommand(self,command):\n if type(command) != str:\n raise ValueError\n self.command = command\n \n def execute(self):\n if self.command == None:\n raise ValueError\n args = shlex.split(self.command)\n if self.debug == True:\n logging.info(\"command=\\\"%s\\\"\",self.command)\n p = Popen(args,stdout=PIPE)\n o,e = p.communicate()\n if self.debug == True:\n logging.info(\"output=\\\"%s\\\"\",o)\n return o","new_contents":"#!\/usr\/bin\/env python\n# Copyright 2014 Boundary, Inc.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom subprocess import Popen,PIPE\nimport shlex\nimport logging\n\nclass ExecProc:\n \n def __init__(self):\n self.command = None\n self.debug = False\n \n def setDebug(self,debug):\n self.debug = debug\n \n def setCommand(self,command):\n if type(command) != str:\n raise ValueError\n self.command = command\n \n def execute(self):\n if self.command == None:\n raise ValueError\n # Remove Carriage Returns\n command = self.command.strip('\\r')\n args = shlex.split(command)\n if self.debug == True:\n logging.info(\"command=\\\"%s\\\"\",args)\n p = Popen(args,stdout=PIPE)\n o,e = p.communicate()\n if self.debug == True:\n logging.info(\"output=\\\"%s\\\"\",o)\n logging.info(':'.join(x.encode('hex') for x in o))\n return o\n\n","subject":"Add output of command in hex","message":"Add output of command in hex\n","lang":"Python","license":"apache-2.0","repos":"boundary\/boundary-plugin-shell,boundary\/boundary-plugin-shell,jdgwartney\/boundary-plugin-shell,jdgwartney\/boundary-plugin-shell"} {"commit":"68b1c3804504ecc14f7c23465ca11db31489e1cd","old_file":"mozcal\/events\/views.py","new_file":"mozcal\/events\/views.py","old_contents":"from django.shortcuts import render, get_object_or_404\n\nfrom mozcal.events.models import Event, Space, FunctionalArea\n\ndef one(request, slug):\n event = get_object_or_404(Event, slug=slug)\n\n return render(request, 'event.html', { 'event': event })\n\n\ndef all(request):\n events = Event.objects.all()\n spaces = Space.objects.all()\n areas = FunctionalArea.objects.all()\n\n return render(request, 'events_all.html', {\n 'events': events,\n 'spaces': spaces,\n 'areas': areas\n })","new_contents":"from django.shortcuts import render, get_object_or_404\n\nfrom mozcal.events.models import Event, Space, FunctionalArea\n\ndef one(request, slug):\n event = get_object_or_404(Event, slug=slug)\n\n return render(request, 'event.html', { 'event': event })\n\n\ndef all(request):\n search_string = request.GET.get('search', '')\n\n events = Event.objects.filter(title__icontains=search_string)\n spaces = Space.objects.all()\n areas = FunctionalArea.objects.all()\n\n return render(request, 'events_all.html', {\n 'events': events,\n 'spaces': spaces,\n 'areas': areas\n })","subject":"Allow filtering of events by title","message":"Allow filtering of events by title\n","lang":"Python","license":"bsd-3-clause","repos":"ppapadeas\/wprevents,yvan-sraka\/wprevents,yvan-sraka\/wprevents,ppapadeas\/wprevents,yvan-sraka\/wprevents,ppapadeas\/wprevents,yvan-sraka\/wprevents"} {"commit":"ce191a9ea7bad7493560a7bdd7f7de2e56f94612","old_file":"fuse_util.py","new_file":"fuse_util.py","old_contents":"import sublime\nimport os\n\ndef getSetting(key,default=None):\n\ts = sublime.load_settings(\"Fuse.sublime-settings\")\n\treturn s.get(key, default)\n\ndef getFusePathFromSettings():\n\tpath = getSetting(\"fuse_path_override\")\n\tif path == \"\" or path == None:\n\t\treturn \"fuse\"\n\telse:\n\t\treturn path+\"\/fuse\"\n\ndef setSetting(key,value):\n\ts = sublime.load_settings(\"Fuse.sublime-settings\")\n\ts.set(key, value)\n\tsublime.save_settings(\"Fuse.sublime-settings\")\n\ndef isSupportedSyntax(syntaxName):\t\n\treturn syntaxName == \"Uno\" or syntaxName == \"UX\"\n\ndef getExtension(path):\n\tbase = os.path.basename(path)\n\treturn os.path.splitext(base)[0]\n\ndef getRowCol(view, pos):\n\trowcol = view.rowcol(pos)\n\trowcol = (rowcol[0] + 1, rowcol[1] + 1)\n\treturn {\"Line\": rowcol[0], \"Character\": rowcol[1]}","new_contents":"import sublime\nimport os\n\ndef getSetting(key,default=None):\n\ts = sublime.load_settings(\"Fuse.sublime-settings\")\n\treturn s.get(key, default)\n\ndef getFusePathFromSettings():\n\tpath = getSetting(\"fuse_path_override\")\n\tif path == \"\" or path == None:\n\t\tif os.path.isfile(\"\/usr\/bin\/fuse\"):\n\t\t\treturn \"\/usr\/bin\/fuse\"\n\t\telse:\n\t\t\treturn \"\/usr\/local\/bin\/fuse\"\n\telse:\n\t\treturn path+\"\/fuse\"\n\ndef setSetting(key,value):\n\ts = sublime.load_settings(\"Fuse.sublime-settings\")\n\ts.set(key, value)\n\tsublime.save_settings(\"Fuse.sublime-settings\")\n\ndef isSupportedSyntax(syntaxName):\t\n\treturn syntaxName == \"Uno\" or syntaxName == \"UX\"\n\ndef getExtension(path):\n\tbase = os.path.basename(path)\n\treturn os.path.splitext(base)[0]\n\ndef getRowCol(view, pos):\n\trowcol = view.rowcol(pos)\n\trowcol = (rowcol[0] + 1, rowcol[1] + 1)\n\treturn {\"Line\": rowcol[0], \"Character\": rowcol[1]}\n","subject":"Make Sublime plugin work with new Fuse install location","message":"Make Sublime plugin work with new Fuse install location\n\nWe changed the location of fuse from \/usr\/bin to \/usr\/local\/bin to be\ncompatible with El Capitan. The latter is not on path in Sublime, so use\nabsolute paths for fuse.\n","lang":"Python","license":"mit","repos":"fusetools\/Fuse.SublimePlugin,fusetools\/Fuse.SublimePlugin"} {"commit":"02fc723b8d459de1bcf94dc02f7b8531dd9e7cfb","old_file":"events\/tests\/test_user_get.py","new_file":"events\/tests\/test_user_get.py","old_contents":"# -*- coding: utf-8 -*-\nimport pytest\nfrom .utils import get, versioned_reverse as reverse, assert_fields_exist\n\n\n# === util methods ===\n\ndef get_list(api_client, version='v1'):\n list_url = reverse('user-list', version=version)\n return get(api_client, list_url)\n\ndef assert_user_fields_exist(data, version='v1'):\n # TODO: incorporate version parameter into version aware\n # parts of test code\n fields = (\n 'last_login',\n 'username',\n 'email',\n 'date_joined',\n 'first_name',\n 'last_name',\n 'uuid',\n 'department_name',\n 'organization',\n 'is_staff',\n 'display_name',\n )\n assert_fields_exist(data, fields)\n\n# === tests ===\n\n@pytest.mark.django_db\ndef test__get_user_list(api_client, user, organization):\n organization.admin_users.add(user)\n api_client.force_authenticate(user=user)\n response = get_list(api_client)\n print(response.data)\n assert_user_fields_exist(response.data['data'][0])\n","new_contents":"# -*- coding: utf-8 -*-\nimport pytest\nfrom .utils import get, versioned_reverse as reverse, assert_fields_exist\n\n\n# === util methods ===\n\ndef get_list(api_client, version='v1'):\n list_url = reverse('user-list', version=version)\n return get(api_client, list_url)\n\ndef get_detail(api_client, detail_pk, version='v1'):\n detail_url = reverse('user-detail', version=version, kwargs={'pk': detail_pk})\n return get(api_client, detail_url)\n\ndef assert_user_fields_exist(data, version='v1'):\n # TODO: incorporate version parameter into version aware\n # parts of test code\n fields = (\n 'last_login',\n 'username',\n 'email',\n 'date_joined',\n 'first_name',\n 'last_name',\n 'uuid',\n 'department_name',\n 'organization',\n 'is_staff',\n 'display_name',\n )\n assert_fields_exist(data, fields)\n\n# === tests ===\n\n@pytest.mark.django_db\ndef test__get_user_list(api_client, user, organization):\n organization.admin_users.add(user)\n api_client.force_authenticate(user=user)\n response = get_detail(api_client, user.pk)\n print(response.data)\n assert_user_fields_exist(response.data)\n","subject":"Check user detail in test","message":"Check user detail in test\n","lang":"Python","license":"mit","repos":"aapris\/linkedevents,tuomas777\/linkedevents,City-of-Helsinki\/linkedevents,tuomas777\/linkedevents,tuomas777\/linkedevents,City-of-Helsinki\/linkedevents,aapris\/linkedevents,City-of-Helsinki\/linkedevents,aapris\/linkedevents"} {"commit":"d0908d1e4e5279579a93772210b001c19fae9b10","old_file":"cogs\/misc.py","new_file":"cogs\/misc.py","old_contents":"import discord\nfrom discord.ext import commands\n\nclass Misc:\n @commands.command()\n async def highfive(self, ctx):\n \"\"\"\n Give Yutu a high-five\n \"\"\"\n await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author))\n\n @commands.command()\n async def cute(self, ctx, user: discord.Member = None):\n \"\"\"\n Tell someone they are cute!\n\n Tells a user that you think they are cute, if you don't give a user, then Yutu will let you know that you are cute.\n \"\"\"\n if user is None:\n first = ctx.me\n second = ctx.author\n else:\n first = ctx.author\n second = user\n post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first,\n second))\n post.set_image(url=\"https:\/\/i.imgur.com\/MuVAkV2.gif\")\n await ctx.send(embed=post)","new_contents":"import discord\nfrom discord.ext import commands\n\nclass Misc:\n @commands.command()\n async def highfive(self, ctx: commands.Context):\n \"\"\"\n Give Yutu a high-five\n \"\"\"\n await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author))\n\n @commands.command()\n async def cute(self, ctx: commands.Context, user: discord.Member = None):\n \"\"\"\n Tell someone they are cute!\n\n Tells a user that you think they are cute, if you don't give a user, then Yutu will let you know that you are cute.\n \"\"\"\n if user is None:\n first = ctx.me\n second = ctx.author\n else:\n first = ctx.author\n second = user\n post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first,\n second))\n post.set_image(url=\"https:\/\/i.imgur.com\/MuVAkV2.gif\")\n await ctx.send(embed=post)","subject":"Add type markers for ctx objects","message":"Add type markers for ctx objects\n","lang":"Python","license":"mit","repos":"HarkonenBade\/yutu"} {"commit":"2263d180184c908b0e96d53f43f6c81aa23a3c92","old_file":"push\/urls.py","new_file":"push\/urls.py","old_contents":"from django.conf.urls import url\nfrom push import views\n\nurlpatterns = [\n url(r'^$', views.index, name = 'index'),\n url(r'^sender', views.sender, name = 'sender'),\n url(r'^notification_list', views.notification_list, name = 'notification_list'),\n url(r'^settings', views.settings, name = 'settings'),\n url(r'^notification', views.notification, name = 'notification'),\n url(r'^register', views.device_token_register, name = 'device_token_register'),\n url(r'^delete\/device_token\/(?P<device_token_id>\\d+)\/$', views.delete_device_token, name = 'delete_device_token'),\n]\n","new_contents":"from django.conf.urls import url\nfrom push import views\n\nurlpatterns = [\n url(r'^$', views.index, name = 'index'),\n url(r'^sender', views.sender, name = 'sender'),\n url(r'^notification_list', views.notification_list, name = 'notification_list'),\n url(r'^settings', views.settings, name = 'settings'),\n url(r'^notification', views.notification, name = 'notification'),\n url(r'^(?P<username>\\w+)\/register', views.device_token_register, name = 'device_token_register'),\n url(r'^delete\/device_token\/(?P<device_token_id>\\d+)\/$', views.delete_device_token, name = 'delete_device_token'),\n]\n","subject":"Modify device_token register URL dispatcher","message":"Modify device_token register URL dispatcher\n","lang":"Python","license":"apache-2.0","repos":"nnsnodnb\/django-mbaas,nnsnodnb\/django-mbaas,nnsnodnb\/django-mbaas"} {"commit":"d5f0b698831e4bfb35b74ef0d8c7af75c91e67d3","old_file":"dadd\/worker\/handlers.py","new_file":"dadd\/worker\/handlers.py","old_contents":"import os\nimport json\nimport socket\n\nimport requests\n\nfrom flask import request, jsonify, Response, abort\n\nfrom dadd.worker import app\nfrom dadd.worker.proc import ChildProcess\n\n\n@app.route('\/run\/', methods=['POST'])\ndef run_process():\n proc = ChildProcess(request.json)\n proc.run()\n return jsonify(proc.info())\n\n\n@app.route('\/register\/', methods=['POST'])\ndef register_with_master():\n register(app)\n return jsonify({'message': 'ok'})\n\n\ndef register(host, port):\n sess = requests.Session()\n\n if 'USERNAME' in app.config and 'PASSWORD' in app.config:\n sess.auth = (app.config['USERNAME'], app.config['PASSWORD'])\n sess.headers = {'content-type': 'application\/json'}\n\n try:\n url = app.config['MASTER_URL'] + '\/api\/hosts\/'\n resp = sess.post(url, data=json.dumps({\n 'host': app.config.get('HOSTNAME', socket.getfqdn()),\n 'port': port\n }))\n if not resp.ok:\n app.logger.warning('Error registering with master: %s' %\n app.config['MASTER_URL'])\n except Exception as e:\n app.logger.warning('Connection Error: %s' % e)\n\n\n@app.route('\/logs\/<path>', methods=['GET'])\ndef tail_log(path):\n if os.path.exists(path) and path.startswith('\/tmp\/'):\n return Response(open(path), content_type='text\/plain')\n abort(404)\n","new_contents":"import os\nimport json\nimport socket\n\nimport requests\n\nfrom flask import request, jsonify, Response, abort\n\nfrom dadd.worker import app\nfrom dadd.worker.proc import ChildProcess\n\n\n@app.route('\/run\/', methods=['POST'])\ndef run_process():\n proc = ChildProcess(request.json)\n proc.run()\n return jsonify(proc.info())\n\n\n@app.route('\/register\/', methods=['POST'])\ndef register_with_master():\n register(app.config['HOST'], app.config['PORT'])\n return jsonify({'message': 'ok'})\n\n\ndef register(host, port):\n sess = requests.Session()\n\n if 'USERNAME' in app.config and 'PASSWORD' in app.config:\n sess.auth = (app.config['USERNAME'], app.config['PASSWORD'])\n sess.headers = {'content-type': 'application\/json'}\n\n try:\n url = app.config['MASTER_URL'] + '\/api\/hosts\/'\n resp = sess.post(url, data=json.dumps({\n 'host': app.config.get('HOSTNAME', socket.getfqdn()),\n 'port': port\n }))\n if not resp.ok:\n app.logger.warning('Error registering with master: %s' %\n app.config['MASTER_URL'])\n except Exception as e:\n app.logger.warning('Connection Error: %s' % e)\n\n\n@app.route('\/logs\/<path:path>', methods=['GET'])\ndef tail_log(path):\n path = '\/' + path\n if os.path.exists(path) and path.startswith('\/tmp\/'):\n return Response(open(path), content_type='text\/plain')\n abort(404)\n","subject":"Fix up the manual register URL in the worker and fix the initial log tail.","message":"Fix up the manual register URL in the worker and fix the initial log tail.\n","lang":"Python","license":"bsd-3-clause","repos":"ionrock\/dadd,ionrock\/dadd,ionrock\/dadd,ionrock\/dadd"} {"commit":"3501f3404aebf6dc7ba349eafdc80602b98f72a9","old_file":"snaek\/ffi.py","new_file":"snaek\/ffi.py","old_contents":"import os\nimport re\nimport cffi\n\n\n_directive_re = re.compile(r'^\\s*#.*?$(?m)')\n\n\ndef make_ffi(module_path, crate_path, cached_header_filename=None):\n \"\"\"Creates a FFI instance for the given configuration.\"\"\"\n if cached_header_filename is not None and \\\n os.path.isfile(cached_header_filename):\n with open(cached_header_filename, 'rb') as f:\n header = f.read()\n else:\n from .bindgen import generate_header\n header = generate_header(crate_path)\n header = _directive_re.sub('', header)\n ffi = cffi.FFI()\n ffi.cdef(header)\n ffi.set_source(module_path, None)\n return ffi\n","new_contents":"import os\nimport re\nimport sys\nimport cffi\n\n\n_directive_re = re.compile(r'^\\s*#.*?$(?m)')\n\n\ndef make_ffi(module_path, crate_path, cached_header_filename=None):\n \"\"\"Creates a FFI instance for the given configuration.\"\"\"\n if cached_header_filename is not None and \\\n os.path.isfile(cached_header_filename):\n with open(cached_header_filename, 'rb') as f:\n header = f.read()\n else:\n from .bindgen import generate_header\n header = generate_header(crate_path)\n header = _directive_re.sub('', header)\n\n if os.environ.get('SNAEK_DEBUG_HEADER') == '1':\n sys.stderr.write('\/* generated header for \"%s\" *\/\\n' % module_path)\n sys.stderr.write(header)\n sys.stderr.write('\\n')\n sys.stderr.flush()\n\n ffi = cffi.FFI()\n ffi.cdef(header)\n ffi.set_source(module_path, None)\n return ffi\n","subject":"Add a way to dump the header during setup.py runs","message":"Add a way to dump the header during setup.py runs\n","lang":"Python","license":"apache-2.0","repos":"mitsuhiko\/snaek,mitsuhiko\/snaek,mitsuhiko\/snaek"} {"commit":"fb1d39ed30e73bef49be7a71945d5dfd67af28e3","old_file":"scripting.py","new_file":"scripting.py","old_contents":"#!\/usr\/bin\/env python2\n\nimport os, shutil\n\ndef print_warning(message, *args, **kwargs):\n import colortext\n if args or kwargs: message = message.format(*args, **kwargs)\n colortext.write(message, color='red')\n\ndef print_error_and_die(message, *args, **kwargs):\n print_warning(message + \" Aborting...\", *args, **kwargs)\n raise SystemExit(1)\n\ndef clear_directory(directory):\n if os.path.exists(directory): shutil.rmtree(directory)\n os.makedirs(directory)\n\n","new_contents":"#!\/usr\/bin\/env python2\n\nimport os, shutil\n\ndef print_warning(message, *args, **kwargs):\n import colortext\n if args or kwargs: message = message.format(*args, **kwargs)\n colortext.write(message, color='red')\n\ndef print_error_and_die(message, *args, **kwargs):\n print_warning(message + \" Aborting...\", *args, **kwargs)\n raise SystemExit(1)\n\ndef clear_directory(directory):\n if os.path.exists(directory): shutil.rmtree(directory)\n os.makedirs(directory)\n\ndef mkdir(newdir):\n if os.path.isdir(newdir):\n pass\n elif os.path.isfile(newdir):\n raise OSError(\"a file with the same name as the desired \" \\\n \"dir, '%s', already exists.\" % newdir)\n else:\n os.makedirs(newdir)\n","subject":"Add a friendly mkdir() function.","message":"Add a friendly mkdir() function.\n","lang":"Python","license":"mit","repos":"Kortemme-Lab\/klab,Kortemme-Lab\/klab,Kortemme-Lab\/klab,Kortemme-Lab\/klab"} {"commit":"0216bfd48fddb9bb7bda611ec5bdfe368bdee55f","old_file":"layout\/tests.py","new_file":"layout\/tests.py","old_contents":"from django.test import TestCase\n\n# Create your tests here.\n","new_contents":"from django.core.urlresolvers import resolve\nfrom django.test import TestCase\n\nfrom layout.views import home\n\n\nclass HomePageTest(TestCase):\n\n def test_root_url_resolves_to_home_page(self):\n found = resolve('\/')\n self.assertEqual(found.func, home)\n","subject":"Add home page resolve test to layout","message":"Add home page resolve test to layout\n","lang":"Python","license":"mit","repos":"jvanbrug\/scout,jvanbrug\/scout"} {"commit":"6556acc2d1be648fdb362f1f1e5000f443642416","old_file":"examples\/mnist-autoencoder.py","new_file":"examples\/mnist-autoencoder.py","old_contents":"#!\/usr\/bin\/env python\n\nimport cPickle\nimport gzip\nimport logging\nimport os\nimport sys\nimport tempfile\nimport urllib\n\nimport lmj.tnn\n\nlogging.basicConfig(\n stream=sys.stdout,\n format='%(levelname).1s %(asctime)s %(message)s',\n level=logging.INFO)\n\nURL = 'http:\/\/www.iro.umontreal.ca\/~lisa\/deep\/data\/mnist\/mnist.pkl.gz'\nDATASET = os.path.join(tempfile.gettempdir(), 'mnist.pkl.gz')\n\nif not os.path.isfile(DATASET):\n logging.info('downloading mnist digit dataset from %s' % URL)\n urllib.urlretrieve(URL, DATASET)\n logging.info('saved mnist digits to %s' % DATASET)\n\nclass Main(lmj.tnn.Main):\n def get_network(self):\n return lmj.tnn.Autoencoder\n\n def get_datasets(self):\n return [(x, ) for x, _ in cPickle.load(gzip.open(DATASET))]\n\npath = os.path.join(tempfile.gettempdir(), 'mnist-autoencoder.pkl.gz')\nMain().train().save(path)\nprint 'saved network to', path\n","new_contents":"#!\/usr\/bin\/env python\n\nimport cPickle\nimport gzip\nimport logging\nimport os\nimport sys\nimport tempfile\nimport urllib\n\nimport lmj.tnn\n\nlogging.basicConfig(\n stream=sys.stdout,\n format='%(levelname).1s %(asctime)s %(message)s',\n level=logging.INFO)\n\nURL = 'http:\/\/www.iro.umontreal.ca\/~lisa\/deep\/data\/mnist\/mnist.pkl.gz'\nDATASET = os.path.join(tempfile.gettempdir(), 'mnist.pkl.gz')\n\nif not os.path.isfile(DATASET):\n logging.info('downloading mnist digit dataset from %s' % URL)\n urllib.urlretrieve(URL, DATASET)\n logging.info('saved mnist digits to %s' % DATASET)\n\nclass Main(lmj.tnn.Main):\n def get_network(self):\n return lmj.tnn.Autoencoder\n\n def get_datasets(self):\n return [x for x, _ in cPickle.load(gzip.open(DATASET))]\n\npath = os.path.join(tempfile.gettempdir(), 'mnist-autoencoder.pkl.gz')\nMain().train().save(path)\nprint 'saved network to', path\n","subject":"Return unlabeled data for the autoencoder as a straight array.","message":"Return unlabeled data for the autoencoder as a straight array.\n","lang":"Python","license":"mit","repos":"lmjohns3\/theanets,chrinide\/theanets,devdoer\/theanets"} {"commit":"9debed5d1d83bdf2098a7a3841ae4ff272e7ea8e","old_file":"lib\/__init__.py","new_file":"lib\/__init__.py","old_contents":"from client import WebHDFSClient\n\n__version__ = '1.0'\n","new_contents":"from errors import WebHDFSError\nfrom client import WebHDFSClient\nfrom attrib import WebHDFSObject\n\n__version__ = '1.0'\n","subject":"Make other API classes available from base module.","message":"Make other API classes available from base module.\n","lang":"Python","license":"mit","repos":"mk23\/webhdfs,mk23\/webhdfs"} {"commit":"2cb2779bfe1ddfcd6651665276ed0a1d513c57de","old_file":"fireplace\/cards\/wog\/shaman.py","new_file":"fireplace\/cards\/wog\/shaman.py","old_contents":"from ..utils import *\n\n\n##\n# Minions\n\nclass OG_023:\n\t\"Primal Fusion\"\n\tplay = Buff(TARGET, \"OG_023t\") * Count(FRIENDLY_MINIONS + TOTEM)\n\nOG_023t = buff(+1, +1)\n\n\nclass OG_209:\n\t\"Hallazeal the Ascended\"\n\tevents = Damage(source=SPELL + FRIENDLY).on(Heal(FRIENDLY_HERO, Damage.AMOUNT))\n","new_contents":"from ..utils import *\n\n\n##\n# Minions\n\nclass OG_023:\n\t\"Primal Fusion\"\n\tplay = Buff(TARGET, \"OG_023t\") * Count(FRIENDLY_MINIONS + TOTEM)\n\nOG_023t = buff(+1, +1)\n\n\nclass OG_026:\n\t\"Eternal Sentinel\"\n\tplay = UnlockOverload(CONTROLLER)\n\n\nclass OG_209:\n\t\"Hallazeal the Ascended\"\n\tevents = Damage(source=SPELL + FRIENDLY).on(Heal(FRIENDLY_HERO, Damage.AMOUNT))\n\n\n##\n# Spells\n\nclass OG_206:\n\t\"Stormcrack\"\n\tplay = Hit(TARGET, 4)\n\n\n##\n# Weapons\n\nclass OG_031:\n\t\"Hammer of Twilight\"\n\tdeathrattle = Summon(CONTROLLER, \"OG_031a\")\n","subject":"Implement Eternal Sentinel, Stormcrack and Hammer of Twilight","message":"Implement Eternal Sentinel, Stormcrack and Hammer of Twilight\n","lang":"Python","license":"agpl-3.0","repos":"NightKev\/fireplace,beheh\/fireplace,jleclanche\/fireplace"} {"commit":"562b56d67d7d292d7c63ec8c3f453bae92a3b073","old_file":"tests\/test_wysiwyg_editor.py","new_file":"tests\/test_wysiwyg_editor.py","old_contents":"from . import TheInternetTestCase\nfrom helium.api import click, Text, press, CONTROL, COMMAND, write\nfrom sys import platform\n\nclass WYSIWYGEditorTest(TheInternetTestCase):\n\tdef get_page(self):\n\t\treturn \"http:\/\/the-internet.herokuapp.com\/tinymce\"\n\tdef test_use_wysiwyg_editor(self):\n\t\tself.assertTrue(Text(\"Your content goes here.\").exists())\n\t\tclick(\"Your content goes here.\")\n\t\tif platform == 'darwin':\n\t\t\tpress(COMMAND + 'a')\n\t\telse:\n\t\t\tpress(CONTROL + 'a')\n\t\twrite(\"Hello Helium!\")\n\t\tself.assertTrue(Text(\"Hello Helium!\").exists())","new_contents":"from . import TheInternetTestCase\nfrom helium.api import click, Text, write\n\nclass WYSIWYGEditorTest(TheInternetTestCase):\n\tdef get_page(self):\n\t\treturn \"http:\/\/the-internet.herokuapp.com\/tinymce\"\n\tdef test_use_wysiwyg_editor(self):\n\t\tself.assertTrue(Text(\"Your content goes here.\").exists())\n\t\tclick(\"File\")\n\t\tclick(\"New document\")\n\t\twrite(\"Hello Helium!\")\n\t\tself.assertTrue(Text(\"Hello Helium!\").exists())","subject":"Simplify the WYSIWYG editor test case.","message":"Simplify the WYSIWYG editor test case.\n","lang":"Python","license":"mit","repos":"bugfree-software\/the-internet-solution-python"} {"commit":"47b3d205931d6ee7fa8062b3e2f01d1ea07df11a","old_file":"pathvalidate\/_error.py","new_file":"pathvalidate\/_error.py","old_contents":"# encoding: utf-8\n\n\"\"\"\n.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import unicode_literals\n\n\nclass NullNameError(ValueError):\n \"\"\"\n Raised when a name is empty.\n \"\"\"\n\n\nclass InvalidCharError(ValueError):\n \"\"\"\n Raised when includes invalid character(s) within a string.\n \"\"\"\n\n\nclass InvalidCharWindowsError(InvalidCharError):\n \"\"\"\n Raised when includes Windows specific invalid character(s) within a string.\n \"\"\"\n\n\nclass InvalidLengthError(ValueError):\n \"\"\"\n Raised when a string too long\/short.\n \"\"\"\n\n\nclass ReservedNameError(ValueError):\n \"\"\"\n Raised when a string is matched a reserved name.\n \"\"\"\n","new_contents":"# encoding: utf-8\n\n\"\"\"\n.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>\n\"\"\"\n\nfrom __future__ import absolute_import\nfrom __future__ import unicode_literals\n\n\nclass InvalidNameError(ValueError):\n \"\"\"\n Base class of invalid name error.\n \"\"\"\n\n\nclass NullNameError(InvalidNameError):\n \"\"\"\n Raised when a name is empty.\n \"\"\"\n\n\nclass InvalidCharError(InvalidNameError):\n \"\"\"\n Raised when includes invalid character(s) within a string.\n \"\"\"\n\n\nclass InvalidCharWindowsError(InvalidCharError):\n \"\"\"\n Raised when includes Windows specific invalid character(s) within a string.\n \"\"\"\n\n\nclass InvalidLengthError(InvalidNameError):\n \"\"\"\n Raised when a string too long\/short.\n \"\"\"\n\n\nclass ReservedNameError(InvalidNameError):\n \"\"\"\n Raised when a string is matched a reserved name.\n \"\"\"\n","subject":"Add base class of invalid name error","message":"Add base class of invalid name error\n","lang":"Python","license":"mit","repos":"thombashi\/pathvalidate"} {"commit":"46fb576c9dcf83c40b67c75dade5b43a4d122e7f","old_file":"platformio\/__init__.py","new_file":"platformio\/__init__.py","old_contents":"# Copyright (C) Ivan Kravets <me@ikravets.com>\n# See LICENSE for details.\n\nVERSION = (0, 7, \"0.dev\")\n__version__ = \".\".join([str(s) for s in VERSION])\n\n__title__ = \"platformio\"\n__description__ = (\"A cross-platform code builder and library manager\")\n__url__ = \"https:\/\/github.com\/ivankravets\/platformio\"\n\n__author__ = \"Ivan Kravets\"\n__email__ = \"me@ikravets.com\"\n\n__license__ = \"MIT License\"\n__copyright__ = \"Copyright (C) 2014 Ivan Kravets\"\n\n# __apiurl__ = \"http:\/\/127.0.0.1:8080\"\n__apiurl__ = \"http:\/\/api.platformio.ikravets.com\"\n__pkgmanifesturl__ = \"http:\/\/platformio.ikravets.com\/packages\/manifest.json\"\n","new_contents":"# Copyright (C) Ivan Kravets <me@ikravets.com>\n# See LICENSE for details.\n\nVERSION = (0, 7, \"0.dev\")\n__version__ = \".\".join([str(s) for s in VERSION])\n\n__title__ = \"platformio\"\n__description__ = (\"A cross-platform code builder and library manager\")\n__url__ = \"https:\/\/github.com\/ivankravets\/platformio\"\n\n__author__ = \"Ivan Kravets\"\n__email__ = \"me@ikravets.com\"\n\n__license__ = \"MIT License\"\n__copyright__ = \"Copyright (C) 2014 Ivan Kravets\"\n\n# __apiurl__ = \"http:\/\/127.0.0.1:8080\"\n__apiurl__ = \"http:\/\/api.platformio.ikravets.com\"\n__pkgmanifesturl__ = \"http:\/\/dl.platformio.ikravets.com\/packages\/manifest.json\"\n","subject":"Switch package manifest to dl.platformio","message":"Switch package manifest to dl.platformio\n","lang":"Python","license":"mit","repos":"aphelps\/platformio,mcanthony\/platformio,TimJay\/platformio,aphelps\/platformio,valeros\/platformio,jrobeson\/platformio,TimJay\/platformio,bkudria\/platformio,platformio\/platformio-core,jrobeson\/platformio,bkudria\/platformio,ZachMassia\/platformio,mplewis\/platformio,mseroczynski\/platformio,bkudria\/platformio,aphelps\/platformio,TimJay\/platformio,bkudria\/platformio,eiginn\/platformio,atyenoria\/platformio,dkuku\/platformio,TimJay\/platformio,TimJay\/platformio,awong1900\/platformio,platformio\/platformio,aphelps\/platformio,awong1900\/platformio,jrobeson\/platformio,platformio\/platformio-core,jrobeson\/platformio,awong1900\/platformio"} {"commit":"752e6cef31ea124f00eced5699fb225501258148","old_file":"reversible.py","new_file":"reversible.py","old_contents":"#!\/usr\/bin\/env python\n\n\"\"\" Some tools for dealing with reversible numbers for problem 145 from Project Euler.\n https:\/\/projecteuler.net\/problem=145\n\"\"\"\n\n\ndef is_odd(num):\n \"\"\" Check if an integer is odd. \"\"\"\n if num % 2 != 0:\n return True\n else:\n return False\n\n\ndef is_reversible(num):\n \"\"\" Check if a number is reversible given the above definition. \"\"\"\n num_str = str(num)\n rev_num = int(\"\".join(reversed(num_str)))\n\n total = num + rev_num\n\n for digit in str(total):\n if not is_odd(int(digit)):\n return False\n\n return True\n\n\nif __name__ == \"__main__\":\n # check some odd and even numbers\n assert is_odd(1), \"1 should be odd\"\n assert not is_odd(2), \"2 should not be odd\"\n assert not is_odd(100), \"100 should not be odd\"\n assert is_odd(10001), \"10001 should be odd\"\n\n # check the example reversible numbers\n assert is_reversible(36), \"36 should be reversible\"\n assert is_reversible(63), \"63 should be reversible\"\n assert is_reversible(409), \"409 should be reversible\"\n assert is_reversible(904), \"904 should be reversible\"\n \n print \"all assertions passed\"\n","new_contents":"#!\/usr\/bin\/env python\n\n\"\"\" Some tools for dealing with reversible numbers for problem 145 from Project Euler.\n https:\/\/projecteuler.net\/problem=145\n\"\"\"\n\n\ndef is_odd(num):\n \"\"\" Check if an integer is odd. \"\"\"\n if num % 2 != 0:\n return True\n else:\n return False\n\n\ndef is_reversible(num):\n \"\"\" Check if a number is reversible given the above definition. \"\"\"\n num_str = str(num)\n rev_str = \"\".join(reversed(num_str))\n\n if int(rev_str[0]) == 0:\n return False\n\n total = num + int(rev_str)\n\n for digit in str(total):\n if not is_odd(int(digit)):\n return False\n\n return True\n\n\nif __name__ == \"__main__\":\n # check some odd and even numbers\n assert is_odd(1), \"1 should be odd\"\n assert not is_odd(2), \"2 should not be odd\"\n assert not is_odd(100), \"100 should not be odd\"\n assert is_odd(10001), \"10001 should be odd\"\n\n # check the example reversible numbers\n assert is_reversible(36), \"36 should be reversible\"\n assert is_reversible(63), \"63 should be reversible\"\n assert is_reversible(409), \"409 should be reversible\"\n assert is_reversible(904), \"904 should be reversible\"\n \n assert not is_reversible(10), \"10 should not be reversible. (leading zero.)\"\n\n print \"all assertions passed\"\n","subject":"Add check for leading zeroes.","message":"Add check for leading zeroes.\n","lang":"Python","license":"mit","repos":"smillet15\/project-euler"} {"commit":"3ffc101a1a8b1ec17e5f2e509a1e5182a1f6f4b9","old_file":"fzn\/utils.py","new_file":"fzn\/utils.py","old_contents":"import subprocess as sp\nimport signal\nimport threading\nimport os\n\n\nSIGTERM_TIMEOUT = 1.0\n\n\nclass Command(object):\n def __init__(self, cmd, memlimit=None):\n self.cmd = cmd\n self.memlimit = memlimit\n self.process = None\n self.stdout = None\n self.stderr = None\n self.exitcode = None\n self.timed_out = False\n\n def run(self, timeout=None):\n def target():\n self.process = sp.Popen(self.cmd,\n stdout=sp.PIPE, stderr=sp.PIPE,\n shell=True, preexec_fn=os.setpgrp)\n self.stdout, self.stderr = self.process.communicate()\n self.exitcode = self.process.returncode\n\n thread = threading.Thread(target=target)\n thread.start()\n thread.join(float(timeout))\n if thread.is_alive():\n self.timed_out = True\n\n # Send the TERM signal to all the process groups\n os.killpg(self.process.pid, signal.SIGTERM)\n thread.join(SIGTERM_TIMEOUT)\n if thread.is_alive():\n # Send the KILL signal if the process hasn't exited by now.\n os.killpg(self.process.pid, signal.SIGKILL)\n self.process.kill()\n thread.join()\n","new_contents":"import subprocess as sp\nimport signal\nimport threading\nimport os\n\n\nSIGTERM_TIMEOUT = 1.0\n\n\nclass Command(object):\n def __init__(self, cmd, memlimit=None):\n self.cmd = cmd\n self.memlimit = memlimit\n self.process = None\n self.stdout = None\n self.stderr = None\n self.exitcode = None\n self.timed_out = False\n\n def run(self, timeout=None):\n def target():\n cmd = self.cmd\n if self.memlimit:\n cmd = \"ulimit -v %d; %s\" % (self.memlimit, cmd)\n self.process = sp.Popen(cmd,\n stdout=sp.PIPE, stderr=sp.PIPE,\n shell=True, preexec_fn=os.setpgrp)\n self.stdout, self.stderr = self.process.communicate()\n self.exitcode = self.process.returncode\n\n thread = threading.Thread(target=target)\n thread.start()\n thread.join(float(timeout))\n if thread.is_alive():\n self.timed_out = True\n\n # Send the TERM signal to all the process groups\n os.killpg(self.process.pid, signal.SIGTERM)\n thread.join(SIGTERM_TIMEOUT)\n if thread.is_alive():\n # Send the KILL signal if the process hasn't exited by now.\n os.killpg(self.process.pid, signal.SIGKILL)\n self.process.kill()\n thread.join()\n","subject":"Enable Command to support memory limiting.","message":"Enable Command to support memory limiting.\n","lang":"Python","license":"lgpl-2.1","repos":"eomahony\/Numberjack,eomahony\/Numberjack,eomahony\/Numberjack,JElchison\/Numberjack,JElchison\/Numberjack,JElchison\/Numberjack,JElchison\/Numberjack,JElchison\/Numberjack,eomahony\/Numberjack,eomahony\/Numberjack"} {"commit":"a89a61620306d3cc38062cf69c56db64aadf0a8d","old_file":"pokedex\/db\/__init__.py","new_file":"pokedex\/db\/__init__.py","old_contents":"import pkg_resources\n\nfrom sqlalchemy import MetaData, Table, create_engine, orm\n\nfrom .tables import metadata\n\ndef connect(uri=None, **kwargs):\n \"\"\"Connects to the requested URI. Returns a session object.\n\n With the URI omitted, attempts to connect to a default SQLite database\n contained within the package directory.\n\n Calling this function also binds the metadata object to the created engine.\n \"\"\"\n\n # Default to a URI within the package, which was hopefully created at some point\n if not uri:\n sqlite_path = pkg_resources.resource_filename('pokedex',\n 'data\/pokedex.sqlite')\n uri = 'sqlite:\/\/\/' + sqlite_path\n\n ### Do some fixery for MySQL\n if uri[0:5] == 'mysql':\n # MySQL uses latin1 for connections by default even if the server is\n # otherwise oozing with utf8; charset fixes this\n if 'charset' not in uri:\n uri += '?charset=utf8'\n\n # Tables should be InnoDB, in the event that we're creating them, and\n # use UTF-8 goddammit!\n for table in metadata.tables.values():\n table.kwargs['mysql_engine'] = 'InnoDB'\n table.kwargs['mysql_charset'] = 'utf8'\n\n ### Connect\n engine = create_engine(uri)\n conn = engine.connect()\n metadata.bind = engine\n\n session_args = dict(autoflush=True, autocommit=False, bind=engine)\n session_args.update(kwargs)\n sm = orm.sessionmaker(**session_args)\n session = orm.scoped_session(sm)\n\n return session\n","new_contents":"import pkg_resources\n\nfrom sqlalchemy import MetaData, Table, create_engine, orm\n\nfrom .tables import metadata\n\ndef connect(uri=None, session_args={}, engine_args={}):\n \"\"\"Connects to the requested URI. Returns a session object.\n\n With the URI omitted, attempts to connect to a default SQLite database\n contained within the package directory.\n\n Calling this function also binds the metadata object to the created engine.\n \"\"\"\n\n # Default to a URI within the package, which was hopefully created at some point\n if not uri:\n sqlite_path = pkg_resources.resource_filename('pokedex',\n 'data\/pokedex.sqlite')\n uri = 'sqlite:\/\/\/' + sqlite_path\n\n ### Do some fixery for MySQL\n if uri[0:5] == 'mysql':\n # MySQL uses latin1 for connections by default even if the server is\n # otherwise oozing with utf8; charset fixes this\n if 'charset' not in uri:\n uri += '?charset=utf8'\n\n # Tables should be InnoDB, in the event that we're creating them, and\n # use UTF-8 goddammit!\n for table in metadata.tables.values():\n table.kwargs['mysql_engine'] = 'InnoDB'\n table.kwargs['mysql_charset'] = 'utf8'\n\n ### Connect\n engine = create_engine(uri, **engine_args)\n conn = engine.connect()\n metadata.bind = engine\n\n all_session_args = dict(autoflush=True, autocommit=False, bind=engine)\n all_session_args.update(session_args)\n sm = orm.sessionmaker(**all_session_args)\n session = orm.scoped_session(sm)\n\n return session\n","subject":"Allow passing engine arguments to connect().","message":"Allow passing engine arguments to connect().\n","lang":"Python","license":"mit","repos":"mschex1\/pokedex,RK905\/pokedex-1,xfix\/pokedex,veekun\/pokedex,DaMouse404\/pokedex,veekun\/pokedex"} {"commit":"9d5abdaefa483574cdd81da8d8d4e63ef68f5ab8","old_file":"crossfolium\/__init__.py","new_file":"crossfolium\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nCrossfolium\n-----------\n\n\"\"\"\n\nimport crossfolium.marker_function as marker_function\n\nfrom crossfolium.crossfolium import (\n Crossfilter,\n PieFilter,\n RowBarFilter,\n BarFilter,\n TableFilter,\n CountFilter,\n ResetFilter,\n GeoChoroplethFilter,\n )\n\nfrom .map import (\n FeatureGroupFilter,\n HeatmapFilter,\n )\n\n__version__ = \"0.0.0\"\n\n__all__ = [\n '__version__',\n 'marker_function',\n 'Crossfilter',\n 'PieFilter',\n 'RowBarFilter',\n 'BarFilter',\n 'FeatureGroupFilter',\n 'TableFilter',\n 'CountFilter',\n 'ResetFilter',\n 'HeatmapFilter',\n 'GeoChoroplethFilter',\n ]\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nCrossfolium\n-----------\n\n\"\"\"\nfrom __future__ import absolute_import\n\nfrom crossfolium import marker_function\n\nfrom crossfolium.crossfolium import (\n Crossfilter,\n PieFilter,\n RowBarFilter,\n BarFilter,\n TableFilter,\n CountFilter,\n ResetFilter,\n GeoChoroplethFilter,\n )\n\nfrom crossfolium.map import (\n FeatureGroupFilter,\n HeatmapFilter,\n )\n\n__version__ = \"0.0.0\"\n\n__all__ = [\n '__version__',\n 'marker_function',\n 'Crossfilter',\n 'PieFilter',\n 'RowBarFilter',\n 'BarFilter',\n 'FeatureGroupFilter',\n 'TableFilter',\n 'CountFilter',\n 'ResetFilter',\n 'HeatmapFilter',\n 'GeoChoroplethFilter',\n ]\n","subject":"Handle absolute import for py27","message":"Handle absolute import for py27\n","lang":"Python","license":"mit","repos":"BibMartin\/crossfolium,BibMartin\/crossfolium"} {"commit":"e325c603e972e6e7cd50eefae23b94594b6c9751","old_file":"Tables\/build_db.py","new_file":"Tables\/build_db.py","old_contents":"import sqlite3\nimport os\nimport pandas as pd\n\nTABLES = [['Natures', 'nature'],\n ['Experience'],\n ]\n\nPATH = os.path.dirname(__file__)+\"\/\"\nCONNECTION = sqlite3.connect(PATH + 'serpyrior.db')\n\n# insert a little jimmy drop tables here\n\n\nfor table in TABLES:\n table_name = table[0]\n print(table_name)\n try:\n table_index = table[1]\n write_index = False\n except IndexError:\n table_index = None\n write_index = True\n\n df = pd.read_csv(PATH + table_name + '.csv')\n df.to_sql(table_name, CONNECTION, index=write_index, index_label=table_index)\n\nCONNECTION.commit()\nCONNECTION.close()\n\n# cur = conn.cursor()\n# cur.execute(\"CREATE TABLE IF NOT EXISTS natures()\")\n# filename.encode('utf-8')\n# with open(filename) as f:\n# reader = csv.reader(f)\n# for field in reader:\n# cur.execute(\"INSERT INTO natures VALUES (?,?,?,?,?,?,?);\", field)\n#\n# conn.commit()\n#\n# df = pd.read_sql_query(\"SELECT * FROM natures\", conn, index_col='nature')\n#\n# print(df.head(25))\n# conn.close()\n","new_contents":"import sqlite3\nimport os\nimport pandas as pd\n\nTABLES = [['Natures', 'nature'],\n ['Experience'],\n ]\n\nPATH = os.path.dirname(__file__)+\"\/\"\n\n\ntry: # Little Bobby Tables\n os.remove(PATH + 'serpyrior.db')\nexcept FileNotFoundError:\n pass\n\nCONNECTION = sqlite3.connect(PATH + 'serpyrior.db')\n\nfor table in TABLES:\n table_name = table[0]\n print(table_name)\n try:\n table_index = table[1]\n write_index = False\n except IndexError:\n table_index = None\n write_index = True\n\n df = pd.read_csv(PATH + table_name + '.csv')\n df.to_sql(table_name, CONNECTION, index=write_index, index_label=table_index)\n\nCONNECTION.commit()\nCONNECTION.close()\n","subject":"Remove db if it already exists","message":"Remove db if it already exists\n","lang":"Python","license":"mit","repos":"Ditoeight\/Pyranitar"} {"commit":"55a8921f3634fe842eddf202d1237f53ca6d003b","old_file":"kobo\/settings\/dev.py","new_file":"kobo\/settings\/dev.py","old_contents":"# coding: utf-8\nfrom .base import *\n\nLOGGING['handlers']['console'] = {\n 'level': 'DEBUG',\n 'class': 'logging.StreamHandler',\n 'formatter': 'verbose'\n}\n\nINSTALLED_APPS = INSTALLED_APPS + ('debug_toolbar',)\nMIDDLEWARE.append('debug_toolbar.middleware.DebugToolbarMiddleware')\n\n# Comment out the line below to use `Django Debug Toolbar`\n# INTERNAL_IPS = ['172.28.0.4'] # Change IP to KPI container's IP\n\nENV = 'dev'\n\n# Expiration time in sec. after which paired data xml file must be regenerated\n# Does not need to match KoBoCAT setting\nPAIRED_DATA_EXPIRATION = 5\n\n# Minimum size (in bytes) of files to allow fast calculation of hashes\n# Should match KoBoCAT setting\nHASH_BIG_FILE_SIZE_THRESHOLD = 200 * 1024 # 200 kB\n\n# Chunk size in bytes to read per iteration when hash of a file is calculated\n# Should match KoBoCAT setting\nHASH_BIG_FILE_CHUNK = 5 * 1024 # 5 kB\n","new_contents":"# coding: utf-8\nfrom .base import *\n\nLOGGING['handlers']['console'] = {\n 'level': 'DEBUG',\n 'class': 'logging.StreamHandler',\n 'formatter': 'verbose'\n}\n\nINSTALLED_APPS = INSTALLED_APPS + ('debug_toolbar',)\nMIDDLEWARE.append('debug_toolbar.middleware.DebugToolbarMiddleware')\n\ndef show_toolbar(request):\n return env.bool(\"DEBUG_TOOLBAR\", False)\nDEBUG_TOOLBAR_CONFIG = {\"SHOW_TOOLBAR_CALLBACK\": show_toolbar}\n\nENV = 'dev'\n\n# Expiration time in sec. after which paired data xml file must be regenerated\n# Does not need to match KoBoCAT setting\nPAIRED_DATA_EXPIRATION = 5\n\n# Minimum size (in bytes) of files to allow fast calculation of hashes\n# Should match KoBoCAT setting\nHASH_BIG_FILE_SIZE_THRESHOLD = 200 * 1024 # 200 kB\n\n# Chunk size in bytes to read per iteration when hash of a file is calculated\n# Should match KoBoCAT setting\nHASH_BIG_FILE_CHUNK = 5 * 1024 # 5 kB\n","subject":"Enable django debug toolbar via env var","message":"Enable django debug toolbar via env var\n","lang":"Python","license":"agpl-3.0","repos":"kobotoolbox\/kpi,kobotoolbox\/kpi,kobotoolbox\/kpi,kobotoolbox\/kpi,kobotoolbox\/kpi"} {"commit":"80fb36f2e8754a07ae2f6f4b454862a8b1852763","old_file":"dadd\/worker\/handlers.py","new_file":"dadd\/worker\/handlers.py","old_contents":"import json\n\nimport requests\n\nfrom flask import request, jsonify\n\nfrom dadd.worker import app\nfrom dadd.worker.proc import ChildProcess\n\n\n@app.route('\/run\/', methods=['POST'])\ndef run_process():\n proc = ChildProcess(request.json)\n proc.run()\n return jsonify(proc.info())\n\n\n@app.route('\/register\/', methods=['POST'])\ndef register_with_master():\n register(app)\n return jsonify({'message': 'ok'})\n\n\ndef register(host, port):\n sess = requests.Session()\n\n if 'USERNAME' in app.config and 'PASSWORD' in app.config:\n sess.auth = (app.config['USERNAME'], app.config['PASSWORD'])\n sess.headers = {'content-type': 'application\/json'}\n\n try:\n url = app.config['MASTER_URL'] + '\/api\/hosts\/'\n resp = sess.post(url, data=json.dumps({\n 'host': host, 'port': port\n }))\n if not resp.ok:\n app.logger.warning('Error registering with master: %s' %\n app.config['MASTER_URL'])\n except Exception as e:\n app.logger.warning('Connection Error: %s' % e)\n","new_contents":"import json\nimport socket\n\nimport requests\n\nfrom flask import request, jsonify\n\nfrom dadd.worker import app\nfrom dadd.worker.proc import ChildProcess\n\n\n@app.route('\/run\/', methods=['POST'])\ndef run_process():\n proc = ChildProcess(request.json)\n proc.run()\n return jsonify(proc.info())\n\n\n@app.route('\/register\/', methods=['POST'])\ndef register_with_master():\n register(app)\n return jsonify({'message': 'ok'})\n\n\ndef register(host, port):\n sess = requests.Session()\n\n if 'USERNAME' in app.config and 'PASSWORD' in app.config:\n sess.auth = (app.config['USERNAME'], app.config['PASSWORD'])\n sess.headers = {'content-type': 'application\/json'}\n\n try:\n url = app.config['MASTER_URL'] + '\/api\/hosts\/'\n resp = sess.post(url, data=json.dumps({\n 'host': socket.getfqdn(), 'port': port\n }))\n if not resp.ok:\n app.logger.warning('Error registering with master: %s' %\n app.config['MASTER_URL'])\n except Exception as e:\n app.logger.warning('Connection Error: %s' % e)\n","subject":"Use the fqdn when registering with the master.","message":"Use the fqdn when registering with the master.\n\nNot all deployment systems will provide a specific hostname via an env\nvar so we'll avoid relying on it by asking the machine.\n","lang":"Python","license":"bsd-3-clause","repos":"ionrock\/dadd,ionrock\/dadd,ionrock\/dadd,ionrock\/dadd"} {"commit":"2a76368054599006c8f7833cda1ec20f85bfcb28","old_file":"hash_table.py","new_file":"hash_table.py","old_contents":"#!\/usr\/bin\/env python\n\n'''Implementation of a simple hash table.\nThe table has `hash`, `get` and `set` methods.\nThe hash function uses a very basic hash algorithm to insert the value\ninto the table.\n'''\n\n\nclass HashItem(object):\n def __init__(self):\n pass\n\n\nclass Hash(object):\n def __init__(self, size=1024):\n self.table = []\n for i in range(size):\n self.table.append(list())\n\n def hash(self):\n pass\n\n def get(self):\n pass\n\n def set(self):\n pass\n","new_contents":"#!\/usr\/bin\/env python\n\n'''Implementation of a simple hash table.\nThe table has `hash`, `get` and `set` methods.\nThe hash function uses a very basic hash algorithm to insert the value\ninto the table.\n'''\n\n\nclass HashItem(object):\n def __init__(self):\n pass\n\n\nclass Hash(object):\n def __init__(self, size=1024):\n self.table = []\n for i in range(size):\n self.table.append(list())\n\n def hash(self, key):\n hash_value = 0\n for i in key:\n hash_value += ord(key)\n return hash_value % len(self.table)\n\n def get(self):\n pass\n\n def set(self):\n pass\n","subject":"Build out hash function of hash table class","message":"Build out hash function of hash table class\n","lang":"Python","license":"mit","repos":"jwarren116\/data-structures-deux"} {"commit":"b0824da73317bae42cb39fad5cfc95574548594a","old_file":"accounts\/models.py","new_file":"accounts\/models.py","old_contents":"# coding: utf-8\n\nfrom __future__ import unicode_literals\nfrom django.contrib.auth.models import AbstractUser, UserManager\nfrom django.db.models import BooleanField\nfrom django.utils.encoding import python_2_unicode_compatible\nfrom django.utils.translation import ungettext_lazy\nfrom mptt.fields import TreeForeignKey\nfrom mptt.managers import TreeManager\nfrom mptt.models import MPTTModel\nfrom cache_tools import cached_ugettext_lazy as _\n\n\nclass HierarchicUserManager(TreeManager, UserManager):\n pass\n\n\n@python_2_unicode_compatible\nclass HierarchicUser(MPTTModel, AbstractUser):\n mentor = TreeForeignKey(\n 'self', null=True, blank=True, related_name='disciples',\n verbose_name=_('mentor'),\n limit_choices_to={'willing_to_be_mentor__exact': True})\n willing_to_be_mentor = BooleanField(\n _('Veut être mentor'), default=False)\n\n objects = HierarchicUserManager()\n\n class MPTTMeta(object):\n parent_attr = 'mentor'\n order_insertion_by = ('username',)\n\n class Meta(object):\n verbose_name = ungettext_lazy('utilisateur', 'utilisateurs', 1)\n verbose_name_plural = ungettext_lazy('utilisateur', 'utilisateurs', 2)\n\n def __str__(self):\n return self.get_full_name() or self.get_username()\n","new_contents":"# coding: utf-8\n\nfrom __future__ import unicode_literals\nfrom django.contrib.auth.models import AbstractUser, UserManager\nfrom django.db.models import BooleanField\nfrom django.utils.encoding import python_2_unicode_compatible\nfrom django.utils.translation import ungettext_lazy\nfrom mptt.fields import TreeForeignKey\nfrom mptt.managers import TreeManager\nfrom mptt.models import MPTTModel\nfrom cache_tools import cached_ugettext_lazy as _\n\n\nclass HierarchicUserManager(TreeManager, UserManager):\n pass\n\n\n@python_2_unicode_compatible\nclass HierarchicUser(MPTTModel, AbstractUser):\n mentor = TreeForeignKey(\n 'self', null=True, blank=True, related_name='disciples',\n verbose_name=_('mentor'),\n limit_choices_to={'willing_to_be_mentor__exact': True})\n willing_to_be_mentor = BooleanField(\n _('Veut être mentor'), default=False)\n\n objects = HierarchicUserManager()\n\n class MPTTMeta(object):\n parent_attr = 'mentor'\n order_insertion_by = ('last_name', 'first_name', 'username')\n\n class Meta(object):\n verbose_name = ungettext_lazy('utilisateur', 'utilisateurs', 1)\n verbose_name_plural = ungettext_lazy('utilisateur', 'utilisateurs', 2)\n\n def __str__(self):\n return self.get_full_name() or self.get_username()\n","subject":"Change l'ordre d'insertion des utilisateurs.","message":"Change l'ordre d'insertion des utilisateurs.\n","lang":"Python","license":"bsd-3-clause","repos":"dezede\/dezede,dezede\/dezede,dezede\/dezede,dezede\/dezede"} {"commit":"c568c4b51f7a20524fb8c5a6184b8fdacb25d613","old_file":"src\/lib\/ghostlines\/windows\/account_window.py","new_file":"src\/lib\/ghostlines\/windows\/account_window.py","old_contents":"from ghostlines.storage.app_storage import AppStorage\nfrom ghostlines.windows.account_details_window import AccountDetailsWindow\nfrom ghostlines.windows.sign_in_window import SignInWindow\n\n\nclass AccountWindow(object):\n def __init__(self, sign_in=SignInWindow, account_details=AccountDetailsWindow):\n if self.is_logged_in:\n self.window = account_details(logout_window=sign_in)\n else:\n self.window = sign_in(success_window=account_details)\n\n @property\n def is_logged_in(self):\n token = AppStorage(\"pm.ghostlines.ghostlines.access_token\").retrieve()\n # TODO: Retrieve returns NSNull if set to None. Empty string used to clear password for now.\n return token != '' and token is not None\n\n def open(self):\n self.window.open()\n\n\nAccountWindow().open()\n","new_contents":"from ghostlines.storage.app_storage import AppStorage\nfrom ghostlines.windows.account_details_window import AccountDetailsWindow\nfrom ghostlines.windows.sign_in_window import SignInWindow\n\n\nclass AccountWindow(object):\n def __init__(self, sign_in=SignInWindow, account_details=AccountDetailsWindow):\n if self.is_logged_in:\n self.window = account_details(logout_window=sign_in)\n else:\n self.window = sign_in(success_window=account_details)\n\n @property\n def is_logged_in(self):\n token = AppStorage(\"pm.ghostlines.ghostlines.access_token\").retrieve()\n # TODO: Retrieve returns NSNull if set to None. Empty string used to clear password for now.\n return token != '' and token is not None\n\n def open(self):\n self.window.open()\n","subject":"Remove call to open account window when loading module","message":"Remove call to open account window when loading module\n","lang":"Python","license":"mit","repos":"ghostlines\/ghostlines-robofont,ghostlines\/ghostlines-robofont,ghostlines\/ghostlines-robofont,ghostlines\/ghostlines-robofont"} {"commit":"9ffc56e947dea40cd49c76beada2ec469a01f8f8","old_file":"__init__.py","new_file":"__init__.py","old_contents":"import base64\r\nimport json\r\nfrom os import path\r\nimport sys\r\n\r\nsys.path.insert(0, path.dirname(path.dirname(path.abspath(__file__))))\r\n\r\napi_file = 'my_api.json'\r\n_api_file = '{}\\{}'.format(path.dirname(path.abspath(__file__)), api_file)\r\n\r\nwith open(_api_file) as fin:\r\n cw_api_settings = json.load(fin)\r\nAPI_URL = cw_api_settings['API_URL']\r\n_cid = cw_api_settings['COMPANYID']\r\n_pubk = cw_api_settings['PUBLICKEY']\r\n_privk = cw_api_settings['PRIVATEKEY']\r\nbasic_auth = base64.b64encode(\"{}+{}:{}\".format(_cid, _pubk, _privk).encode('utf-8'))\r\nbasic_auth = {'Authorization': 'Basic {}'.format(str(basic_auth, 'utf-8'))}\r\n","new_contents":"import base64\r\nimport json\r\nfrom os import path\r\nimport sys\r\n\r\nsys.path.insert(0, path.dirname(path.dirname(path.abspath(__file__))))\r\n\r\napi_file = 'my_api.json'\r\n_api_file = path.join(path.dirname(path.abspath(__file__)), api_file)\r\n\r\nwith open(_api_file) as fin:\r\n cw_api_settings = json.load(fin)\r\nAPI_URL = cw_api_settings['API_URL']\r\n_cid = cw_api_settings['COMPANYID']\r\n_pubk = cw_api_settings['PUBLICKEY']\r\n_privk = cw_api_settings['PRIVATEKEY']\r\nbasic_auth = base64.b64encode(\"{}+{}:{}\".format(_cid, _pubk, _privk).encode('utf-8'))\r\nbasic_auth = {'Authorization': 'Basic {}'.format(str(basic_auth, 'utf-8'))}\r\n","subject":"Make api file path OS safe","message":"Make api file path OS safe\n","lang":"Python","license":"mit","repos":"joshuamsmith\/ConnectPyse"} {"commit":"f99847f363eed36713f657a4cb15a103ffcc6623","old_file":"web\/server.py","new_file":"web\/server.py","old_contents":"import http.client\nimport os\n\nfrom flask import Flask\nfrom pymongo import MongoClient\n\nMONGO_URL = os.environ.get('MONGO_URL', 'mongodb:\/\/mongo:27017\/')\nMONGO_DATABASE = os.environ.get('MONGO_DATABASE', 'whistleblower')\nDATABASE = MongoClient(MONGO_URL)[MONGO_DATABASE]\napp = Flask(__name__)\n\n\n@app.route('\/')\ndef hello_world():\n return 'Hello, World!'\n\n\n@app.route('\/facebook_webhook', methods=['POST'])\ndef facebook_webhook():\n DATABASE.facebook_webhook.insert(request.form)\n return ('', http.client.NO_CONTENT)\n","new_contents":"import http.client\nimport os\n\nfrom flask import Flask\nfrom pymongo import MongoClient\n\nMONGO_URL = os.environ.get('MONGO_URL', 'mongodb:\/\/mongo:27017\/')\nMONGO_DATABASE = os.environ.get('MONGO_DATABASE', 'whistleblower')\nDATABASE = MongoClient(MONGO_URL)[MONGO_DATABASE]\napp = Flask(__name__)\n\n\n@app.route('\/')\ndef hello_world():\n return 'Hello, World!'\n\n\n# @app.route('\/facebook_webhook', methods=['POST'])\n@app.route('\/facebook_webhook')\ndef facebook_webhook():\n DATABASE.facebook_webhook.insert(request.form)\n return ('', http.client.NO_CONTENT)\n","subject":"Save every request coming in the facebook webroot endpoint","message":"Save every request coming in the facebook webroot endpoint\n","lang":"Python","license":"unlicense","repos":"datasciencebr\/whistleblower"} {"commit":"483cf7f91a89e040184bd71a0a1c59c0e0926e34","old_file":"elasticmapping\/types.py","new_file":"elasticmapping\/types.py","old_contents":"# ElasticMapping\n# File: types.py\n# Desc: base Elasticsearch types\n\n\nclass CallableDict(dict):\n BASE = None\n OVERRIDES = None\n\n def __call__(self, overrides):\n new_dict = CallableDict(self)\n new_dict.OVERRIDES = overrides\n new_dict.BASE = self\n return new_dict\n\n\nBASE_TYPE = {\n 'store': False,\n 'doc_values': False\n}\n\nSTRING = CallableDict({\n 'type': 'string',\n 'index': 'analyzed'\n})\n\nFLOAT = CallableDict({\n 'type': 'float'\n})\n\nDOUBLE = CallableDict({\n 'type': 'double'\n})\n\nINTEGER = CallableDict({\n 'type': 'integer'\n})\n\nLONG = CallableDict({\n 'type': 'long'\n})\n\nSHORT = CallableDict({\n 'type': 'short'\n})\n\nBYTE = CallableDict({\n 'type': 'byte'\n})\n\nBOOLEAN = CallableDict({\n 'type': 'boolean'\n})\n\nDATE = CallableDict({\n 'type': 'date',\n 'format': 'date'\n})\n\nDATETIME = CallableDict({\n 'type': 'date',\n 'format': 'date_hour_minute_second_fraction'\n})\n\n\nTYPES = {\n name: type\n for name, type in locals().items()\n if isinstance(type, CallableDict)\n}\n","new_contents":"# ElasticMapping\n# File: types.py\n# Desc: base Elasticsearch types\n\n\nclass CallableDict(dict):\n BASE = None\n OVERRIDES = None\n\n def __call__(self, overrides):\n new_dict = CallableDict(self)\n new_dict.OVERRIDES = overrides\n new_dict.BASE = self\n return new_dict\n\n\nBASE_TYPE = {\n 'store': False,\n 'doc_values': False\n}\n\nSTRING = CallableDict({\n 'type': 'string',\n 'index': 'analyzed'\n})\n\nFLOAT = CallableDict({\n 'type': 'float'\n})\n\nDOUBLE = CallableDict({\n 'type': 'double'\n})\n\nINTEGER = CallableDict({\n 'type': 'integer'\n})\n\nLONG = CallableDict({\n 'type': 'long'\n})\n\nSHORT = CallableDict({\n 'type': 'short'\n})\n\nBYTE = CallableDict({\n 'type': 'byte'\n})\n\nBOOLEAN = CallableDict({\n 'type': 'boolean'\n})\n\nDATE = CallableDict({\n 'type': 'date',\n 'format': 'date'\n})\n\nTIME = CallableDict({\n 'type': 'date',\n 'format': 'time'\n})\n\nDATETIME = CallableDict({\n 'type': 'date',\n 'format': 'date_optional_time'\n})\n\n\nTYPES = {\n name: type\n for name, type in locals().items()\n if isinstance(type, CallableDict)\n}\n","subject":"Switch default to actual ES default (date_optional_time) and add TIME type","message":"Switch default to actual ES default (date_optional_time) and add TIME type\n","lang":"Python","license":"mit","repos":"Fizzadar\/ElasticMapping,Fizzadar\/ElasticMapping"} {"commit":"a5609a81b25425fdff1ac8aab120d8d6a4313f8b","old_file":"__init__.py","new_file":"__init__.py","old_contents":"\"\"\"distutils\n\nThe main package for the Python Module Distribution Utilities. Normally\nused from a setup script as\n\n from distutils.core import setup\n\n setup (...)\n\"\"\"\n\n# Distutils version\n#\n# Updated automatically by the Python release process.\n#\n#--start constants--\n__version__ = \"3.3.0rc3\"\n#--end constants--\n","new_contents":"\"\"\"distutils\n\nThe main package for the Python Module Distribution Utilities. Normally\nused from a setup script as\n\n from distutils.core import setup\n\n setup (...)\n\"\"\"\n\n# Distutils version\n#\n# Updated automatically by the Python release process.\n#\n#--start constants--\n__version__ = \"3.3.0\"\n#--end constants--\n","subject":"Bump version to 3.3.0 final.","message":"Bump version to 3.3.0 final.\n","lang":"Python","license":"mit","repos":"pypa\/setuptools,pypa\/setuptools,pypa\/setuptools"} {"commit":"b3bae8e48618e487ce9c8a90a555d5c6d6664872","old_file":"app\/management\/commands\/cleanapptables.py","new_file":"app\/management\/commands\/cleanapptables.py","old_contents":"from django.core.management.base import BaseCommand, CommandError\nfrom app.models import Author, Location, AutoComment, Comment, Idea, Vote\n\n\nclass Command(BaseCommand):\n def handle(self, *args, **options):\n self.stdout.write('Starting to clean app tables...')\n try:\n Idea.objects.all().delete()\n self.stdout.write('Ideas deleted')\n Comment.objects.all().delete()\n self.stdout.write('Comments deleted')\n Vote.objects.all().delete()\n self.stdout.write('Votes deleted')\n Location.objects.all().delete()\n self.stdout.write('Locations deleted')\n Author.objects.all().delete()\n self.stdout.write('Authors deleted')\n AutoComment.objects.all().delete()\n self.stdout.write('Automatic Comments deleted')\n except Exception as e:\n raise CommandError('The cleaning procedure couldn\\'t finished. Error {}'.format(e))\n\n self.stdout.write('The procedure has finished successfully...')\n\n\n\n\n\n","new_contents":"from django.core.management.base import BaseCommand, CommandError\nfrom app.models import Author, Location, AutoComment, Comment, Idea, Vote, SocialNetworkAppUser\n\n\nclass Command(BaseCommand):\n def handle(self, *args, **options):\n self.stdout.write('Starting to clean app tables...')\n try:\n Idea.objects.all().delete()\n self.stdout.write('Ideas were deleted')\n Comment.objects.all().delete()\n self.stdout.write('Comments were deleted')\n Vote.objects.all().delete()\n self.stdout.write('Votes were deleted')\n Location.objects.all().delete()\n self.stdout.write('Locations were deleted')\n Author.objects.all().delete()\n self.stdout.write('Authors were deleted')\n AutoComment.objects.all().delete()\n self.stdout.write('Automatic Comments were deleted')\n SocialNetworkAppUser.objects.all().delete()\n self.stdout.write('App users were deleted')\n except Exception as e:\n raise CommandError('The cleaning procedure couldn\\'t finished. Error {}'.format(e))\n\n self.stdout.write('The procedure has finished successfully...')\n\n\n\n\n\n","subject":"Add the deletion of the app users","message":"Add the deletion of the app users\n","lang":"Python","license":"mit","repos":"joausaga\/social-ideation,joausaga\/social-ideation,rebearteta\/social-ideation,rebearteta\/social-ideation,rebearteta\/social-ideation,joausaga\/social-ideation,joausaga\/social-ideation,rebearteta\/social-ideation"} {"commit":"cde145c95f604cd1936fe0d61e6f6b5858177a80","old_file":"app\/util\/danger.py","new_file":"app\/util\/danger.py","old_contents":"from itsdangerous import TimedJSONWebSignatureSerializer as Serializer\nfrom flask import request as flask_request\nfrom flask import abort\n\nimport logging\nimport os\n\ndef gen_auth_token(id,expiration=10000):\n \"\"\"Generate auth token\"\"\"\n try:\n s = Serializer(os.environ['API_KEY'],expires_in=expiration)\n except KeyError:\n logging.fatal(\"No API_KEY env\")\n abort(500)\n\n return s.dumps({'id':id})\n\ndef verify_auth_token(token):\n \"\"\"Verify auth token\"\"\"\n try:\n s = Serializer(os.environ['API_KEY'])\n except KeyError:\n logging.fatal(\"No API_KEY env\")\n abort(500)\n # check the token and throw respective exception\n try:\n user = s.loads(token)\n except Exception as e:\n logging.info(e)\n abort(401)\n return user\n\ndef enable_auth(func):\n \"\"\"Decorator to enable auth\"\"\"\n def wrapper(*args,**kwargs):\n re = flask_request\n # deny if not authorized\n if not re.headers.has_key(\"Authorization\"):\n logging.info(\"No token found\")\n abort(401)\n auth = re.headers.get(\"Authorization\").split(\" \")\n # proces token \n validate = verify_auth_token(auth[1])\n logging.debug(\"Valid auth! Yay\")\n return func(*args,**kwargs)\n return wrapper","new_contents":"from itsdangerous import TimedJSONWebSignatureSerializer as Serializer\nfrom flask import request as flask_request\nfrom flask import abort\n\nimport logging\nimport os\n\ndef gen_auth_token(id,expiration=10000):\n \"\"\"Generate auth token\"\"\"\n try:\n s = Serializer(os.environ['API_KEY'],expires_in=expiration)\n except KeyError:\n logging.fatal(\"No API_KEY env\")\n abort(500)\n\n return s.dumps({'id':id})\n\ndef verify_auth_token(token):\n \"\"\"Verify auth token\"\"\"\n try:\n s = Serializer(os.environ['API_KEY'])\n except KeyError:\n logging.fatal(\"No API_KEY env\")\n abort(500)\n # check the token and throw respective exception\n try:\n user = s.loads(token)\n except Exception as e:\n logging.warning(\"Bad token\")\n abort(401)\n return user\n\ndef enable_auth(func):\n \"\"\"Decorator to enable auth\"\"\"\n def wrapper(*args,**kwargs):\n re = flask_request\n # deny if not authorized\n if not re.headers.has_key(\"Authorization\"):\n logging.warning(\"No token found\")\n abort(401)\n auth = re.headers.get(\"Authorization\").split(\" \")\n # proces token \n validate = verify_auth_token(auth[1])\n logging.debug(\"Valid auth! Yay\")\n return func(*args,**kwargs)\n return wrapper","subject":"Add warning log level for auth checks","message":"Add warning log level for auth checks\n","lang":"Python","license":"mit","repos":"tforrest\/soda-automation,tforrest\/soda-automation"} {"commit":"c9003940c583a19861c1dff20498aa4c6aae1efb","old_file":"scikits\/crab\/tests\/test_base.py","new_file":"scikits\/crab\/tests\/test_base.py","old_contents":"#-*- coding:utf-8 -*-\n\n\"\"\" \nBase Recommender Models.\n\"\"\"\n\n# Authors: Marcel Caraciolo <marcel@muricoca.com>\n# Bruno Melo <bruno@muricoca.com>\n# License: BSD Style.\n\nimport unittest\nimport sys\n\nsys.path.append('\/Users\/marcelcaraciolo\/Desktop\/Orygens\/crab\/crab\/scikits\/craba')\n\nfrom base import BaseRecommender\n\n#test classes\n\nclass MyRecommender(BaseRecommender):\n def __init__(self,model):\n BaseRecommender.__init__(self,model)\n \n################################################################################\n# The tests\n\n\nclass testBaseRecommender(unittest.TestCase):\n pass\n\n\nif __name__ == '__main__':\n unittest.main()\n\n","new_contents":"#-*- coding:utf-8 -*-\n\n\"\"\" \nBase Recommender Models.\n\"\"\"\n\n# Authors: Marcel Caraciolo <marcel@muricoca.com>\n# Bruno Melo <bruno@muricoca.com>\n# License: BSD Style.\n\nimport unittest\n\nfrom base import BaseRecommender\n\n#test classes\n\nclass MyRecommender(BaseRecommender):\n def __init__(self,model):\n BaseRecommender.__init__(self,model)\n \n################################################################################\n# The tests\n\n\nclass testBaseRecommender(unittest.TestCase):\n pass\n\n\nif __name__ == '__main__':\n unittest.main()\n\n","subject":"Fix the test removing paths.","message":"Fix the test removing paths.\n","lang":"Python","license":"bsd-3-clause","repos":"Lawrence-Liu\/crab,muricoca\/crab,hi2srihari\/crab,echogreens\/crab,muricoca\/crab,augustoppimenta\/crab,imouren\/crab,rcarmo\/crab,Flowerowl\/Crab,wnyc\/crab,wnyc\/crab"} {"commit":"b2e4b0eaf67714c180feec53e5b0021d911d8f8d","old_file":"pandora\/py2compat.py","new_file":"pandora\/py2compat.py","old_contents":"\"\"\"\nPython 2 Compatibility Layer\n\nThis module exists to work around compatibility issues between Python 2 and\nPython 3. The main code-base will use Python 3 idioms and this module will\npatch Python 2 code to support those changes. When Python 2 support is\ndropped this module can be removed and imports can be updated.\n\"\"\"\n\ntry:\n from configparser import ConfigParser\nexcept ImportError:\n from ConfigParser import SafeConfigParser\n\n class ConfigParser(SafeConfigParser):\n\n def read_file(self, fp):\n return self.readfp(fp)\n\n\n# Only used in tests\ntry:\n from unittest.mock import Mock, MagicMock, call\nexcept ImportError:\n try:\n from mock import Mock, MagicMock, call, patch\n except ImportError:\n pass\n","new_contents":"\"\"\"\nPython 2 Compatibility Layer\n\nThis module exists to work around compatibility issues between Python 2 and\nPython 3. The main code-base will use Python 3 idioms and this module will\npatch Python 2 code to support those changes. When Python 2 support is\ndropped this module can be removed and imports can be updated.\n\"\"\"\n\ntry:\n from configparser import ConfigParser\nexcept ImportError:\n from ConfigParser import SafeConfigParser\n\n class ConfigParser(SafeConfigParser):\n\n def read_file(self, fp):\n return self.readfp(fp)\n\n\n# Only used in tests\ntry:\n from unittest.mock import Mock, MagicMock, call, patch\nexcept ImportError:\n try:\n from mock import Mock, MagicMock, call, patch\n except ImportError:\n pass\n","subject":"Add missing import for unittest.mock.patch.","message":"Add missing import for unittest.mock.patch.\n","lang":"Python","license":"mit","repos":"mcrute\/pydora"} {"commit":"84dd763d5d2aec1c4248e42106ef4f68439bc4cd","old_file":"server\/api\/serializers\/rides.py","new_file":"server\/api\/serializers\/rides.py","old_contents":"import requests\nfrom django.conf import settings\nfrom rest_framework import serializers\nfrom rest_framework.validators import UniqueTogetherValidator\nfrom server.api.serializers.chapters import ChapterSerializer\nfrom server.core.models.rides import Ride, RideRiders\n\n\nclass RideSerializer(serializers.ModelSerializer):\n chapter = ChapterSerializer()\n riders = serializers.PrimaryKeyRelatedField(source='registered_riders', many=True, read_only=True)\n\n class Meta:\n model = Ride\n fields = ('id', 'name', 'slug', 'description_html', 'start_location', 'end_location', 'start_date', 'end_date',\n 'chapter', 'rider_capacity', 'riders', 'spaces_left', 'price', 'full_cost', 'currency', 'is_over',\n 'fundraising_total', 'fundraising_target')\n\n\nclass RideRiderSerializer(serializers.ModelSerializer):\n user = serializers.PrimaryKeyRelatedField(read_only=True, default=serializers.CurrentUserDefault())\n\n class Meta:\n model = RideRiders\n fields = ('ride', 'user', 'signup_date', 'signup_expires', 'status', 'paid', 'expired', 'payload')\n validators = [\n UniqueTogetherValidator(\n queryset=RideRiders.objects.all(),\n fields=('user', 'ride'),\n message='You have already registered for this ride.'\n )\n ]","new_contents":"import requests\nfrom django.conf import settings\nfrom rest_framework import serializers\nfrom rest_framework.validators import UniqueTogetherValidator\nfrom server.api.serializers.chapters import ChapterSerializer\nfrom .riders import RiderSerializer\nfrom server.core.models.rides import Ride, RideRiders\n\n\nclass RideSerializer(serializers.ModelSerializer):\n chapter = ChapterSerializer()\n riders = RiderSerializer(source='registered_riders', many=True, read_only=True)\n\n class Meta:\n model = Ride\n fields = ('id', 'name', 'slug', 'description_html', 'start_location', 'end_location', 'start_date', 'end_date',\n 'chapter', 'rider_capacity', 'riders', 'spaces_left', 'price', 'full_cost', 'currency', 'is_over',\n 'fundraising_total', 'fundraising_target')\n\n\nclass RideRiderSerializer(serializers.ModelSerializer):\n user = serializers.PrimaryKeyRelatedField(read_only=True, default=serializers.CurrentUserDefault())\n\n class Meta:\n model = RideRiders\n fields = ('ride', 'user', 'signup_date', 'signup_expires', 'status', 'paid', 'expired', 'payload')\n validators = [\n UniqueTogetherValidator(\n queryset=RideRiders.objects.all(),\n fields=('user', 'ride'),\n message='You have already registered for this ride.'\n )\n ]","subject":"Send through serialised users with each ride api request","message":"Send through serialised users with each ride api request\n","lang":"Python","license":"mit","repos":"mwillmott\/techbikers,mwillmott\/techbikers,mwillmott\/techbikers,Techbikers\/techbikers,Techbikers\/techbikers,Techbikers\/techbikers,Techbikers\/techbikers,mwillmott\/techbikers"} {"commit":"823347e9c5bcc93710b7bdc2656f438a0e74c2b8","old_file":"invocations\/_version.py","new_file":"invocations\/_version.py","old_contents":"__version_info__ = (0, 14, 0)\n__version__ = '.'.join(map(str, __version_info__))\n","new_contents":"__version_info__ = (0, 14, 1)\n__version__ = '.'.join(map(str, __version_info__))\n","subject":"Cut 0.14.1 because setup.py bug","message":"Cut 0.14.1 because setup.py bug\n","lang":"Python","license":"bsd-2-clause","repos":"pyinvoke\/invocations"} {"commit":"12bbc7e10ae52328feb766e2bed5f5f20fa0d354","old_file":"pyramid_es\/__init__.py","new_file":"pyramid_es\/__init__.py","old_contents":"from pyramid.settings import asbool\n\nfrom .client import ElasticClient\n\n\ndef client_from_config(settings, prefix='elastic.'):\n \"\"\"\n Instantiate and configure an Elasticsearch from settings.\n\n In typical Pyramid usage, you shouldn't use this directly: instead, just\n include ``pyramid_es`` and use the :py:func:`get_client` function to get\n access to the shared :py:class:`.client.ElasticClient` instance.\n \"\"\"\n return ElasticClient(\n servers=settings.get(prefix + 'servers', ['localhost:9200']),\n timeout=settings.get(prefix + 'timeout', 1.0),\n index=settings[prefix + 'index'],\n use_transaction=asbool(settings.get(prefix + 'use_transaction', True)),\n disable_indexing=settings.get(prefix + 'disable_indexing', False))\n\n\ndef includeme(config):\n registry = config.registry\n settings = registry.settings\n\n client = client_from_config(settings)\n client.ensure_index()\n\n registry.pyramid_es_client = client\n\n\ndef get_client(request):\n \"\"\"\n Get the registered Elasticsearch client. The supplied argument can be\n either a ``Request`` instance or a ``Registry``.\n \"\"\"\n registry = getattr(request, 'registry', None)\n if registry is None:\n registry = request\n return registry.pyramid_es_client\n","new_contents":"from pyramid.settings import asbool\n\nfrom .client import ElasticClient\n\n\ndef client_from_config(settings, prefix='elastic.'):\n \"\"\"\n Instantiate and configure an Elasticsearch from settings.\n\n In typical Pyramid usage, you shouldn't use this directly: instead, just\n include ``pyramid_es`` and use the :py:func:`get_client` function to get\n access to the shared :py:class:`.client.ElasticClient` instance.\n \"\"\"\n return ElasticClient(\n servers=settings.get(prefix + 'servers', ['localhost:9200']),\n timeout=settings.get(prefix + 'timeout', 1.0),\n index=settings[prefix + 'index'],\n use_transaction=asbool(settings.get(prefix + 'use_transaction', True)),\n disable_indexing=settings.get(prefix + 'disable_indexing', False))\n\n\ndef includeme(config):\n registry = config.registry\n settings = registry.settings\n\n client = client_from_config(settings)\n if asbool(settings.get('elastic.ensure_index_on_start')):\n client.ensure_index()\n\n registry.pyramid_es_client = client\n\n\ndef get_client(request):\n \"\"\"\n Get the registered Elasticsearch client. The supplied argument can be\n either a ``Request`` instance or a ``Registry``.\n \"\"\"\n registry = getattr(request, 'registry', None)\n if registry is None:\n registry = request\n return registry.pyramid_es_client\n","subject":"Add a settings key to ensure index at start","message":"Add a settings key to ensure index at start\n","lang":"Python","license":"mit","repos":"storborg\/pyramid_es"} {"commit":"bfecf498c30c08d8ede18fd587e192f0961c334c","old_file":"invoke\/run.py","new_file":"invoke\/run.py","old_contents":"from subprocess import PIPE\n\nfrom .monkey import Popen\nfrom .exceptions import Failure\n\n\nclass Result(object):\n def __init__(self, stdout=None, stderr=None, exited=None):\n self.exited = self.return_code = exited\n self.stdout = stdout\n self.stderr = stderr\n\n def __nonzero__(self):\n # Holy mismatch between name and implementation, Batman!\n return self.exited == 0\n\n\ndef run(command, warn=False):\n \"\"\"\n Execute ``command`` in a local subprocess.\n\n By default, raises an exception if the subprocess terminates with a nonzero\n return code. This may be disabled by setting ``warn=True``.\n \"\"\"\n process = Popen(command,\n shell=True,\n stdout=PIPE,\n stderr=PIPE\n )\n stdout, stderr = process.communicate()\n result = Result(stdout=stdout, stderr=stderr, exited=process.returncode)\n if not (result or warn):\n raise Failure(result)\n return result\n","new_contents":"from subprocess import PIPE\n\nfrom .monkey import Popen\nfrom .exceptions import Failure\n\n\nclass Result(object):\n def __init__(self, stdout=None, stderr=None, exited=None):\n self.exited = self.return_code = exited\n self.stdout = stdout\n self.stderr = stderr\n\n def __nonzero__(self):\n # Holy mismatch between name and implementation, Batman!\n return self.exited == 0\n\n def __str__(self):\n ret = [\"Command exited with status %s.\" % self.exited]\n for x in ('stdout', 'stderr'):\n val = getattr(self, x)\n ret.append(\"\"\"=== %s ===\n%s\n\"\"\" % (x, val.rstrip()) if val else \"(no %s)\" % x)\n return \"\\n\".join(ret)\n\ndef run(command, warn=False):\n \"\"\"\n Execute ``command`` in a local subprocess.\n\n By default, raises an exception if the subprocess terminates with a nonzero\n return code. This may be disabled by setting ``warn=True``.\n \"\"\"\n process = Popen(command,\n shell=True,\n stdout=PIPE,\n stderr=PIPE\n )\n stdout, stderr = process.communicate()\n result = Result(stdout=stdout, stderr=stderr, exited=process.returncode)\n if not (result or warn):\n raise Failure(result)\n return result\n","subject":"Add semi-useful `__str__` for Result","message":"Add semi-useful `__str__` for Result\n","lang":"Python","license":"bsd-2-clause","repos":"pyinvoke\/invoke,mkusz\/invoke,singingwolfboy\/invoke,tyewang\/invoke,mattrobenolt\/invoke,kejbaly2\/invoke,sophacles\/invoke,pfmoore\/invoke,pyinvoke\/invoke,mkusz\/invoke,pfmoore\/invoke,kejbaly2\/invoke,alex\/invoke,frol\/invoke,mattrobenolt\/invoke,frol\/invoke"} {"commit":"87d099f8094d5fb2c78729adfc6df9c68f68b450","old_file":"pythonforandroid\/recipes\/regex\/__init__.py","new_file":"pythonforandroid\/recipes\/regex\/__init__.py","old_contents":"from pythonforandroid.recipe import CompiledComponentsPythonRecipe\n\n\nclass RegexRecipe(CompiledComponentsPythonRecipe):\n name = 'regex'\n version = '2017.07.28'\n url = 'https:\/\/pypi.python.org\/packages\/d1\/23\/5fa829706ee1d4452552eb32e0bfc1039553e01f50a8754c6f7152e85c1b\/regex-{version}.tar.gz'\n\n depends = ['setuptools']\n\n\nrecipe = RegexRecipe()\n","new_contents":"from pythonforandroid.recipe import CompiledComponentsPythonRecipe\n\n\nclass RegexRecipe(CompiledComponentsPythonRecipe):\n name = 'regex'\n version = '2017.07.28'\n url = 'https:\/\/pypi.python.org\/packages\/d1\/23\/5fa829706ee1d4452552eb32e0bfc1039553e01f50a8754c6f7152e85c1b\/regex-{version}.tar.gz'\n\n depends = ['setuptools']\n call_hostpython_via_targetpython = False\n\n\nrecipe = RegexRecipe()\n","subject":"Fix compilation for regex recipe","message":"[recipes] Fix compilation for regex recipe\n\nThe error was: build\/other_builds\/hostpython3\/desktop\/hostpython3\/Include\/Python.h:39:19: fatal error: crypt.h: No such file or directory\n","lang":"Python","license":"mit","repos":"rnixx\/python-for-android,kronenpj\/python-for-android,kivy\/python-for-android,kivy\/python-for-android,kivy\/python-for-android,kivy\/python-for-android,germn\/python-for-android,PKRoma\/python-for-android,kronenpj\/python-for-android,rnixx\/python-for-android,germn\/python-for-android,germn\/python-for-android,rnixx\/python-for-android,PKRoma\/python-for-android,kronenpj\/python-for-android,PKRoma\/python-for-android,germn\/python-for-android,PKRoma\/python-for-android,kronenpj\/python-for-android,kronenpj\/python-for-android,PKRoma\/python-for-android,rnixx\/python-for-android,rnixx\/python-for-android,kivy\/python-for-android,germn\/python-for-android,rnixx\/python-for-android,germn\/python-for-android"} {"commit":"5f5be04adc9e17aa497022ed3b19371075c63d85","old_file":"relay_api\/api\/backend.py","new_file":"relay_api\/api\/backend.py","old_contents":"import json\nfrom relay_api.core.relay import relay\nfrom relay_api.conf.config import relays\n\n\ndef init_relays():\n for r in relays:\n relays[r][\"object\"] = relay(relays[r][\"gpio\"])\n relays[r][\"state\"] = relays[r][\"object\"].get_state()\n\n\ndef get_all_relays():\n relays_dict = __get_relay_dict()\n return json.dumps(relays_dict, indent=4)\n\n\ndef get_relay(relay_name):\n if relay_name not in relays:\n return None\n relay_dict = __get_relay_dict(relay_name)\n return json.dumps(relay_dict, indent=4)\n\n\ndef __get_relay_dict(relay_name=None):\n if relay_name:\n relay_dict = dict.copy(relays[\"relay_name\"])\n del(relay_dict[\"object\"])\n return relay_dict\n relays_dict = dict.copy(relays)\n for r in relays_dict:\n del(relays_dict[r][\"object\"])\n return relays_dict\n","new_contents":"import json\nimport copy\nfrom relay_api.core.relay import relay\nfrom relay_api.conf.config import relays\n\n\ndef init_relays():\n for r in relays:\n relays[r][\"object\"] = relay(relays[r][\"gpio\"])\n relays[r][\"state\"] = relays[r][\"object\"].get_state()\n\n\ndef get_all_relays():\n relays_dict = __get_relay_dict()\n return json.dumps(relays_dict, indent=4)\n\n\ndef get_relay(relay_name):\n if relay_name not in relays:\n return None\n relay_dict = __get_relay_dict(relay_name)\n return json.dumps(relay_dict, indent=4)\n\n\ndef __get_relay_dict(relay_name=None):\n if relay_name:\n relay_dict = copy.deepcopy(relays[relay_name])\n del(relay_dict[\"object\"])\n return relay_dict\n relays_dict = copy.deepcopy(relays)\n for r in relays_dict:\n del(relays_dict[r][\"object\"])\n return relays_dict\n","subject":"Use deppcopy to copy dicts","message":"Use deppcopy to copy dicts\n","lang":"Python","license":"mit","repos":"pahumadad\/raspi-relay-api"} {"commit":"c100d61f084ea759654baf0e9414f73567577f68","old_file":"src\/robot.py","new_file":"src\/robot.py","old_contents":"from sr import *\n\nprint \"Hello, world!\"\n\n","new_contents":"import time\nfrom sr import *\n\nprint \"Hello, world!\"\n\nR = Robot()\n\n# This is the configuration for Elizabeth.\n# TODO: make this generic\nBOARD_RIGHT = R.motors[\"SR0HL17\"]\nBOARD_LEFT = R.motors[\"SR0YK1C\"]\n\nWHEEL_FRONT_LEFT = BOARD_LEFT.m1 # positive is towards the front of the robot\nWHEEL_FRONT_RIGHT = BOARD_RIGHT.m0 # positive is towards the front of the robot\nWHEEL_BACK = BOARD_RIGHT.m1 # positive is to the right of the robot\n\n# enable the brakes\nWHEEL_FRONT_LEFT.use_brake = True\nWHEEL_FRONT_RIGHT.use_brake = True\nWHEEL_BACK.use_brake = True\n\nWHEEL_FRONT_LEFT_CALIBRATION = -1\nWHEEL_FRONT_RIGHT_CALIBRATION = -1\nWHEEL_BACK_CALIBRATION = 1\n\ndef set_motors(front_left, front_right, back):\n WHEEL_FRONT_LEFT.power = int(front_left * WHEEL_FRONT_LEFT_CALIBRATION)\n WHEEL_FRONT_RIGHT.power = int(front_right * WHEEL_FRONT_RIGHT_CALIBRATION)\n WHEEL_BACK.power = int(back * WHEEL_BACK_CALIBRATION)\n\ndef forward(speed):\n set_motors(speed, speed, 0)\n\ndef reverse(speed):\n forward(-speed)\n\ndef stop():\n forward(0)\n\ndef rotate(speed):\n set_motors(speed, -speed, speed)\n\ndef can_see_block():\n ACCEPTABLE_MARKER_TYPES = (MARKER_TOKEN_TOP,\n MARKER_TOKEN_BOTTOM,\n MARKER_TOKEN_SIDE)\n markers = R.see()\n return any(marker.info.marker_type in ACCEPTABLE_MARKER_TYPES\n for marker in R.see())\n\n# FIXME: debug while we don't have a marker\nsearch_count = 0\ndef can_see_block():\n global search_count\n search_count += 1\n return search_count % 5 == 0\n\ndef state_search():\n rotate(40)\n time.sleep(0.4)\n stop()\n time.sleep(0.3)\n return state_advance if can_see_block() else state_search\n\ndef state_advance():\n forward(30)\n time.sleep(1)\n stop()\n time.sleep(0.8)\n return state_advance if can_see_block() else state_backoff\n\ndef state_backoff():\n reverse(20)\n time.sleep(2)\n stop()\n rotate(-80)\n time.sleep(4)\n stop()\n return state_search\n\ncurrent_state = state_search\nwhile True:\n current_state = current_state()\n\n","subject":"Implement a basic search-and-grab algorithm","message":"Implement a basic search-and-grab algorithm\n","lang":"Python","license":"mit","repos":"prophile\/tudor-block-chase"} {"commit":"0d4e619a11a084f83ab42d45e528f7b38777fcae","old_file":"linter.py","new_file":"linter.py","old_contents":"#\n# linter.py\n# Linter for SublimeLinter3, a code checking framework for Sublime Text 3\n#\n# Written by Andrew Grim\n# Copyright (c) 2014 Andrew Grim\n#\n# License: MIT\n#\n\n\"\"\"This module exports the Puppet plugin class.\"\"\"\n\nfrom SublimeLinter.lint import Linter, util\n\n\nclass Puppet(Linter):\n\n \"\"\"Provides an interface to puppet.\"\"\"\n\n syntax = 'puppet'\n cmd = ('puppet', 'parser', 'validate', '--color=false')\n regex = r'^(?P<error>Error:).+?(?P<message>Syntax error at \\'(?P<near>.+?)\\'; expected \\'.+\\').+?line (?P<line>\\d+)'\n error_stream = util.STREAM_STDERR\n defaults = {}\n","new_contents":"#\n# linter.py\n# Linter for SublimeLinter3, a code checking framework for Sublime Text 3\n#\n# Written by Andrew Grim\n# Copyright (c) 2014 Andrew Grim\n#\n# License: MIT\n#\n\n\"\"\"This module exports the Puppet plugin class.\"\"\"\n\nfrom SublimeLinter.lint import Linter, util\n\n\nclass Puppet(Linter):\n\n \"\"\"Provides an interface to puppet.\"\"\"\n\n syntax = 'puppet'\n cmd = ('puppet', 'parser', 'validate', '--color=false')\n regex = r'^Error:.+?(?P<message>Syntax error at \\'(?P<near>.+?)\\'; expected \\'.+\\').+?line (?P<line>\\d+)'\n error_stream = util.STREAM_STDERR\n","subject":"Remove unnecessary defaults and simplify regexp","message":"Remove unnecessary defaults and simplify regexp","lang":"Python","license":"mit","repos":"dylanratcliffe\/SublimeLinter-puppet,travisgroth\/SublimeLinter-puppet,stopdropandrew\/SublimeLinter-puppet"} {"commit":"b6b506e8250078664bdefdcf7d9d380e950e3730","old_file":"linter.py","new_file":"linter.py","old_contents":"#\n# linter.py\n# Linter for SublimeLinter3, a code checking framework for Sublime Text 3\n#\n# Written by Jack Brewer\n# Copyright (c) 2015 Jack Brewer\n#\n# License: MIT\n\n\"\"\"Exports the Stylint plugin class.\"\"\"\n\nfrom SublimeLinter.lint import NodeLinter, util\n\n\nclass Stylint(NodeLinter):\n\n \"\"\"Provides an interface to stylint.\"\"\"\n\n npm_name = 'stylint'\n syntax = ('stylus', 'vue')\n selectors = {'vue': 'source.stylus.embedded.html'}\n cmd = 'stylint @ *'\n executable = 'stylint'\n version_requirement = '>= 1.5.0'\n regex = r'''(?xi)\n # Comments show example output for each line of a Stylint warning\n # \/path\/to\/file\/example.styl\n ^.*$\\s*\n # 177:24 colors warning hexidecimal color should be a variable\n ^(?P<line>\\d+):?(?P<col>\\d+)?\\s*((?P<warning>warning)|(?P<error>error))\\s*(?P<message>.+)$\\s*\n '''\n multiline = True\n error_stream = util.STREAM_STDOUT\n tempfile_suffix = 'styl'\n config_file = ('--config', '.stylintrc', '~')\n","new_contents":"#\n# linter.py\n# Linter for SublimeLinter3, a code checking framework for Sublime Text 3\n#\n# Written by Jack Brewer\n# Copyright (c) 2015 Jack Brewer\n#\n# License: MIT\n\n\"\"\"Exports the Stylint plugin class.\"\"\"\n\nfrom SublimeLinter.lint import NodeLinter, util\n\n\nclass Stylint(NodeLinter):\n\n \"\"\"Provides an interface to stylint.\"\"\"\n\n npm_name = 'stylint'\n syntax = ('stylus', 'vue')\n selectors = {'vue': 'source.stylus.embedded.html'}\n cmd = 'stylint @ *'\n executable = 'stylint'\n version_requirement = '>= 1.5.0'\n regex = r'''(?xi)\n # Comments show example output for each line of a Stylint warning\n # \/path\/to\/file\/example.styl\n ^.*$\\s*\n # 177:24 colors warning hexidecimal color should be a variable\n ^(?P<line>\\d+):?(?P<col>\\d+)?\\s*(?P<rule>\\w+)?\\s*((?P<warning>warning)|(?P<error>error))\\s*(?P<message>.+)$\\s*\n '''\n multiline = True\n error_stream = util.STREAM_STDOUT\n tempfile_suffix = 'styl'\n config_file = ('--config', '.stylintrc', '~')\n","subject":"Handle case where rule shows before severity","message":"Handle case where rule shows before severity\n\nThank you @suprMax !\n","lang":"Python","license":"mit","repos":"jackbrewer\/SublimeLinter-contrib-stylint"} {"commit":"f7172424977d0166ec4dd7946a360a5a426f4a72","old_file":"bin\/migrate-tips.py","new_file":"bin\/migrate-tips.py","old_contents":"from gratipay.wireup import db, env\nfrom gratipay.models.team import AlreadyMigrated\n\ndb = db(env())\n\nteams = db.all(\"\"\"\n SELECT distinct ON (t.slug) t.*::teams\n FROM teams t\n JOIN tips ON t.owner = tips.tippee -- Only fetch teams whose owner have tips.\n WHERE t.is_approved IS TRUE -- Only fetch approved teams.\n AND NOT EXISTS ( -- Make sure not already migrated.\n SELECT 1\n FROM payment_instructions pi\n WHERE t.slug = pi.team\n AND pi.ctime < t.ctime\n\t )\n\"\"\")\n\nfor team in teams:\n try:\n ntips = team.migrate_tips()\n print(\"Migrated {} tip(s) for '{}'\".format(ntips, team.slug))\n except AlreadyMigrated:\n print(\"'%s' already migrated.\" % team.slug)\n\nprint(\"Done.\")\n","new_contents":"from gratipay.wireup import db, env\nfrom gratipay.models.team import AlreadyMigrated\n\ndb = db(env())\n\nteams = db.all(\"\"\"\n SELECT distinct ON (t.slug) t.*::teams\n FROM teams t\n JOIN tips ON t.owner = tips.tippee -- Only fetch teams whose owners had tips under Gratipay 1.0\n WHERE t.is_approved IS TRUE -- Only fetch approved teams\n AND NOT EXISTS ( -- Make sure tips haven't been migrated for any teams with same owner\n SELECT 1\n FROM payment_instructions pi\n JOIN teams t2 ON t2.slug = pi.team\n WHERE t2.owner = t.owner\n AND pi.ctime < t2.ctime\n )\n\"\"\")\n\nfor team in teams:\n try:\n ntips = team.migrate_tips()\n print(\"Migrated {} tip(s) for '{}'\".format(ntips, team.slug))\n except AlreadyMigrated:\n print(\"'%s' already migrated.\" % team.slug)\n\nprint(\"Done.\")\n","subject":"Exclude teams if owner has other teams with migrated tips","message":"Exclude teams if owner has other teams with migrated tips\n","lang":"Python","license":"mit","repos":"gratipay\/gratipay.com,gratipay\/gratipay.com,gratipay\/gratipay.com,gratipay\/gratipay.com"} {"commit":"0ce14be170e09530b225f2f7526ad68ee1758095","old_file":"peering\/migrations\/0027_auto_20190105_1600.py","new_file":"peering\/migrations\/0027_auto_20190105_1600.py","old_contents":"# Generated by Django 2.1.4 on 2019-01-05 15:00\n\nimport django.contrib.postgres.fields\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n (\n \"peering\",\n \"0026_autonomoussystem_potential_internet_exchange_peering_sessions\",\n )\n ]\n\n operations = [\n migrations.AlterField(\n model_name=\"autonomoussystem\",\n name=\"potential_internet_exchange_peering_sessions\",\n field=django.contrib.postgres.fields.ArrayField(\n base_field=models.GenericIPAddressField(),\n blank=True,\n default=list,\n size=None,\n ),\n )\n ]\n","new_contents":"# Generated by Django 2.1.4 on 2019-01-05 15:00\n\nimport django.contrib.postgres.fields\nfrom django.db import migrations, models\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n (\n \"peering\",\n \"0026_autonomoussystem_potential_internet_exchange_peering_sessions\",\n )\n ]\n\n def forwards_func(apps, schema_editor):\n AutonomousSystem = apps.get_model(\"peering\", \"AutonomousSystem\")\n db_alias = schema_editor.connection.alias\n AutonomousSystem.objects.using(db_alias).filter(\n potential_internet_exchange_peering_sessions=None\n ).update(potential_internet_exchange_peering_sessions=[])\n\n def reverse_func(apps, schema_editor):\n AutonomousSystem = apps.get_model(\"peering\", \"AutonomousSystem\")\n db_alias = schema_editor.connection.alias\n AutonomousSystem.objects.using(db_alias).filter(\n potential_internet_exchange_peering_sessions=[]\n ).update(potential_internet_exchange_peering_sessions=None)\n\n operations = [\n migrations.AlterField(\n model_name=\"autonomoussystem\",\n name=\"potential_internet_exchange_peering_sessions\",\n field=django.contrib.postgres.fields.ArrayField(\n base_field=models.GenericIPAddressField(),\n blank=True,\n default=list,\n size=None,\n ),\n ),\n migrations.RunPython(forwards_func, reverse_func),\n ]\n","subject":"Fix issue with migrations introduced lately.","message":"Fix issue with migrations introduced lately.\n","lang":"Python","license":"apache-2.0","repos":"respawner\/peering-manager,respawner\/peering-manager,respawner\/peering-manager,respawner\/peering-manager"} {"commit":"e8b8c257c71b6c02fa691557618261e6832fba94","old_file":"faker\/providers\/ssn\/uk_UA\/__init__.py","new_file":"faker\/providers\/ssn\/uk_UA\/__init__.py","old_contents":"# coding=utf-8\nfrom __future__ import unicode_literals\nfrom .. import Provider as SsnProvider\n\n\n# Note: as there no SSN in Ukraine\n# we get value added tax identification number (VATIN) here.\n# It is also called \"Ідентифікаційний номер платника податків\" (in ukrainian).\n# It contains only digits and length if 12.\n\n\nclass Provider(SsnProvider):\n ssn_formats = (\"############\",)\n","new_contents":"# coding=utf-8\nfrom __future__ import unicode_literals\n\nfrom datetime import date\n\nfrom .. import Provider as SsnProvider\nfrom faker.providers.date_time import Provider as DateTimeProvider\n\n\nclass Provider(SsnProvider):\n @classmethod\n def ssn(cls):\n \"\"\"\n Ukrainian \"Реєстраційний номер облікової картки платника податків\"\n also known as \"Ідентифікаційний номер фізичної особи\".\n \"\"\"\n digits = []\n\n # Number of days between 1899-12-31 and a birth date\n for digit in str((DateTimeProvider.date_object() -\n date(1899, 12, 31)).days):\n digits.append(int(digit))\n\n # Person's sequence number\n for _ in range(4):\n digits.append(cls.random_int(0, 9))\n\n checksum = (digits[0]*-1 + digits[1]*5 + digits[2]*7 + digits[3]*9 +\n digits[4]*4 + digits[5]*6 + digits[6]*10 + digits[7]*5 +\n digits[8]*7)\n # Remainder of a checksum divided by 11 or 1 if it equals to 10\n digits.append(checksum % 11 % 10)\n\n return ''.join(str(digit) for digit in digits)\n","subject":"Make the Ukrainian SSN provider realer","message":"Make the Ukrainian SSN provider realer\n\n","lang":"Python","license":"mit","repos":"joke2k\/faker,danhuss\/faker,trtd\/faker,joke2k\/faker"} {"commit":"b698f6925b4629d7473fbe42806f54068d98428a","old_file":"tests\/component\/test_component_identidock.py","new_file":"tests\/component\/test_component_identidock.py","old_contents":"import sys\n\nprint(sys.path)\n","new_contents":"import pytest\nimport requests\nfrom time import sleep\n\nCOMPONENT_INDEX_URL = \"http:\/\/identidock:5000\"\n\nCOMPONENT_MONSTER_BASE_URL = COMPONENT_INDEX_URL + '\/monster'\n\n\n\n\ndef test_get_mainpage():\n\tprint('component tester sleeping for 1 sec to let the identidock app to be ready adn also start its server')\n\tsleep(1)\n\tpage = requests.get(COMPONENT_INDEX_URL)\n\tassert page.status_code == 200\n\tassert 'Joe Bloggs' in str(page.text)\n\n\n\n\ndef test_post_mainpage():\n\tpage = requests.post(COMPONENT_INDEX_URL, data=dict(name=\"Moby Dock\"))\n\tassert page.status_code == 200\n\tassert 'Moby Dock' in str(page.text)\n\n\n\n\ndef test_mainpage_html_escaping():\n\tpage = requests.post(COMPONENT_INDEX_URL, data=dict(name='\"><b>TEST<\/b><!--'))\n\tassert page.status_code == 200\n\tassert '<b>' not in str(page.text)\n\n\n\ndef test_get_identicon_with_valid_name_and_invalid_post_method_should_return_405():\n\tname_hash = 'ABCDEF123456789'\n\t\n\tpage = requests.post('{0}\/{1}'.format(COMPONENT_MONSTER_BASE_URL, name_hash))\n\t\n\tassert page.status_code == 405\n\n\n\n\ndef test_get_identicon_with_valid_name_and_cache_miss():\t\n\tname_hash = 'ABCDEF123456789'\n\tpage = requests.get('{0}\/{1}'.format(COMPONENT_MONSTER_BASE_URL, name_hash))\n\t\n\t# print('page.content : {0}'.format(page.content))\n\tassert page.status_code == 200\n\n\n\n\ndef test_get_identicon_with_valid_name_and_cache_hit():\t\n\tname_hash = 'ABCDEF123456789'\n\tpage = requests.get('{0}\/{1}'.format(COMPONENT_MONSTER_BASE_URL, name_hash))\n\t\n\t# print('page.content : {0}'.format(page.content))\n\tassert page.status_code == 200\n\n\n\n\ndef test_get_identicon_with_insecure_and_unescaped_invalid_name_hash():\n\tinvalid_name_hash = '<b>;i_am_invalid|name <{\"'\n\n\tpage = requests.get('{0}\/{1}'.format(COMPONENT_MONSTER_BASE_URL, invalid_name_hash))\n\t\n\t# print('page.content : {0}'.format(page.content))\n\tassert page.status_code == 200\n\n\n\n\nif __name__ == '__main__':\n\t# unittest.main()\n\tpytest.main()\n","subject":"Add component test functions using pytest","message":"Add component test functions using pytest\n","lang":"Python","license":"mit","repos":"anirbanroydas\/ci-testing-python,anirbanroydas\/ci-testing-python,anirbanroydas\/ci-testing-python"} {"commit":"024ea3b2e9e373abdcd78e44a163a2c32345073f","old_file":"unittests.py","new_file":"unittests.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nimport unittest\n\nimport const\nimport uniformdh\n\nimport obfsproxy.network.buffer as obfs_buf\n\nclass UniformDHTest( unittest.TestCase ):\n\n def setUp( self ):\n weAreServer = True\n self.udh = uniformdh.new(\"A\" * const.SHARED_SECRET_LENGTH, weAreServer)\n\n def test1_createHandshake( self ):\n handshake = self.udh.createHandshake()\n self.failUnless((const.PUBLIC_KEY_LENGTH +\n const.MARKER_LENGTH +\n const.HMAC_LENGTH) <= len(handshake) <=\n (const.MARKER_LENGTH +\n const.HMAC_LENGTH +\n const.MAX_PADDING_LENGTH))\n\n def test2_receivePublicKey( self ):\n buf = obfs_buf.Buffer(self.udh.createHandshake())\n\n self.failUnless(self.udh.receivePublicKey(buf, lambda x: x) == True)\n\n publicKey = self.udh.getRemotePublicKey()\n self.failUnless(len(publicKey) == const.PUBLIC_KEY_LENGTH)\n\n def test3_invalidHMAC( self ):\n # Make the HMAC invalid.\n handshake = self.udh.createHandshake()\n if handshake[-1] != 'a':\n handshake = handshake[:-1] + 'a'\n else:\n handshake = handshake[:-1] + 'b'\n\n buf = obfs_buf.Buffer(handshake)\n\n self.failIf(self.udh.receivePublicKey(buf, lambda x: x) == True)\n\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\n\nimport unittest\n\nimport const\nimport uniformdh\n\nimport obfsproxy.network.buffer as obfs_buf\n\nclass UniformDHTest( unittest.TestCase ):\n\n def setUp( self ):\n weAreServer = True\n self.udh = uniformdh.new(\"A\" * const.SHARED_SECRET_LENGTH, weAreServer)\n\n def test1_createHandshake( self ):\n handshake = self.udh.createHandshake()\n self.failUnless((const.PUBLIC_KEY_LENGTH +\n const.MARKER_LENGTH +\n const.HMAC_LENGTH) <= len(handshake) <=\n (const.MARKER_LENGTH +\n const.HMAC_LENGTH +\n const.MAX_PADDING_LENGTH))\n\n def test2_receivePublicKey( self ):\n buf = obfs_buf.Buffer(self.udh.createHandshake())\n\n def callback( masterKey ):\n self.failUnless(len(masterKey) == const.MASTER_KEY_LENGTH)\n\n self.failUnless(self.udh.receivePublicKey(buf, callback) == True)\n\n publicKey = self.udh.getRemotePublicKey()\n self.failUnless(len(publicKey) == const.PUBLIC_KEY_LENGTH)\n\n def test3_invalidHMAC( self ):\n # Make the HMAC invalid.\n handshake = self.udh.createHandshake()\n if handshake[-1] != 'a':\n handshake = handshake[:-1] + 'a'\n else:\n handshake = handshake[:-1] + 'b'\n\n buf = obfs_buf.Buffer(handshake)\n\n self.failIf(self.udh.receivePublicKey(buf, lambda x: x) == True)\n\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Extend UniformDH test to also verify the length of the shared master secret.","message":"Extend UniformDH test to also verify the length of the shared master secret.\n","lang":"Python","license":"bsd-3-clause","repos":"isislovecruft\/scramblesuit,isislovecruft\/scramblesuit"} {"commit":"8a9c4585d633e5c7858071fe1420850f930c7614","old_file":"manage.py","new_file":"manage.py","old_contents":"from flask_script import Manager\nfrom flask_skeleton_api.main import app\nimport subprocess\nimport os\n\nmanager = Manager(app)\n\n\n@manager.command\ndef unittest(report=False):\n \"\"\"Run unit tests\"\"\"\n\n if report:\n subprocess.call([\"py.test\", \"--junitxml=test-output\/unit-test-output.xml\",\n \"--cov-report=html:test-output\/unit-test-cov-report\"])\n else:\n subprocess.call([\"py.test\"])\n\n\n@manager.command\ndef integrationtest(report=False):\n \"\"\"Run integration tests\"\"\"\n\n if report:\n subprocess.call([\"py.test\", \"--junitxml=test-output\/integration-test-output.xml\",\n \"integration_tests\"])\n else:\n subprocess.call([\"py.test\", \"integration_tests\"])\n\n\n@manager.command\ndef runserver(port=9998):\n \"\"\"Run the app using flask server\"\"\"\n\n os.environ[\"PYTHONUNBUFFERED\"] = \"yes\"\n os.environ[\"LOG_LEVEL\"] = \"DEBUG\"\n os.environ[\"COMMIT\"] = \"LOCAL\"\n\n app.run(debug=True, port=int(port))\n\nif __name__ == \"__main__\":\n manager.run()\n","new_contents":"from flask_script import Manager\nfrom flask_skeleton_api.main import app\nimport subprocess\nimport os\n# Using Alembic?\n# See what extra lines are needed here:\n# http:\/\/192.168.249.38\/gadgets\/gadget-api\/blob\/master\/manage.py\n\nmanager = Manager(app)\n\n@manager.command\ndef unittest(report=False):\n \"\"\"Run unit tests\"\"\"\n\n if report:\n subprocess.call([\"py.test\", \"--junitxml=test-output\/unit-test-output.xml\",\n \"--cov-report=html:test-output\/unit-test-cov-report\"])\n else:\n subprocess.call([\"py.test\"])\n\n\n@manager.command\ndef integrationtest(report=False):\n \"\"\"Run integration tests\"\"\"\n\n if report:\n subprocess.call([\"py.test\", \"--junitxml=test-output\/integration-test-output.xml\",\n \"integration_tests\"])\n else:\n subprocess.call([\"py.test\", \"integration_tests\"])\n\n\n@manager.command\ndef runserver(port=9998):\n \"\"\"Run the app using flask server\"\"\"\n\n os.environ[\"PYTHONUNBUFFERED\"] = \"yes\"\n os.environ[\"LOG_LEVEL\"] = \"DEBUG\"\n os.environ[\"COMMIT\"] = \"LOCAL\"\n\n app.run(debug=True, port=int(port))\n\nif __name__ == \"__main__\":\n manager.run()\n","subject":"Add link to gadget for Alembic code","message":"Add link to gadget for Alembic code\n","lang":"Python","license":"mit","repos":"matthew-shaw\/thing-api"} {"commit":"61450328583cfb8e5ceee94a03502cef54bb11d6","old_file":"learning_journal\/tests\/test_models.py","new_file":"learning_journal\/tests\/test_models.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom learning_journal.models import Entry, DBSession\n\n\ndef test_create_entry(dbtransaction):\n \"\"\"Assert entry was entered into database.\"\"\"\n new_entry = Entry(title=\"Entry1\", text=\"Hey, this works. Awesome.\")\n assert new_entry.id is None\n DBSession.flush\n assert new_entry.id is not None\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom learning_journal.models import Entry, DBSession\n\n\ndef test_create_entry(dbtransaction):\n \"\"\"Assert entry was entered into database.\"\"\"\n new_entry = Entry(title=\"Entry1\", text=\"Hey, this works. Awesome.\")\n assert new_entry.id is None\n DBSession.add(new_entry)\n DBSession.flush\n assert new_entry.id is not None\n","subject":"Modify test file, still doesn't work. Messed around in pshell.","message":"Modify test file, still doesn't work. Messed around in pshell.\n","lang":"Python","license":"mit","repos":"DZwell\/learning_journal,DZwell\/learning_journal,DZwell\/learning_journal"} {"commit":"e45f394c61620db13bae579a29043dfdd6ae2d0f","old_file":"SLA_bot\/alertfeed.py","new_file":"SLA_bot\/alertfeed.py","old_contents":"import asyncio\nimport json\n\nimport aiohttp\n\nimport SLA_bot.config as cf\n\nclass AlertFeed:\n source_url = 'http:\/\/pso2emq.flyergo.eu\/api\/v2\/'\n\n async def download(url):\n try:\n async with aiohttp.get(url) as response:\n return await response.json()\n except json.decoder.JSONDecodeError:\n pass\n\n def parse_data(data):\n latest_alert = data[0]['text']\n lines = latest_alert.splitlines()\n header = '-' * len(lines[0])\n lines.insert(1, header)\n text = '\\n'.join(lines)\n return '```fix\\n{}\\n```'.format(text)\n \n async def fetch():\n header = cf.get('PSO2 Feed', 'header')\n raw_data = await AlertFeed.download(AlertFeed.source_url)\n return '** **\\n' + header + '\\n' + AlertFeed.parse_data(raw_data)\n","new_contents":"import asyncio\nimport json\n\nimport aiohttp\n\nimport SLA_bot.config as cf\n\nclass AlertFeed:\n source_url = 'http:\/\/pso2emq.flyergo.eu\/api\/v2\/'\n\n async def download(url):\n try:\n async with aiohttp.get(url) as response:\n return await response.json()\n except json.decoder.JSONDecodeError:\n pass\n\n def parse_data(data):\n latest_alert = data[0]['text']\n lines = latest_alert.splitlines()\n code_color = 'fix' if len(lines) >= 10 else ''\n header = '-' * len(lines[0])\n lines.insert(1, header)\n text = '\\n'.join(lines)\n return '```{}\\n{}\\n```'.format(code_color, text)\n \n async def fetch():\n header = cf.get('PSO2 Feed', 'header')\n raw_data = await AlertFeed.download(AlertFeed.source_url)\n return '** **\\n' + header + '\\n' + AlertFeed.parse_data(raw_data)\n","subject":"Remove text coloring in AlertFeed if it seems like scheduled text","message":"Remove text coloring in AlertFeed if it seems like scheduled text\n","lang":"Python","license":"mit","repos":"EsqWiggles\/SLA-bot,EsqWiggles\/SLA-bot"} {"commit":"d06b80227e404bd0ad36e6fd9d382c247e570ca9","old_file":"runtime\/Python2\/setup.py","new_file":"runtime\/Python2\/setup.py","old_contents":"from setuptools import setup\n\nv = '4.10.1'\nsetup(\n name='antlr4-python2-runtime',\n version=v,\n url='http:\/\/www.antlr.org',\n license='BSD',\n packages=['antlr4', 'antlr4.atn', 'antlr4.dfa', 'antlr4.tree', 'antlr4.error', 'antlr4.xpath'],\n package_dir={'': 'src'},\n author='Eric Vergnaud, Terence Parr, Sam Harwell',\n author_email='eric.vergnaud@wanadoo.fr',\n description=f'ANTLR {v} runtime for Python 2.7.12'\n)","new_contents":"from setuptools import setup\n\nv = '4.10.1'\nsetup(\n name='antlr4-python2-runtime',\n version=v,\n url='http:\/\/www.antlr.org',\n license='BSD',\n packages=['antlr4', 'antlr4.atn', 'antlr4.dfa', 'antlr4.tree', 'antlr4.error', 'antlr4.xpath'],\n package_dir={'': 'src'},\n author='Eric Vergnaud, Terence Parr, Sam Harwell',\n author_email='eric.vergnaud@wanadoo.fr',\n description='ANTLR %s runtime for Python 2.7.12' % v\n)\n","subject":"Fix SyntaxError due to F string","message":"[py2] Fix SyntaxError due to F string\n\nSigned-off-by: Travis Thieman <f1ef50ba1343ab5680bff0994219d82815f791bd@gmail.com>\n","lang":"Python","license":"bsd-3-clause","repos":"parrt\/antlr4,parrt\/antlr4,ericvergnaud\/antlr4,ericvergnaud\/antlr4,ericvergnaud\/antlr4,ericvergnaud\/antlr4,antlr\/antlr4,parrt\/antlr4,antlr\/antlr4,antlr\/antlr4,ericvergnaud\/antlr4,parrt\/antlr4,antlr\/antlr4,parrt\/antlr4,antlr\/antlr4,parrt\/antlr4,ericvergnaud\/antlr4,parrt\/antlr4,antlr\/antlr4,parrt\/antlr4,ericvergnaud\/antlr4,ericvergnaud\/antlr4,antlr\/antlr4,antlr\/antlr4,antlr\/antlr4,antlr\/antlr4,ericvergnaud\/antlr4,ericvergnaud\/antlr4,parrt\/antlr4,parrt\/antlr4"} {"commit":"9f503c6e54686f0fe48a3f68379bfda98dc5bb23","old_file":"pca\/__init__.py","new_file":"pca\/__init__.py","old_contents":"from os import path\n\nfrom devops.utils.version import Version\n\n\nPROJECT_NAME = 'python-clean-architecture'\nPACKAGE_NAME = 'pca'\nVERSION = Version(0, 0, 3)\n\nPACKAGE_DIR = path.dirname(__file__)\nPROJECT_DIR = path.dirname(PACKAGE_DIR)\n","new_contents":"from os import path\n\nfrom devops.utils.version import Version\n\n\nPROJECT_NAME = 'python-clean-architecture'\nPACKAGE_NAME = 'pca'\nVERSION = Version(0, 0, 4)\n\nPACKAGE_DIR = path.dirname(__file__)\nPROJECT_DIR = path.dirname(PACKAGE_DIR)\n","subject":"Bump version: 0.0.3 -> 0.0.4","message":"Bump version: 0.0.3 -> 0.0.4","lang":"Python","license":"mit","repos":"lhaze\/dharma"} {"commit":"c86e22a16eb2c1f2c95f81c232ae8535e447e935","old_file":"solutions\/pybasic_ex1_3_1.py","new_file":"solutions\/pybasic_ex1_3_1.py","old_contents":"# Use the codon variables you defined previously\nS = \"TCT\"\nL = \"CTT\"\nY = \"TAT\"\nC = \"TGT\"\n\n# Create a list for the protein sequence CLYSY\ncodons = [C, L, Y, S, Y]\n\n# Print the DNA sequence of the protein\nprint(\"DNA sequence:\", codons)\n\n# Print the DNA sequence of the last amino acid\nprint(\"Last codon:\", codons[-1])\n\n# Create two more variables containing the DNA sequence for a stop codon and a start codon\nstart = \"ATG\"\nstop = \"TGA\"\n\n# Replace the first element of the list with the start codon\ncodons[0] = start\n\n# Append the stop codon to the end of the list\ncodons.append(stop)\n\n# Print the resulting DNA sequence\nprint(\"DNA sequence after alteration:\", \"\".join(codons))\n","new_contents":"# Use the codon variables you defined previously\nS = \"TCT\"\nL = \"CTT\"\nY = \"TAT\"\nC = \"TGT\"\n\n# Create a list for the protein sequence CLYSY\ncodons = [C, L, Y, S, Y]\n\n# Print the DNA sequence of the protein\nprint(\"DNA sequence:\", codons)\n\n# Print the DNA sequence of the last amino acid\nprint(\"Last codon:\", codons[-1])\n\n# Create two more variables containing the DNA sequence for a stop codon and a start codon\nstart = \"ATG\"\nstop = \"TGA\"\n\n# Replace the first element of the list with the start codon\ncodons[0] = start\n\n# Append the stop codon to the end of the list\ncodons.append(stop)\n\n# Print the resulting DNA sequence\nprint(\"DNA sequence after alteration:\", codons)\n","subject":"Remove join in exercise 1.3.1 not seen yet in course","message":"Remove join in exercise 1.3.1 not seen yet in course\n","lang":"Python","license":"unlicense","repos":"pycam\/python-basic,pycam\/python-basic"} {"commit":"8dc69dca8538eb992989da396b65ade4fe2e5088","old_file":"polls\/models.py","new_file":"polls\/models.py","old_contents":"from django.db import models\nfrom django.utils import timezone\nfrom datetime import timedelta\n\n\nclass Poll(models.Model):\n text = models.CharField(max_length=200)\n created_ts = models.DateTimeField()\n updated_ts = models.DateTimeField(null=True, default=None)\n is_published = models.BooleanField(default=False)\n publication_date = models.DateTimeField(\n 'date published',\n default=None,\n )\n\n def __str__(self):\n return self.text\n\n def was_published_recently(self):\n return self.publication_date >= timezone.now() - timedelta(days=1)\n\n def save(self, *args, **kwargs):\n ''' On save, update timestamps '''\n if not self.id:\n self.created_ts = timezone.now()\n self.updated_ts = timezone.now()\n return super(Poll, self).save(*args, **kwargs)\n\n\nclass Choice(models.Model):\n poll = models.ForeignKey(Poll, on_delete=models.CASCADE)\n text = models.CharField(max_length=200)\n votes = models.IntegerField(default=0)\n\n def __str__(self):\n return self.text\n","new_contents":"from django.db import models\nfrom django.utils import timezone\nfrom datetime import timedelta\n\n\nclass Poll(models.Model):\n text = models.CharField(max_length=200)\n created_ts = models.DateTimeField()\n updated_ts = models.DateTimeField(null=True, default=None)\n is_published = models.BooleanField(default=False)\n publication_date = models.DateTimeField(\n 'date published',\n default=None,\n )\n\n def __str__(self):\n return self.text\n\n def was_published_recently(self):\n now = timezone.now()\n return now - timedelta(days=1) <= self.publication_date <= now\n\n def save(self, *args, **kwargs):\n ''' On save, update timestamps '''\n if not self.id:\n self.created_ts = timezone.now()\n self.updated_ts = timezone.now()\n return super(Poll, self).save(*args, **kwargs)\n\n\nclass Choice(models.Model):\n poll = models.ForeignKey(Poll, on_delete=models.CASCADE)\n text = models.CharField(max_length=200)\n votes = models.IntegerField(default=0)\n\n def __str__(self):\n return self.text\n","subject":"Fix was_published_recently reporting polls from the future","message":"Fix was_published_recently reporting polls from the future\n","lang":"Python","license":"mit","repos":"fernandocanizo\/django-poll-site,fernandocanizo\/django-poll-site,fernandocanizo\/django-poll-site"} {"commit":"c898d3f3d142727d0a55303238cda8044d729437","old_file":"motobot\/core_plugins\/commands.py","new_file":"motobot\/core_plugins\/commands.py","old_contents":"from motobot import command, Notice, split_response, IRCBot\n\n\n@command('commands')\ndef commands_command(bot, database, context, message, args):\n userlevel = bot.get_userlevel(context.channel, context.nick)\n\n valid_command = lambda plugin: plugin.type == IRCBot.command_plugin \\\n and plugin.level <= userlevel and not plugin.arg.hidden\n key = lambda plugin: (plugin.arg.trigger, plugin.func)\n\n command_groups = {}\n for command, func in map(key, filter(valid_command, bot.plugins)):\n value = command_groups.get(func, [])\n value.append(command)\n command_groups[func] = value\n\n format_group = lambda group: '({})'.format(', '.join(group)) \\\n if len(group) != 1 else group[0]\n commands = map(format_group, sorted(command_groups.values(), key=lambda x: x[0]))\n response = split_response(commands, \"Bot Commands: {};\")\n\n return response, Notice(context.nick)\n","new_contents":"from motobot import command, Notice, split_response, IRCBot\nfrom collections import defaultdict\n\n\ndef filter_plugins(plugins, userlevel):\n return map(\n lambda plugin: (plugin.arg.trigger, plugin.func), filter(\n lambda plugin: plugin.type == IRCBot.command_plugin and\n plugin.level <= userlevel and not plugin.arg.hidden,\n plugins\n )\n )\n\n\ndef format_group(group):\n return '({})'.format(', '.join(group)) if len(group) != 1 else group[0]\n\n\n@command('commands')\ndef commands_command(bot, database, context, message, args):\n userlevel = bot.get_userlevel(context.channel, context.nick)\n groups = defaultdict(lambda: [])\n\n for command, func in filter_plugins(bot.plugins, userlevel):\n groups[func].append(command)\n\n commands = map(format_group, sorted(groups.values(), key=lambda x: x[0]))\n response = split_response(commands, \"Bot Commands: {};\")\n\n return response, Notice(context.nick)\n","subject":"Revert \"Revert \"Cleans up split_response\"\"","message":"Revert \"Revert \"Cleans up split_response\"\"\n\nThis reverts commit c3c62feb9fbd8b7ff35d70eaaa5fecfb2093dbb0.\n","lang":"Python","license":"mit","repos":"Motoko11\/MotoBot"} {"commit":"7e78408dad1aab6bb42fd62601ee52e5f0ab3bd9","old_file":"stanczyk\/proxy.py","new_file":"stanczyk\/proxy.py","old_contents":"from twisted.internet import endpoints, reactor\nfrom txampext.multiplexing import ProxyingFactory\n\n\ndef connectProxy(namespace, identifier, _reactor=reactor):\n \"\"\"Start listening on some free local port; connections will be\n proxied to the virtual server with the given identifier.\n\n \"\"\"\n remote = namespace.get(\"remote\")\n if remote is None:\n raise RuntimeError(\"You are not connected to the exercise server. \"\n \"Call ``connect``.\")\n\n factory = ProxyingFactory(remote, identifier)\n endpoint = endpoints.TCP4ServerEndpoint(_reactor, 0, interface=\"localhost\")\n d = endpoint.listen(factory)\n d.addCallback(_listening, namespace, identifier)\n return None\n\n\ndef _listening(listeningPort, namespace, identifier):\n \"\"\"Started listening; report success to terminal.\n\n \"\"\"\n host = listeningPort.getHost()\n template = \"{id} is now listening on {h.host}:{h.port}\"\n namespace[\"manhole\"].writeLine(template.format(h=host, id=identifier))\n","new_contents":"from stanczyk.util import _getRemote\nfrom twisted.internet import endpoints, reactor\nfrom txampext.multiplexing import ProxyingFactory\n\n\ndef connectProxy(namespace, identifier, _reactor=reactor):\n \"\"\"Start listening on some free local port; connections will be\n proxied to the virtual server with the given identifier.\n\n \"\"\"\n endpoint = endpoints.TCP4ServerEndpoint(_reactor, 0, interface=\"localhost\")\n factory = ProxyingFactory(_getRemote(namespace), identifier)\n d = endpoint.listen(factory)\n d.addCallback(_listening, namespace, identifier)\n return None\n\n\ndef _listening(listeningPort, namespace, identifier):\n \"\"\"Started listening; report success to terminal.\n\n \"\"\"\n host = listeningPort.getHost()\n template = \"{id} is now listening on {h.host}:{h.port}\"\n namespace[\"manhole\"].writeLine(template.format(h=host, id=identifier))\n","subject":"Use the new fancy refactored remote logic","message":"Use the new fancy refactored remote logic\n","lang":"Python","license":"isc","repos":"crypto101\/stanczyk"} {"commit":"7a582488a3f8d86820dca7c3b44ff86b8dbe4412","old_file":"changes\/__init__.py","new_file":"changes\/__init__.py","old_contents":"import os\nimport subprocess\n\ntry:\n VERSION = __import__('pkg_resources') \\\n .get_distribution('changes').version\nexcept Exception, e:\n VERSION = 'unknown'\n\n\ndef _get_git_revision(path):\n try:\n r = subprocess.check_output('git rev-parse HEAD', cwd=path, shell=True)\n except Exception:\n return None\n return r.strip()\n\n\ndef get_revision():\n \"\"\"\n :returns: Revision number of this branch\/checkout, if available. None if\n no revision number can be determined.\n \"\"\"\n package_dir = os.path.dirname(__file__)\n checkout_dir = os.path.normpath(os.path.join(package_dir, os.pardir))\n path = os.path.join(checkout_dir, '.git')\n if os.path.exists(path):\n return _get_git_revision(path)\n return None\n\n\ndef get_version():\n base = VERSION\n if __build__:\n base = '%s (%s)' % (base, __build__)\n return base\n\n__build__ = get_revision()\n__docformat__ = 'restructuredtext en'\n","new_contents":"import os\nimport subprocess\n\ntry:\n VERSION = __import__('pkg_resources') \\\n .get_distribution('changes').version\nexcept Exception:\n VERSION = 'unknown'\n\n\ndef _get_git_revision(path):\n try:\n r = subprocess.check_output('git rev-parse HEAD', cwd=path, shell=True)\n except Exception:\n return None\n return r.strip()\n\n\ndef get_revision():\n \"\"\"\n :returns: Revision number of this branch\/checkout, if available. None if\n no revision number can be determined.\n \"\"\"\n package_dir = os.path.dirname(__file__)\n checkout_dir = os.path.normpath(os.path.join(package_dir, os.pardir))\n path = os.path.join(checkout_dir, '.git')\n if os.path.exists(path):\n return _get_git_revision(path)\n return None\n\n\ndef get_version():\n base = VERSION\n if __build__:\n base = '%s (%s)' % (base, __build__)\n return base\n\n__build__ = get_revision()\n__docformat__ = 'restructuredtext en'\n","subject":"Update exception syntax to be py3 compat","message":"Update exception syntax to be py3 compat\n","lang":"Python","license":"apache-2.0","repos":"bowlofstew\/changes,bowlofstew\/changes,dropbox\/changes,wfxiang08\/changes,dropbox\/changes,bowlofstew\/changes,dropbox\/changes,bowlofstew\/changes,wfxiang08\/changes,dropbox\/changes,wfxiang08\/changes,wfxiang08\/changes"} {"commit":"f5613b2b03f20f9d8f2a8d221ba1fae86664839c","old_file":"modules\/mpi-ring\/bin\/onramp_status.py","new_file":"modules\/mpi-ring\/bin\/onramp_status.py","old_contents":"#!\/usr\/bin\/env python\n\n#\n# Curriculum Module Status Script\n# - Run while the job is running\n# - Run -outside- of the allocation\n# - onramp_run_params.ini file is available in current working directory\n#\nimport sys\nimport re\n\n#\n# Display any special message you want the user to see, or leave blank if nothing.\n# Please restrict status messages to 1 line of text.\n#\n\n# Read in the output file\nlines = [line.rstrip('\\n') for line in open('onramp\/output.txt')]\n\n# If the file is empty then nothing to do\nif len(lines) <= 0:\n sys.exit(0)\n\n\n#print \"Status) Number of Lines: %d\" % len(lines)\n\n# Find the last line of 'Increment value'\nlast_status = None\nfor line in lines:\n searchObj = re.search( r'Increment value (.*)', line)\n if searchObj:\n last_status = searchObj.group(1).strip()\n\nif last_status is not None:\n print \"%s\" % last_status\n\n#\n# Exit successfully\n#\nsys.exit(0)\n","new_contents":"#!\/usr\/bin\/env python\n\n#\n# Curriculum Module Status Script\n# - Run while the job is running\n# - Run -outside- of the allocation\n# - onramp_run_params.ini file is available in current working directory\n#\nimport sys\nimport re\n\n#\n# Display any special message you want the user to see, or leave blank if nothing.\n# Please restrict status messages to 1 line of text.\n#\n\n# Read in the output file\nlines = [line.rstrip('\\n') for line in open('output.txt')]\n\n# If the file is empty then nothing to do\nif len(lines) <= 0:\n sys.exit(0)\n\n\n#print \"Status) Number of Lines: %d\" % len(lines)\n\n# Find the last line of 'Increment value'\nlast_status = None\nfor line in lines:\n searchObj = re.search( r'Increment value (.*)', line)\n if searchObj:\n last_status = searchObj.group(1).strip()\n\nif last_status is not None:\n print \"%s\" % last_status\n\n#\n# Exit successfully\n#\nsys.exit(0)\n","subject":"Update the status.py to look for the output.txt in the new location","message":"Update the status.py to look for the output.txt in the new location\n","lang":"Python","license":"bsd-3-clause","repos":"OnRampOrg\/onramp,koepked\/onramp,OnRampOrg\/onramp,ssfoley\/onramp,OnRampOrg\/onramp,koepked\/onramp,ssfoley\/onramp,koepked\/onramp,OnRampOrg\/onramp,koepked\/onramp,ssfoley\/onramp,ssfoley\/onramp,OnRampOrg\/onramp,koepked\/onramp,OnRampOrg\/onramp,OnRampOrg\/onramp,koepked\/onramp"} {"commit":"00b798c309d8807a562efb31751e82e5149ac7c8","old_file":"molo\/core\/api\/tests\/test_importers.py","new_file":"molo\/core\/api\/tests\/test_importers.py","old_contents":"\"\"\"\nTest the importing module.\nThis module relies heavily on an external service and requires\nquite a bit of mocking.\n\"\"\"\nimport json\n\nfrom django.test import TestCase\n\nfrom molo.core.tests.base import MoloTestCaseMixin\nfrom molo.core.api import importers\nfrom molo.core.api.tests import constants\n\n\nclass ArticleImportTestCase(MoloTestCaseMixin, TestCase):\n\n def setUp(self):\n self.mk_main()\n\n def test_importer_initializtion(self):\n content = json.dumps(constants.AVAILABLE_ARTICLES)\n importer = importers.ArticlePageImporter(content=content)","new_contents":"\"\"\"\nTest the importing module.\nThis module relies heavily on an external service and requires\nquite a bit of mocking.\n\"\"\"\nimport json\n\nfrom django.test import TestCase\n\nfrom molo.core.tests.base import MoloTestCaseMixin\nfrom molo.core.api import importers\nfrom molo.core.api.tests import constants\n\n\nclass ArticleImportTestCase(MoloTestCaseMixin, TestCase):\n\n def setUp(self):\n self.mk_main()\n\n def test_importer_initializtion(self):\n content = json.dumps(constants.AVAILABLE_ARTICLES)\n importer = importers.ArticlePageImporter(content=content)\n\n self.assertEqual(importer.articles(), content[\"items\"])","subject":"Write test for importer initialisation","message":"Write test for importer initialisation\n","lang":"Python","license":"bsd-2-clause","repos":"praekelt\/molo,praekelt\/molo,praekelt\/molo,praekelt\/molo"} {"commit":"190b4b193a2b33d7904310d24891e8aec18a126f","old_file":"pipreq\/cli.py","new_file":"pipreq\/cli.py","old_contents":"import argparse\nimport sys\n\nfrom pipreq.command import Command\n\n\ndef create_parser():\n parser = argparse.ArgumentParser(\n description='Manage Python package requirements across multiple environments using '\n 'per-environment requirements files.')\n\n parser.add_argument('-g', '--generate', action='store_true', default=False,\n help='Generate requirements files')\n parser.add_argument('-c', '--create', action='store_true', default=False,\n help='Create or update rc file (requires list of packages)')\n parser.add_argument('-U', '--upgrade', action='store_true', default=False,\n help='Upgrade packages (requires list of packages)')\n parser.add_argument('packages', nargs='?', type=argparse.FileType('r'), default=sys.stdin)\n\n return parser\n\n\ndef verify_args(args):\n if not args.create and not args.generate and not args.upgrade:\n return u'Must specify generate (-g) or create\/upgrade (-[cu]) with packages'\n return None\n\n\ndef error(parser, message):\n parser.print_help()\n parser.exit(message=\"\\nERROR: %s\\n\" % message)\n\n\ndef main():\n try:\n parser = create_parser()\n parsed_args = parser.parse_args()\n error_message = verify_args(parsed_args)\n if error_message:\n error(parser, error_message)\n command = Command(parsed_args, \".requirementsrc\")\n command.run()\n except KeyboardInterrupt:\n sys.exit()\n","new_contents":"import argparse\nimport sys\n\nfrom pipreq.command import Command\n\n\ndef create_parser():\n parser = argparse.ArgumentParser(\n description='Manage Python package requirements across multiple environments using '\n 'per-environment requirements files.')\n\n parser.add_argument('-g', '--generate', action='store_true', default=False,\n help='Generate requirements files')\n parser.add_argument('-c', '--create', action='store_true', default=False,\n help='Create or update rc file (requires list of packages)')\n parser.add_argument('-U', '--upgrade', action='store_true', default=False,\n help='Upgrade packages (requires list of packages)')\n parser.add_argument('packages', nargs='?', type=argparse.FileType('r'), default=sys.stdin)\n\n return parser\n\n\ndef verify_args(args):\n if not args.create and not args.generate and not args.upgrade:\n return 'Must specify generate (-g) or create\/upgrade (-[cu]) with packages'\n return None\n\n\ndef error(parser, message):\n parser.print_help()\n parser.exit(message=\"\\nERROR: %s\\n\" % message)\n\n\ndef main():\n try:\n parser = create_parser()\n parsed_args = parser.parse_args()\n error_message = verify_args(parsed_args)\n if error_message:\n error(parser, error_message)\n command = Command(parsed_args, \".requirementsrc\")\n command.run()\n except KeyboardInterrupt:\n sys.exit()\n","subject":"Remove unnecessary u on string","message":"Remove unnecessary u on string\n","lang":"Python","license":"mit","repos":"jessamynsmith\/pipwrap,jessamynsmith\/pipreq,jessamynsmith\/pipwrap,jessamynsmith\/pipreq"} {"commit":"cb6f11ad05ef07facf651f8fbccae9e86e0a77c8","old_file":"processing.py","new_file":"processing.py","old_contents":"#!\/usr\/bin\/env python\n\"\"\"\nProcessing routines for the waveFlapper case.\n\n\"\"\"\n\nimport foampy\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nwidth_2d = 0.1\nwidth_3d = 3.66\n\ndef plot_force():\n \"\"\"Plots the streamwise force on the paddle over time.\"\"\"\n \ndef plot_moment():\n data = foampy.load_forces_moments()\n i = 10\n t = data[\"time\"][i:]\n m = data[\"moment\"][\"pressure\"][\"z\"] + data[\"moment\"][\"viscous\"][\"z\"]\n m = m[i:]*width_3d\/width_2d\n plt.figure()\n plt.plot(t, m)\n plt.xlabel(\"t (s)\")\n plt.ylabel(\"Flapper moment (Nm)\")\n print(\"Max moment from CFD =\", m.max(), \"Nm\")\n print(\"Theoretical max moment (including inertia) =\", 5500*3.3, \"Nm\") \n plt.show()\n \nif __name__ == \"__main__\":\n plot_moment()\n","new_contents":"#!\/usr\/bin\/env python\n\"\"\"\nProcessing routines for the waveFlapper case.\n\n\"\"\"\n\nimport foampy\nimport numpy as np\nimport matplotlib.pyplot as plt\n\nwidth_2d = 0.1\nwidth_3d = 3.66\nm_paddle = 1270.0 # Paddle mass in kg, from OMB manual\nh_piston = 3.3147 \nI_paddle = 1\/3*m_paddle*h_piston**2\n\ndef plot_force():\n \"\"\"Plots the streamwise force on the paddle over time.\"\"\"\n \ndef plot_moment():\n data = foampy.load_forces_moments()\n i = 10\n t = data[\"time\"][i:]\n m = data[\"moment\"][\"pressure\"][\"z\"] + data[\"moment\"][\"viscous\"][\"z\"]\n m = m[i:]*width_3d\/width_2d\n period = 2.2\n omega = 2*np.pi\/period\n theta = 0.048*np.sin(omega*t)\n theta_doubledot = -0.048*omega**2*np.sin(omega*t)\n m_inertial = I_paddle*theta_doubledot\n m += m_inertial\n plt.figure()\n plt.plot(t, m)\n plt.xlabel(\"t (s)\")\n plt.ylabel(\"Flapper moment (Nm)\")\n print(\"Max moment from CFD =\", m.max(), \"Nm\")\n print(\"Theoretical max moment (including inertia) =\", 5500*3.3, \"Nm\") \n plt.show()\n \nif __name__ == \"__main__\":\n plot_moment()\n","subject":"Add paddle inertia to calculations","message":"Add paddle inertia to calculations\n","lang":"Python","license":"cc0-1.0","repos":"petebachant\/waveFlapper-OpenFOAM,petebachant\/waveFlapper-OpenFOAM,petebachant\/waveFlapper-OpenFOAM"} {"commit":"db977f65a6f986508c826b645b9c94e5eff4f83f","old_file":"oidc_provider\/management\/commands\/creatersakey.py","new_file":"oidc_provider\/management\/commands\/creatersakey.py","old_contents":"from Crypto.PublicKey import RSA\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\n\n\nclass Command(BaseCommand):\n help = 'Randomly generate a new RSA key for the OpenID server'\n\n def handle(self, *args, **options):\n try:\n key = RSA.generate(1024)\n file_path = settings.BASE_DIR + '\/OIDC_RSA_KEY.pem'\n with open(file_path, 'w') as f:\n f.write(key.exportKey('PEM'))\n self.stdout.write('RSA key successfully created at: ' + file_path)\n except Exception as e:\n self.stdout.write('Something goes wrong: {0}'.format(e))\n","new_contents":"from Crypto.PublicKey import RSA\n\nfrom django.conf import settings\nfrom django.core.management.base import BaseCommand\n\n\nclass Command(BaseCommand):\n help = 'Randomly generate a new RSA key for the OpenID server'\n\n def handle(self, *args, **options):\n try:\n key = RSA.generate(1024)\n file_path = settings.BASE_DIR + '\/OIDC_RSA_KEY.pem'\n with open(file_path, 'wb') as f:\n f.write(key.exportKey('PEM'))\n self.stdout.write('RSA key successfully created at: ' + file_path)\n except Exception as e:\n self.stdout.write('Something goes wrong: {0}'.format(e))\n","subject":"Append binary file mode to write RSA exported key needed by Python 3","message":"Append binary file mode to write RSA exported key needed by Python 3\n","lang":"Python","license":"mit","repos":"ByteInternet\/django-oidc-provider,bunnyinc\/django-oidc-provider,wayward710\/django-oidc-provider,ByteInternet\/django-oidc-provider,wojtek-fliposports\/django-oidc-provider,wayward710\/django-oidc-provider,juanifioren\/django-oidc-provider,bunnyinc\/django-oidc-provider,torreco\/django-oidc-provider,wojtek-fliposports\/django-oidc-provider,juanifioren\/django-oidc-provider,torreco\/django-oidc-provider"} {"commit":"90bc04a92bbe6f29d1487fbd87a4fad811f22c93","old_file":"setup\/setup-test-docs.py","new_file":"setup\/setup-test-docs.py","old_contents":"#!\/usr\/bin\/python\n#\n# SCRIPT FOR POPULATING TEST SOLR SERVER CORE WITH TEST DOCUMENTS\n#\n# Usage: python setup-test-docs.py <Solr Endpoint Url>\n#\n# Solr endpoint URL should be in the form:\n# https:\/\/example.com\/solr\/<core-name>\/\n#\n# .txt files in the directory .\/txt\/ will be committed to user-provided Solr\n# core matching the name <core-name>.\n\nimport os\nfrom os import listdir\nfrom os.path import isfile, join\nimport json\nimport sys\n\narguments = sys.argv\n\nsolrApiUrl = arguments[1]\n\nfilePaths = [f for f in listdir('txt') if isfile(join('txt', f))]\n\nTEMPLATE = \"\"\"\n{\n \"add\": {\n \"doc\":\n {\"title\":\"%s\", \"body\": %s},\n \"boost\":1.0,\n \"overwrite\":true,\n \"commitWithin\":1000\n }\n}\n\"\"\"\n\nheaders = {'Content-type': 'application\/json'}\n\nfor i, path in enumerate(filePaths):\n print str(i) + '\\tProcessing ' + path\n f = open('txt\/' + path)\n text = f.read()\n\n commandJson = TEMPLATE % (path.replace('.txt', ''), json.dumps(text))\n\n os.system(\"curl \" + solrApiUrl + \"update?commit=true -H 'Content-type:application\/json' -d '%s'\" % commandJson)\n print '\\nDone.\\n----------------------------------'\n","new_contents":"#!\/usr\/bin\/python\n#\n# SCRIPT FOR POPULATING TEST SOLR SERVER CORE WITH TEST DOCUMENTS\n#\n# Usage: python setup-test-docs.py <Solr Endpoint Url>\n#\n# Solr endpoint URL should be in the form:\n# https:\/\/example.com\/solr\/<core-name>\/\n#\n# .txt files in the directory .\/txt\/ will be committed to user-provided Solr\n# core matching the name <core-name>.\n\nimport os\nfrom os import listdir\nfrom os.path import isfile, join\nimport json\nimport sys\n\nTEST_DOC_DIR = 'test_docs'\n\narguments = sys.argv\n\nsolrApiUrl = arguments[1]\n\nfilePaths = [f for f in listdir(TEST_DOC_DIR) if isfile(join(TEST_DOC_DIR, f))]\n\nTEMPLATE = \"\"\"\n{\n \"add\": {\n \"doc\":\n {\"title\":\"%s\", \"body\": %s},\n \"boost\":1.0,\n \"overwrite\":true,\n \"commitWithin\":1000\n }\n}\n\"\"\"\n\nheaders = {'Content-type': 'application\/json'}\n\nfor i, path in enumerate(filePaths):\n print str(i) + '\\tProcessing ' + path\n f = open(TEST_DOC_DIR + '\/' + path)\n text = f.read()\n\n commandJson = TEMPLATE % (path.replace('.txt', ''), json.dumps(text))\n\n os.system(\"curl \" + solrApiUrl + \"update?commit=true -H 'Content-type:application\/json' -d '%s'\" % commandJson)\n print '\\nDone.\\n----------------------------------'\n","subject":"Use test_docs as directory for test documents for solr server","message":"Use test_docs as directory for test documents for solr server\n","lang":"Python","license":"mit","repos":"gios-asu\/search-api"} {"commit":"f34de068e71c57b434c48c9c2b90471112bb4a2b","old_file":"common\/djangoapps\/util\/bad_request_rate_limiter.py","new_file":"common\/djangoapps\/util\/bad_request_rate_limiter.py","old_contents":"\"\"\"\nA utility class which wraps the RateLimitMixin 3rd party class to do bad request counting\nwhich can be used for rate limiting\n\"\"\"\nfrom ratelimitbackend.backends import RateLimitMixin\nfrom django.conf import settings\n\nif settings.FEATURES.get('EDRAAK_RATELIMIT_APP', False):\n from edraak_ratelimit.backends import EdraakRateLimitMixin\n RateLimitMixin = EdraakRateLimitMixin\n\n\nclass BadRequestRateLimiter(RateLimitMixin):\n \"\"\"\n Use the 3rd party RateLimitMixin to help do rate limiting on the Password Reset flows\n \"\"\"\n\n def is_rate_limit_exceeded(self, request):\n \"\"\"\n Returns if the client has been rated limited\n \"\"\"\n counts = self.get_counters(request)\n is_exceeded = sum(counts.values()) >= self.requests\n\n if is_exceeded:\n self.db_log_failed_attempt(request)\n\n return is_exceeded\n\n def tick_bad_request_counter(self, request):\n \"\"\"\n Ticks any counters used to compute when rate limt has been reached\n \"\"\"\n self.cache_incr(self.get_cache_key(request))\n","new_contents":"\"\"\"\nA utility class which wraps the RateLimitMixin 3rd party class to do bad request counting\nwhich can be used for rate limiting\n\"\"\"\nfrom ratelimitbackend.backends import RateLimitMixin\nfrom django.conf import settings\n\nif settings.FEATURES.get('EDRAAK_RATELIMIT_APP', False):\n from edraak_ratelimit.backends import EdraakRateLimitMixin\n RateLimitMixin = EdraakRateLimitMixin\n\n\nclass BadRequestRateLimiter(RateLimitMixin):\n \"\"\"\n Use the 3rd party RateLimitMixin to help do rate limiting on the Password Reset flows\n \"\"\"\n\n def is_rate_limit_exceeded(self, request):\n \"\"\"\n Returns if the client has been rated limited\n \"\"\"\n counts = self.get_counters(request)\n is_exceeded = sum(counts.values()) >= self.requests\n\n if is_exceeded and settings.FEATURES.get('EDRAAK_RATELIMIT_APP', False):\n self.db_log_failed_attempt(request)\n\n return is_exceeded\n\n def tick_bad_request_counter(self, request):\n \"\"\"\n Ticks any counters used to compute when rate limt has been reached\n \"\"\"\n self.cache_incr(self.get_cache_key(request))\n","subject":"Fix object has no db_log_failed_attempt","message":"Fix object has no db_log_failed_attempt\n","lang":"Python","license":"agpl-3.0","repos":"Edraak\/edraak-platform,Edraak\/edraak-platform,Edraak\/edraak-platform,Edraak\/edraak-platform"} {"commit":"35201e71037d544893a59bfda8c4538fcb6fb4b7","old_file":"api\/tests\/test_scrape_item.py","new_file":"api\/tests\/test_scrape_item.py","old_contents":"from api.scrapers.item import scrape_item_by_id\nfrom api import app\n\nfrom flask.json import loads\n\nimport unittest\n\napp.config['TESTING'] = True\n\n\nclass ScrapeItem(unittest.TestCase):\n\n def test_scrape_item_by_id(self):\n item = scrape_item_by_id('d19447e548d')\n self.assertEqual('d19447e548d', item.lodestone_id)\n self.assertEqual('Thyrus Zenith', item.name)\n self.assertEqual('Two-handed Conjurer\\'s Arm', item.type)\n self.assertEqual(90, item.ilvl)\n\n def test_item_json(self):\n with app.test_client() as client:\n response = client.get('\/scrape\/item\/cada9ec7074')\n self.assertEqual(response.status_code, 200)\n\n json = loads(response.data)\n self.assertEqual(json, {\n \"id\": \"cada9ec7074\",\n \"ilvl\": 110,\n \"name\": \"Arachne Robe\",\n \"type\": \"Body\"\n })\n\n def test_invalid_lodestone_id(self):\n with app.test_client() as client:\n response = client.post('\/scrape\/item\/23fh032hf0oi1so3a012r1')\n self.assertEqual(response.status_code, 405)\n","new_contents":"from api.scrapers.item import scrape_item_by_id\nfrom api import app, db\n\nfrom flask.json import loads\n\nimport unittest\n\napp.config['TESTING'] = True\ndb.create_all()\n\n\nclass ScrapeItem(unittest.TestCase):\n\n def test_scrape_item_by_id(self):\n item = scrape_item_by_id('d19447e548d')\n self.assertEqual('d19447e548d', item.lodestone_id)\n self.assertEqual('Thyrus Zenith', item.name)\n self.assertEqual('Two-handed Conjurer\\'s Arm', item.type)\n self.assertEqual(90, item.ilvl)\n\n def test_item_json(self):\n with app.test_client() as client:\n response = client.get('\/scrape\/item\/cada9ec7074')\n self.assertEqual(response.status_code, 200)\n\n json = loads(response.data)\n self.assertEqual(json, {\n \"id\": \"cada9ec7074\",\n \"ilvl\": 110,\n \"name\": \"Arachne Robe\",\n \"type\": \"Body\"\n })\n\n def test_invalid_lodestone_id(self):\n with app.test_client() as client:\n response = client.post('\/scrape\/item\/23fh032hf0oi1so3a012r1')\n self.assertEqual(response.status_code, 405)\n","subject":"Create tables in database before running tests","message":"Create tables in database before running tests\n","lang":"Python","license":"mit","repos":"Demotivated\/loadstone"} {"commit":"3e7d83d51fa43f8e93ad548b07193f13791f8abe","old_file":"django_lightweight_queue\/middleware\/transaction.py","new_file":"django_lightweight_queue\/middleware\/transaction.py","old_contents":"from django.db import transaction, connection\n\nclass TransactionMiddleware(object):\n def process_job(self, job):\n if not connection.in_atomic_block:\n transaction.set_autocommit(False)\n\n def process_result(self, job, result, duration):\n if not connection.in_atomic_block:\n transaction.commit()\n\n def process_exception(self, job, time_taken, *exc_info):\n if not connection.in_atomic_block:\n transaction.rollback()\n","new_contents":"from django.db import transaction, connection\n\nclass TransactionMiddleware(object):\n def process_job(self, job):\n if not connection.in_atomic_block:\n transaction.set_autocommit(False)\n\n def process_result(self, job, result, duration):\n if not connection.in_atomic_block:\n transaction.commit()\n\n def process_exception(self, job, time_taken, *exc_info):\n if not connection.in_atomic_block:\n transaction.rollback()\n\n# Legacy\nif not hasattr(connection, 'in_atomic_block'):\n class TransactionMiddleware(object):\n def process_job(self, job):\n transaction.enter_transaction_management()\n transaction.managed(True)\n\n def process_result(self, job, result, duration):\n if not transaction.is_managed():\n return\n if transaction.is_dirty():\n transaction.commit()\n transaction.leave_transaction_management()\n\n def process_exception(self, job, time_taken, *exc_info):\n if transaction.is_dirty():\n transaction.rollback()\n transaction.leave_transaction_management()\n","subject":"Add a legacy version for older versions of Django.","message":"Add a legacy version for older versions of Django.\n\nSigned-off-by: Chris Lamb <711c73f64afdce07b7e38039a96d2224209e9a6c@thread.com>\n","lang":"Python","license":"bsd-3-clause","repos":"thread\/django-lightweight-queue,lamby\/django-lightweight-queue,prophile\/django-lightweight-queue,thread\/django-lightweight-queue,prophile\/django-lightweight-queue"} {"commit":"b6c98dd016aa440f96565ceaee2716cd530beae5","old_file":"pages\/search_indexes.py","new_file":"pages\/search_indexes.py","old_contents":"\"\"\"Django haystack `SearchIndex` module.\"\"\"\nfrom pages.models import Page, Content\n\nfrom haystack.indexes import SearchIndex, CharField, DateTimeField\nfrom haystack import site\n\nimport datetime\n\nclass PageIndex(SearchIndex):\n \"\"\"Search index for pages content.\"\"\"\n text = CharField(document=True, use_template=True)\n title = CharField(model_attr='title')\n publication_date = DateTimeField(model_attr='publication_date')\n\n def get_queryset(self):\n \"\"\"Used when the entire index for model is updated.\"\"\"\n return Page.objects.published()\n\n\nsite.register(Page, PageIndex)","new_contents":"\"\"\"Django haystack `SearchIndex` module.\"\"\"\nfrom pages.models import Page, Content\n\nfrom haystack.indexes import SearchIndex, CharField, DateTimeField\nfrom haystack import site\n\nimport datetime\n\nclass PageIndex(SearchIndex):\n \"\"\"Search index for pages content.\"\"\"\n text = CharField(document=True, use_template=True)\n title = CharField(model_attr='title')\n url = CharField(model_attr='get_absolute_url')\n publication_date = DateTimeField(model_attr='publication_date')\n\n def get_queryset(self):\n \"\"\"Used when the entire index for model is updated.\"\"\"\n return Page.objects.published()\n\n\nsite.register(Page, PageIndex)","subject":"Add a url attribute to the SearchIndex for pages.","message":"Add a url attribute to the SearchIndex for pages.\n\nThis is useful when displaying a list of search results because we\ncan create a link to the result without having to hit the database\nfor every object in the result list.\n","lang":"Python","license":"bsd-3-clause","repos":"remik\/django-page-cms,akaihola\/django-page-cms,batiste\/django-page-cms,akaihola\/django-page-cms,batiste\/django-page-cms,akaihola\/django-page-cms,remik\/django-page-cms,pombredanne\/django-page-cms-1,pombredanne\/django-page-cms-1,pombredanne\/django-page-cms-1,oliciv\/django-page-cms,remik\/django-page-cms,remik\/django-page-cms,oliciv\/django-page-cms,oliciv\/django-page-cms,batiste\/django-page-cms"} {"commit":"7f86ab26fb1c6ba01f81fdc3f5b66a0f079c23ff","old_file":"tests\/test_app.py","new_file":"tests\/test_app.py","old_contents":"import asyncio\nfrom unittest import mock\n\nimport aiohttp\nimport pytest\n\nfrom bottery.app import App\n\n\ndef test_app_session():\n app = App()\n assert isinstance(app.session, aiohttp.ClientSession)\n\n\ndef test_app_already_configured_session():\n app = App()\n app._session = 'session'\n assert app.session == 'session'\n\n\ndef test_app_loop():\n app = App()\n assert isinstance(app.loop, asyncio.AbstractEventLoop)\n\n\ndef test_app_already_configured_loop():\n app = App()\n app._loop = 'loop'\n assert app.loop == 'loop'\n\n\n@mock.patch('bottery.app.settings')\ndef test_app_configure_without_platforms(mocked_settings):\n \"\"\"Should raise Exception if no platform was found at settings\"\"\"\n\n mocked_settings.PLATFORMS = {}\n app = App()\n with pytest.raises(Exception):\n app.configure_platforms()\n","new_contents":"import asyncio\nimport sys\nfrom unittest import mock\n\nimport aiohttp\nimport pytest\n\nfrom bottery.app import App\n\n\n@pytest.fixture\ndef mocked_engine():\n mocked_engine_module = mock.MagicMock()\n mocked_engine_instance = mocked_engine_module.engine.return_value\n mocked_engine_instance.tasks.return_value = [(mock.MagicMock(), )]\n sys.modules['tests.fake_engine'] = mocked_engine_module\n\n yield {\n 'module': mocked_engine_module,\n 'instance': mocked_engine_instance\n }\n\n del sys.modules['tests.fake_engine']\n\n\ndef test_app_session():\n app = App()\n assert isinstance(app.session, aiohttp.ClientSession)\n\n\ndef test_app_already_configured_session():\n app = App()\n app._session = 'session'\n assert app.session == 'session'\n\n\ndef test_app_loop():\n app = App()\n assert isinstance(app.loop, asyncio.AbstractEventLoop)\n\n\ndef test_app_already_configured_loop():\n app = App()\n app._loop = 'loop'\n assert app.loop == 'loop'\n\n\n@mock.patch('bottery.app.settings')\ndef test_app_configure_without_platforms(mocked_settings):\n \"\"\"Should raise Exception if no platform was found at settings\"\"\"\n\n mocked_settings.PLATFORMS = {}\n app = App()\n with pytest.raises(Exception):\n app.configure_platforms()\n\n@mock.patch('bottery.app.settings')\ndef test_app_configure_with_platforms(mocked_settings, mocked_engine):\n \"\"\"Should call the platform interface methods\"\"\"\n\n mocked_settings.PLATFORMS = {\n 'test': {\n 'ENGINE': 'tests.fake_engine',\n 'OPTIONS': {\n 'token': 'should-be-a-valid-token'\n }\n }\n }\n\n app = App()\n app.configure_platforms()\n\n mocked_engine['module'].engine.assert_called_with(\n session=app.session,\n token='should-be-a-valid-token'\n )\n mocked_engine['instance'].configure.assert_called_with()\n mocked_engine['instance'].tasks.assert_called_with()\n","subject":"Increase the code coverage of App.configure_platforms method","message":"Increase the code coverage of App.configure_platforms method\n","lang":"Python","license":"mit","repos":"rougeth\/bottery"} {"commit":"2e9c6c883de12b7293b9e932e5268a2d806e714c","old_file":"chatterbot\/logic\/time_adapter.py","new_file":"chatterbot\/logic\/time_adapter.py","old_contents":"from __future__ import unicode_literals\nfrom datetime import datetime\nfrom .logic_adapter import LogicAdapter\n\n\nclass TimeLogicAdapter(LogicAdapter):\n \"\"\"\n The TimeLogicAdapter returns the current time.\n \"\"\"\n\n def __init__(self, **kwargs):\n super(TimeLogicAdapter, self).__init__(**kwargs)\n from textblob.classifiers import NaiveBayesClassifier\n\n training_data = [\n ('what time is it', 1),\n ('do you know the time', 1),\n ('do you know what time it is', 1),\n ('what is the time', 1),\n ('it is time to go to sleep', 0),\n ('what is your favorite color', 0),\n ('i had a great time', 0),\n ('what is', 0)\n ]\n\n self.classifier = NaiveBayesClassifier(training_data)\n\n def process(self, statement):\n from chatterbot.conversation import Statement\n\n now = datetime.now()\n\n confidence = self.classifier.classify(statement.text.lower())\n response = Statement('The current time is ' + now.strftime('%I:%M %p'))\n\n return confidence, response\n","new_contents":"from __future__ import unicode_literals\nfrom datetime import datetime\nfrom .logic_adapter import LogicAdapter\n\n\nclass TimeLogicAdapter(LogicAdapter):\n \"\"\"\n The TimeLogicAdapter returns the current time.\n \"\"\"\n\n def __init__(self, **kwargs):\n super(TimeLogicAdapter, self).__init__(**kwargs)\n from nltk import NaiveBayesClassifier\n\n self.positive = [\n 'what time is it',\n 'do you know the time',\n 'do you know what time it is',\n 'what is the time'\n ]\n\n self.negative = [\n 'it is time to go to sleep',\n 'what is your favorite color',\n 'i had a great time',\n 'what is'\n ]\n\n labeled_data = (\n [(name, 0) for name in self.negative] +\n [(name, 1) for name in self.positive]\n )\n\n # train_set = apply_features(self.time_question_features, training_data)\n train_set = [(self.time_question_features(n), text) for (n, text) in labeled_data]\n\n self.classifier = NaiveBayesClassifier.train(train_set)\n\n def time_question_features(self, text):\n \"\"\"\n Provide an analysis of significan features in the string.\n \"\"\"\n features = {}\n\n all_words = \" \".join(self.positive + self.negative).split()\n\n for word in text.split():\n features['contains({})'.format(word)] = (word in all_words)\n\n for letter in 'abcdefghijklmnopqrstuvwxyz':\n features['count({})'.format(letter)] = text.lower().count(letter)\n features['has({})'.format(letter)] = (letter in text.lower())\n\n return features\n\n def process(self, statement):\n from chatterbot.conversation import Statement\n\n now = datetime.now()\n\n time_features = self.time_question_features(statement.text.lower())\n confidence = self.classifier.classify(time_features)\n response = Statement('The current time is ' + now.strftime('%I:%M %p'))\n\n return confidence, response\n","subject":"Remove textblob dependency in time logic adapter","message":"Remove textblob dependency in time logic adapter\n","lang":"Python","license":"bsd-3-clause","repos":"gunthercox\/ChatterBot,vkosuri\/ChatterBot,Gustavo6046\/ChatterBot,davizucon\/ChatterBot,Reinaesaya\/OUIRL-ChatBot,Reinaesaya\/OUIRL-ChatBot,maclogan\/VirtualPenPal"} {"commit":"025c95a59b079d630c778646d5c82f5e0679b47c","old_file":"sale_automatic_workflow\/models\/account_invoice.py","new_file":"sale_automatic_workflow\/models\/account_invoice.py","old_contents":"# -*- coding: utf-8 -*-\n# © 2011 Akretion Sébastien BEAU <sebastien.beau@akretion.com>\n# © 2013 Camptocamp SA (author: Guewen Baconnier)\n# © 2016 Sodexis\n# License AGPL-3.0 or later (http:\/\/www.gnu.org\/licenses\/agpl).\n\nfrom odoo import models, fields\n\n\nclass AccountInvoice(models.Model):\n _inherit = \"account.invoice\"\n\n workflow_process_id = fields.Many2one(\n comodel_name='sale.workflow.process',\n string='Sale Workflow Process'\n )\n","new_contents":"# -*- coding: utf-8 -*-\n# © 2011 Akretion Sébastien BEAU <sebastien.beau@akretion.com>\n# © 2013 Camptocamp SA (author: Guewen Baconnier)\n# © 2016 Sodexis\n# License AGPL-3.0 or later (http:\/\/www.gnu.org\/licenses\/agpl).\nfrom odoo import models, fields\n\n\nclass AccountInvoice(models.Model):\n _inherit = \"account.invoice\"\n\n workflow_process_id = fields.Many2one(\n comodel_name='sale.workflow.process',\n string='Sale Workflow Process',\n copy=False,\n )\n","subject":"Fix issue on account.invoice about workflow_process_id: if a user duplicate an invoice, it copy also the workflow and validations (the reason of bugs)","message":"[FIX] Fix issue on account.invoice about workflow_process_id: if a user duplicate an invoice, it copy also the workflow and validations (the reason of bugs)\n","lang":"Python","license":"agpl-3.0","repos":"kittiu\/sale-workflow,kittiu\/sale-workflow"} {"commit":"7e2440c00ce75dc3ff0eac53e63d629981a9873a","old_file":"raven\/contrib\/celery\/__init__.py","new_file":"raven\/contrib\/celery\/__init__.py","old_contents":"\"\"\"\nraven.contrib.celery\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n:copyright: (c) 2010 by the Sentry Team, see AUTHORS for more details.\n:license: BSD, see LICENSE for more details.\n\"\"\"\n\ntry:\n from celery.task import task\nexcept ImportError:\n from celery.decorators import task\nfrom celery.signals import task_failure\nfrom raven.base import Client\n\n\nclass CeleryMixin(object):\n def send_encoded(self, message):\n \"Errors through celery\"\n self.send_raw.delay(message)\n\n @task(routing_key='sentry')\n def send_raw(self, message):\n return super(CeleryMixin, self).send_encoded(message)\n\n\nclass CeleryClient(CeleryMixin, Client):\n pass\n\n\ndef register_signal(client):\n def process_failure_signal(exception, traceback, sender, task_id,\n signal, args, kwargs, einfo, **kw):\n exc_info = (type(exception), exception, traceback)\n client.captureException(\n exc_info=exc_info,\n extra={\n 'task_id': task_id,\n 'sender': sender,\n 'args': args,\n 'kwargs': kwargs,\n })\n task_failure.connect(process_failure_signal)\n","new_contents":"\"\"\"\nraven.contrib.celery\n~~~~~~~~~~~~~~~~~~~~~~~~~~~~\n\n:copyright: (c) 2010 by the Sentry Team, see AUTHORS for more details.\n:license: BSD, see LICENSE for more details.\n\"\"\"\n\ntry:\n from celery.task import task\nexcept ImportError:\n from celery.decorators import task\nfrom celery.signals import task_failure\nfrom raven.base import Client\n\n\nclass CeleryMixin(object):\n def send_encoded(self, message):\n \"Errors through celery\"\n self.send_raw.delay(message)\n\n @task(routing_key='sentry')\n def send_raw(self, message):\n return super(CeleryMixin, self).send_encoded(message)\n\n\nclass CeleryClient(CeleryMixin, Client):\n pass\n\n\ndef register_signal(client):\n @task_failure.connect(weak=False)\n def process_failure_signal(sender, task_id, exception, args, kwargs,\n traceback, einfo, **kw):\n client.captureException(\n exc_info=einfo.exc_info,\n extra={\n 'task_id': task_id,\n 'task': sender,\n 'args': args,\n 'kwargs': kwargs,\n })\n\n","subject":"Fix celery task_failure signal definition","message":"Fix celery task_failure signal definition\n","lang":"Python","license":"bsd-3-clause","repos":"lepture\/raven-python,recht\/raven-python,lepture\/raven-python,beniwohli\/apm-agent-python,dbravender\/raven-python,patrys\/opbeat_python,recht\/raven-python,jbarbuto\/raven-python,getsentry\/raven-python,akalipetis\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,patrys\/opbeat_python,ewdurbin\/raven-python,nikolas\/raven-python,daikeren\/opbeat_python,arthurlogilab\/raven-python,icereval\/raven-python,ronaldevers\/raven-python,jmp0xf\/raven-python,danriti\/raven-python,danriti\/raven-python,smarkets\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,akheron\/raven-python,dbravender\/raven-python,smarkets\/raven-python,ticosax\/opbeat_python,lepture\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,lopter\/raven-python-old,icereval\/raven-python,ronaldevers\/raven-python,johansteffner\/raven-python,jmagnusson\/raven-python,tarkatronic\/opbeat_python,collective\/mr.poe,someonehan\/raven-python,jbarbuto\/raven-python,inspirehep\/raven-python,akalipetis\/raven-python,dirtycoder\/opbeat_python,smarkets\/raven-python,arthurlogilab\/raven-python,daikeren\/opbeat_python,Photonomie\/raven-python,inspirehep\/raven-python,beniwohli\/apm-agent-python,jbarbuto\/raven-python,daikeren\/opbeat_python,inspirehep\/raven-python,danriti\/raven-python,someonehan\/raven-python,nikolas\/raven-python,patrys\/opbeat_python,ewdurbin\/raven-python,icereval\/raven-python,akheron\/raven-python,jbarbuto\/raven-python,inspirehep\/raven-python,beniwohli\/apm-agent-python,someonehan\/raven-python,hzy\/raven-python,percipient\/raven-python,openlabs\/raven,ticosax\/opbeat_python,getsentry\/raven-python,tarkatronic\/opbeat_python,dirtycoder\/opbeat_python,recht\/raven-python,jmp0xf\/raven-python,jmagnusson\/raven-python,arthurlogilab\/raven-python,percipient\/raven-python,nikolas\/raven-python,getsentry\/raven-python,ronaldevers\/raven-python,johansteffner\/raven-python,akheron\/raven-python,hzy\/raven-python,ewdurbin\/raven-python,dirtycoder\/opbeat_python,percipient\/raven-python,jmagnusson\/raven-python,Goldmund-Wyldebeast-Wunderliebe\/raven-python,nikolas\/raven-python,johansteffner\/raven-python,icereval\/raven-python,hzy\/raven-python,smarkets\/raven-python,beniwohli\/apm-agent-python,Photonomie\/raven-python,Photonomie\/raven-python,jmp0xf\/raven-python,ticosax\/opbeat_python,tarkatronic\/opbeat_python,alex\/raven,arthurlogilab\/raven-python,akalipetis\/raven-python,dbravender\/raven-python,patrys\/opbeat_python"} {"commit":"9d68808a363ad00c3fc0b0907d625e5c75bdb8ae","old_file":"ptt_preproc_sampling.py","new_file":"ptt_preproc_sampling.py","old_contents":"#!\/usr\/bin\/env python\n\n\nfrom pathlib import Path\nfrom random import shuffle\nfrom shutil import copy\n\n\n# configs\n\nN = 10000\nSAMPLED_DIR_PATH = Path('sampled\/')\n\n# mkdir if doesn't exist\n\nif not SAMPLED_DIR_PATH.exists():\n SAMPLED_DIR_PATH.mkdir()\n\n# sample and copy\n\npaths = [p for p in Path('preprocessed\/').iterdir()]\nshuffle(paths)\n\nfor p in paths[:N]:\n copy(str(p), str(SAMPLED_DIR_PATH \/ p.name))\n","new_contents":"#!\/usr\/bin\/env python\n\n\nfrom pathlib import Path\nfrom random import sample\nfrom os import remove\n\n\n# configs\n\nN = 10000\n\n# remove unsampled\n\npaths = [path for path in Path('preprocessed\/').iterdir()]\npaths_len = len(paths)\n\nif paths_len <= N:\n raise RuntimeError('file count {:,} <= N {:,}'.format(paths_len, N))\n\nfor path in sample(paths, paths_len-N):\n remove(str(path))\n","subject":"Use removing rather than copying","message":"Use removing rather than copying\n","lang":"Python","license":"mit","repos":"moskytw\/mining-news"} {"commit":"b5e4af74bfc12eb3ae9ca14ab4cebc49daf05fdc","old_file":"api\/wb\/urls.py","new_file":"api\/wb\/urls.py","old_contents":"from django.conf.urls import url\nfrom api.wb import views\n\napp_name = 'osf'\n\nurlpatterns = [\n url(r'^move\/', views.MoveFile.as_view(), name=views.MoveFile.view_name),\n url(r'^copy\/', views.MoveFile.as_view(), name=views.MoveFile.view_name),\n]\n","new_contents":"from django.conf.urls import url\nfrom api.wb import views\n\napp_name = 'osf'\n\nurlpatterns = [\n url(r'^(?P<node_id>\\w+)\/move\/', views.MoveFile.as_view(), name=views.MoveFile.view_name),\n url(r'^(?P<node_id>\\w+)\/copy\/', views.MoveFile.as_view(), name=views.MoveFile.view_name),\n]\n","subject":"Add node id to url.","message":"Add node id to url.\n","lang":"Python","license":"apache-2.0","repos":"baylee-d\/osf.io,adlius\/osf.io,CenterForOpenScience\/osf.io,cslzchen\/osf.io,caseyrollins\/osf.io,erinspace\/osf.io,pattisdr\/osf.io,erinspace\/osf.io,icereval\/osf.io,adlius\/osf.io,erinspace\/osf.io,HalcyonChimera\/osf.io,HalcyonChimera\/osf.io,mattclark\/osf.io,adlius\/osf.io,felliott\/osf.io,Johnetordoff\/osf.io,felliott\/osf.io,felliott\/osf.io,HalcyonChimera\/osf.io,mattclark\/osf.io,adlius\/osf.io,cslzchen\/osf.io,Johnetordoff\/osf.io,mfraezz\/osf.io,aaxelb\/osf.io,aaxelb\/osf.io,saradbowman\/osf.io,icereval\/osf.io,saradbowman\/osf.io,cslzchen\/osf.io,Johnetordoff\/osf.io,mfraezz\/osf.io,caseyrollins\/osf.io,aaxelb\/osf.io,mfraezz\/osf.io,sloria\/osf.io,brianjgeiger\/osf.io,aaxelb\/osf.io,mfraezz\/osf.io,pattisdr\/osf.io,CenterForOpenScience\/osf.io,brianjgeiger\/osf.io,brianjgeiger\/osf.io,CenterForOpenScience\/osf.io,sloria\/osf.io,icereval\/osf.io,brianjgeiger\/osf.io,CenterForOpenScience\/osf.io,baylee-d\/osf.io,pattisdr\/osf.io,sloria\/osf.io,caseyrollins\/osf.io,HalcyonChimera\/osf.io,baylee-d\/osf.io,felliott\/osf.io,mattclark\/osf.io,Johnetordoff\/osf.io,cslzchen\/osf.io"} {"commit":"44893be528063d25d0b2305c9d24be4605c49f3c","old_file":"mcserver\/config\/core.py","new_file":"mcserver\/config\/core.py","old_contents":"\"\"\"\nMCServer Tools config loader\n\"\"\"\n\nimport json\nimport os.path\n\nclass CoreConfig(object):\n\t\"\"\"\n\tMCServer Tools configuration\n\t\"\"\"\n\n\tSETTINGS_FILE = 'mcserver.settings'\n\n\tdef __init__(self, path):\n\t\t\"\"\"\n\t\tLoad configuration from the given file path\n\t\t\"\"\"\n\n\t\tself.settings_file = os.path.join(path, self.SETTINGS_FILE)\n\t\tself._settings = {}\n\n\t\tself._load_settings()\n\n\tdef _load_settings(self):\n\t\t\"\"\"\n\t\tLoad the settings from disk\n\t\t\"\"\"\n\n\t\twith open(self.settings_file, 'r') as fh:\n\t\t\tself._settings = json.load(fh)\n\n\tdef get(self, property, default = None):\n\t\t\"\"\"\n\t\tTry to get the property value. If the property was not found\n\t\tthen return the given default.\n\t\t\"\"\"\n\n\t\tif property not in self._settings:\n\t\t\treturn default\n\n\t\treturn self._settings[property]\n\n\tdef has(self, property):\n\t\t\"\"\"\n\t\tCheck if the config has the given property.\n\t\t\"\"\"\n\n\t\treturn property in self._settings\n\n","new_contents":"\"\"\"\nMCServer Tools config loader\n\"\"\"\n\nimport json\nimport os.path\n\nfrom mcserver import MCServerError\n\nclass CoreConfig(object):\n\t\"\"\"\n\tMCServer Tools configuration\n\t\"\"\"\n\n\tSETTINGS_FILE = 'mcserver.settings'\n\n\tdef __init__(self, path):\n\t\t\"\"\"\n\t\tLoad configuration from the given file path\n\t\t\"\"\"\n\n\t\tself.settings_file = os.path.join(path, self.SETTINGS_FILE)\n\t\tself._settings = {}\n\n\t\tself._load_settings()\n\n\tdef _load_settings(self):\n\t\t\"\"\"\n\t\tLoad the settings from disk\n\t\t\"\"\"\n\n\t\ttry:\n\t\t\twith open(self.settings_file, 'r') as fh:\n\t\t\t\tself._settings = json.load(fh)\n\t\texcept:\n\t\t\traise MCServerError('Could not open settings file: {}'.format(self.settings_file))\n\n\tdef get(self, property, default = None):\n\t\t\"\"\"\n\t\tTry to get the property value. If the property was not found\n\t\tthen return the given default.\n\t\t\"\"\"\n\n\t\tif property not in self._settings:\n\t\t\treturn default\n\n\t\treturn self._settings[property]\n\n\tdef has(self, property):\n\t\t\"\"\"\n\t\tCheck if the config has the given property.\n\t\t\"\"\"\n\n\t\treturn property in self._settings\n\n","subject":"Check for the existance of the settings file and report if its not there","message":"Check for the existance of the settings file and report if its not there\n","lang":"Python","license":"mit","repos":"cadyyan\/mcserver-tools,cadyyan\/mcserver-tools"} {"commit":"20224e4fe8b93dee087dd7a455f9709b9795a026","old_file":"app\/models.py","new_file":"app\/models.py","old_contents":"from app import database\n\n\nclass Talk(database.Model):\n id = database.Column(database.Integer, primary_key=True, autoincrement=True)\n title = database.Column(database.String(128), nullable=False)\n description = database.Column(database.String(512))\n speaker_facebook_id = database.Column(database.BIGINT, database.ForeignKey('speaker.facebook_id'),\n nullable=False)\n\n liked_by = database.relationship('Liker_Talk', backref='talk', lazy='dynamic')\n\n def __repr__(self):\n return '<Talk %r>' % self.id\n\n\nclass Speaker(database.Model):\n facebook_id = database.Column(database.BIGINT, primary_key=True)\n name = database.Column(database.String(128), nullable=False)\n\n talks = database.relationship('Talk', backref='speaker', lazy='dynamic')\n\n def __repr__(self):\n return '<Speaker %r>' % self.facebook_id\n\n\nclass Liker_Talk(database.Model):\n liker_facebook_id = database.Column(database.BIGINT, primary_key=True)\n talk_id = database.Column(database.Integer, database.ForeignKey('talk.id'), primary_key=True)\n\n def __repr__(self):\n liker = repr(self.liker_facebook_id)\n talk = repr(self.talk_id)\n return '<Liker_Talk %r>' % ', '.join((liker, talk))\n","new_contents":"from app import database\n\n\nclass Talk(database.Model):\n id = database.Column(database.Integer, primary_key=True, autoincrement=True)\n title = database.Column(database.String(128), unique=True, nullable=False)\n description = database.Column(database.String(512))\n speaker_facebook_id = database.Column(database.BIGINT, database.ForeignKey('speaker.facebook_id'),\n nullable=False)\n\n liked_by = database.relationship('Liker_Talk', backref='talk', lazy='dynamic')\n\n def __repr__(self):\n return '<Talk %r>' % self.id\n\n\nclass Speaker(database.Model):\n facebook_id = database.Column(database.BIGINT, primary_key=True)\n name = database.Column(database.String(128), nullable=False)\n\n talks = database.relationship('Talk', backref='speaker', lazy='dynamic')\n\n def __repr__(self):\n return '<Speaker %r>' % self.facebook_id\n\n\nclass Liker_Talk(database.Model):\n liker_facebook_id = database.Column(database.BIGINT, primary_key=True)\n talk_id = database.Column(database.Integer, database.ForeignKey('talk.id'), primary_key=True)\n\n def __repr__(self):\n liker = repr(self.liker_facebook_id)\n talk = repr(self.talk_id)\n return '<Liker_Talk %r>' % ', '.join((liker, talk))\n","subject":"Make title unique Talk property","message":"Make title unique Talk property\n","lang":"Python","license":"mit","repos":"Stark-Mountain\/meetup-facebook-bot,Stark-Mountain\/meetup-facebook-bot"} {"commit":"3611e8a1b6477d251ddb2c90211e0cfee370671d","old_file":"cal_pipe\/easy_RFI_flagging.py","new_file":"cal_pipe\/easy_RFI_flagging.py","old_contents":"\nimport sys\nimport os\n\n'''\nEasier searching for good RFI flagging values\n'''\n\ntry:\n vis = sys.argv[1]\nexcept IndexError:\n vis = raw_input(\"Input vis? : \")\n\n# Just want the number of SPWs\ntb.open(os.path.join(vis, \"SPECTRAL_WINDOW\"))\nnchans = tb.getcol('NUM_CHAN')\ntb.close()\n\nspws = range(len(nchans))\n\ndefault('flagdata')\n\nfor spw in spws:\n print(\"On spw \"+str(spw)+\" of \"+str(len(nchans)))\n freqdevscale = 4.0\n timedevscale = 4.0\n while True:\n\n print(\"Starting at \")\n flagdata(vis=vis, mode='rflag', field='3C48*',\n spw=str(spw), datacolumn='corrected',\n action='calculate', display='both',\n freqdevscale=freqdevscale, timedevscale=timedevscale,\n flagbackup=False)\n\n adjust = True if raw_input(\"New thresholds? : \") == \"T\" else False\n\n if adjust:\n print(\"Current freqdevscale and timedevscale: %s %s\" % (freqdevscale, timedevscale))\n freqdevscale = float(raw_input(\"New freqdevscale : \"))\n timedevscale = float(raw_input(\"New timedevscale : \"))\n else:\n break\n","new_contents":"\nimport sys\nimport os\n\n'''\nEasier searching for good RFI flagging values\n'''\n\ntry:\n ms_name = sys.argv[1]\nexcept IndexError:\n ms_name = raw_input(\"Input vis? : \")\n\n# Just want the number of SPWs\ntb.open(os.path.join(ms_name, \"SPECTRAL_WINDOW\"))\nnchans = tb.getcol('NUM_CHAN')\ntb.close()\n\nspws = range(len(nchans))\n\ndefault('flagdata')\n\nfor spw in spws:\n print(\"On spw \"+str(spw)+\" of \"+str(len(nchans)))\n freqdevscale = 4.0\n timedevscale = 4.0\n while True:\n\n print(\"Starting at \")\n flagdata(vis=ms_name, mode='rflag', field='3C48*',\n spw=str(spw), datacolumn='corrected',\n action='calculate', display='both',\n freqdevscale=freqdevscale, timedevscale=timedevscale,\n flagbackup=False)\n\n adjust = True if raw_input(\"New thresholds? : \") == \"T\" else False\n\n if adjust:\n print(\"Current freqdevscale and timedevscale: %s %s\" % (freqdevscale, timedevscale))\n freqdevscale = float(raw_input(\"New freqdevscale : \"))\n timedevscale = float(raw_input(\"New timedevscale : \"))\n else:\n break\n","subject":"CHange name so it isn't reset","message":"CHange name so it isn't reset\n","lang":"Python","license":"mit","repos":"e-koch\/canfar_scripts,e-koch\/canfar_scripts"} {"commit":"be458ff809f6f49e21be06054ad096ff3f5430f9","old_file":"masters\/master.client.syzygy\/master_site_config.py","new_file":"masters\/master.client.syzygy\/master_site_config.py","old_contents":"# Copyright 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\n\"\"\"ActiveMaster definition.\"\"\"\n\nclass Syzygy(object):\n project_name = 'Syzygy'\n master_port = 8142\n slave_port = 8242\n master_port_alt = 8342\n tree_closing_notification_recipients = []\n from_address = 'buildbot@chromium.org'\n master_host = 'master3.golo.chromium.org'\n buildslave_version = 'buildbot_slave_8_4'\n twisted_version = 'twisted_10_2'\n","new_contents":"# Copyright 2013 The Chromium Authors. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\n\"\"\"ActiveMaster definition.\"\"\"\n\nclass Syzygy(object):\n project_name = 'Syzygy'\n master_port = 8042\n slave_port = 8142\n master_port_alt = 8242\n tree_closing_notification_recipients = []\n from_address = 'buildbot@chromium.org'\n master_host = 'master3.golo.chromium.org'\n buildslave_version = 'buildbot_slave_8_4'\n twisted_version = 'twisted_10_2'\n","subject":"Fix ports for syzygy master to match previous ports.","message":"Fix ports for syzygy master to match previous ports.\n\n\nTBR=chrisha@chromium.org\nBUG=\n\n\nReview URL: https:\/\/chromiumcodereview.appspot.com\/12315047\n\ngit-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@183944 0039d316-1c4b-4281-b951-d872f2087c98\n","lang":"Python","license":"bsd-3-clause","repos":"eunchong\/build,eunchong\/build,eunchong\/build,eunchong\/build"} {"commit":"80acc483f9b5d7fb462d81a2df092d16f5dbf035","old_file":"openprocurement\/tender\/limited\/subscribers.py","new_file":"openprocurement\/tender\/limited\/subscribers.py","old_contents":"from pyramid.events import subscriber\nfrom openprocurement.tender.core.events import TenderInitializeEvent\nfrom openprocurement.tender.core.utils import get_now, calculate_business_date\n\n\n@subscriber(TenderInitializeEvent, procurementMethodType=\"reporting\")\ndef tender_init_handler(event):\n \"\"\" initialization handler for tenders \"\"\"\n event.tender.date = get_now()\n\n\n@subscriber(TenderInitializeEvent, procurementMethodType=\"negotiation\")\ndef tender_init_handler(event):\n \"\"\" initialization handler for tenders \"\"\"\n tender = event.tender\n tender.date = get_now()\n if tender.lots:\n for lot in tender.lots:\n lot.date = get_now()\n\n\n@subscriber(TenderInitializeEvent, procurementMethodType=\"negotiation.quick\")\ndef tender_init_handler(event):\n \"\"\" initialization handler for tenders \"\"\"\n tender = event.tender\n tender.date = get_now()\n if tender.lots:\n for lot in tender.lots:\n lot.date = get_now()\n","new_contents":"from pyramid.events import subscriber\nfrom openprocurement.tender.core.events import TenderInitializeEvent\nfrom openprocurement.tender.core.utils import get_now, calculate_business_date\n\n\n@subscriber(TenderInitializeEvent, procurementMethodType=\"reporting\")\ndef tender_init_handler_1(event):\n \"\"\" initialization handler for tenders \"\"\"\n event.tender.date = get_now()\n\n\n@subscriber(TenderInitializeEvent, procurementMethodType=\"negotiation\")\ndef tender_init_handler_2(event):\n \"\"\" initialization handler for tenders \"\"\"\n tender = event.tender\n tender.date = get_now()\n if tender.lots:\n for lot in tender.lots:\n lot.date = get_now()\n\n\n@subscriber(TenderInitializeEvent, procurementMethodType=\"negotiation.quick\")\ndef tender_init_handler_3(event):\n \"\"\" initialization handler for tenders \"\"\"\n tender = event.tender\n tender.date = get_now()\n if tender.lots:\n for lot in tender.lots:\n lot.date = get_now()\n","subject":"Change tender init handlers names","message":"Change tender init handlers names\n","lang":"Python","license":"apache-2.0","repos":"openprocurement\/openprocurement.tender.limited"} {"commit":"689dd5cb67516fd091a69e39708b547c66f96750","old_file":"nap\/dataviews\/models.py","new_file":"nap\/dataviews\/models.py","old_contents":"\nfrom .fields import Field\nfrom .views import DataView\n\nfrom django.utils.six import with_metaclass\n\n\nclass MetaView(type):\n\n def __new__(mcs, name, bases, attrs):\n meta = attrs.get('Meta', None)\n\n try:\n model = meta.model\n except AttributeError:\n if name != 'ModelDataView':\n raise\n else:\n include = getattr(meta, 'fields', None)\n exclude = getattr(meta, 'exclude', [])\n\n # XXX Does the top base have all fields?\n\n for model_field in model._meta.fields:\n if model_field.name in attrs:\n continue\n if model_field.name in exclude:\n continue\n if include != '__all__' and model_field.name not in include:\n continue\n\n # XXX Magic for field types\n attrs[model_field.name] = Field(model_field.name)\n\n attrs['_meta'] = meta\n\n return super(MetaView, mcs).__new__(mcs, name, bases, attrs)\n\n\nclass ModelDataView(with_metaclass(MetaView, DataView)):\n\n pass\n","new_contents":"\nfrom django.db.models.fields import NOT_PROVIDED\nfrom django.utils.six import with_metaclass\n\nfrom . import filters\nfrom .fields import Field\nfrom .views import DataView\n\n\n# Map of ModelField name -> list of filters\nFIELD_FILTERS = {\n 'DateField': [filters.DateFilter],\n 'TimeField': [filters.TimeFilter],\n 'DateTimeField': [filters.DateTimeFilter],\n}\n\n\nclass Options(object):\n def __init__(self, meta):\n self.model = getattr(meta, 'model', None)\n self.fields = getattr(meta, 'fields', [])\n self.exclude = getattr(meta, 'exclude', [])\n self.required = getattr(meta, 'required', {})\n\n\nclass MetaView(type):\n\n def __new__(mcs, name, bases, attrs):\n meta = Options(attrs.get('Meta', None))\n\n if meta.model is None:\n if name != 'ModelDataView':\n raise ValueError('model not defined on class Meta')\n else:\n # XXX Does the top base have all fields?\n\n for model_field in meta.model._meta.fields:\n if model_field.name in attrs:\n continue\n if model_field.name in meta.exclude:\n continue\n if meta.fields != '__all__' and model_field.name not in meta.fields:\n continue\n\n # XXX Magic for field types\n kwargs = {}\n kwargs['default'] = model_field.default\n kwargs['required'] = any([\n not model_field.blank,\n model_field.default is not NOT_PROVIDED,\n ])\n kwargs['filters'] = FIELD_FILTERS.get(model_field.__class__.__name__, [])\n attrs[model_field.name] = Field(model_field.name, **kwargs)\n\n attrs['_meta'] = meta\n\n return super(MetaView, mcs).__new__(mcs, name, bases, attrs)\n\n\nclass ModelDataView(with_metaclass(MetaView, DataView)):\n\n pass\n","subject":"Add Options class Add field filters lists Start proper model field introspection","message":"Add Options class\nAdd field filters lists\nStart proper model field introspection\n","lang":"Python","license":"bsd-3-clause","repos":"limbera\/django-nap,MarkusH\/django-nap"} {"commit":"10e23fdd5c0427ad1ff5a5284410c755378a0e6d","old_file":"SoftLayer\/CLI\/object_storage\/list_accounts.py","new_file":"SoftLayer\/CLI\/object_storage\/list_accounts.py","old_contents":"\"\"\"List Object Storage accounts.\"\"\"\n# :license: MIT, see LICENSE for more details.\n\nimport click\n\nimport SoftLayer\nfrom SoftLayer.CLI import environment\nfrom SoftLayer.CLI import formatting\n\n\n@click.command()\n@environment.pass_env\ndef cli(env):\n \"\"\"List object storage accounts.\"\"\"\n\n mgr = SoftLayer.ObjectStorageManager(env.client)\n accounts = mgr.list_accounts()\n table = formatting.Table(['id', 'name', 'apiType'])\n table.sortby = 'id'\n global api_type\n for account in accounts:\n if 'vendorName' in account and 'Swift' == account['vendorName']:\n api_type = 'Swift'\n elif 'Cleversafe' in account['serviceResource']['name']:\n api_type = 'S3'\n\n table.add_row([\n account['id'],\n account['username'],\n api_type,\n ])\n\n env.fout(table)\n","new_contents":"\"\"\"List Object Storage accounts.\"\"\"\n# :license: MIT, see LICENSE for more details.\n\nimport click\n\nimport SoftLayer\nfrom SoftLayer.CLI import environment\nfrom SoftLayer.CLI import formatting\n\n\n@click.command()\n@environment.pass_env\ndef cli(env):\n \"\"\"List object storage accounts.\"\"\"\n\n mgr = SoftLayer.ObjectStorageManager(env.client)\n accounts = mgr.list_accounts()\n table = formatting.Table(['id', 'name', 'apiType'])\n table.sortby = 'id'\n api_type = None\n for account in accounts:\n if 'vendorName' in account and account['vendorName'] == 'Swift':\n api_type = 'Swift'\n elif 'Cleversafe' in account['serviceResource']['name']:\n api_type = 'S3'\n\n table.add_row([\n account['id'],\n account['username'],\n api_type,\n ])\n\n env.fout(table)\n","subject":"Fix object storage apiType for S3 and Swift.","message":"Fix object storage apiType for S3 and Swift.\n","lang":"Python","license":"mit","repos":"allmightyspiff\/softlayer-python,softlayer\/softlayer-python,kyubifire\/softlayer-python"} {"commit":"e9386e24bea91b8659b5184fe146002f555ccd15","old_file":"versions\/xmlib.py","new_file":"versions\/xmlib.py","old_contents":"#!\/usr\/bin\/env python\n#\n# Xm\n#\n# The xm library is hard to find and requires this special code.\n#\n# Author P G Jones - 11\/07\/2012 <p.g.jones@qmul.ac.uk> : First revision\n# Author P G Jones - 22\/09\/2012 <p.g.jones@qmul.ac.uk> : Major refactor of snoing.\n####################################################################################################\nimport systempackage\nimport system\nimport os\n\nclass Xm(systempackage.SystemPackage):\n \"\"\" Package for the Open Motif\/Xm library.\"\"\"\n def __init__(self, system):\n super(Xm, self).__init__(\"Xm\", system, \"Install Xm-dev (OpenMotif) on this system.\")\n def check_state(self):\n \"\"\" Check the Xm state, slightly more involved on macs.\"\"\"\n if self._system.get_os_type() == system.System.Mac:\n if os.path.exists(\"\/sw\/include\/Xm\"):\n flags = [ \"-I%s\" % \"\/sw\/include\/Xm\", \"-L%s\" % \"\/sw\/lib\" ]\n elif os.path.exists(\"\/usr\/OpenMotif\"):\n flags = [ \"-I%s\" % \"\/usr\/OpenMotif\/include\", \"-L%s\" % \"\/usr\/OpenMotif\/lib\" ]\n self._installed = self._system._test_compile([\"Xm.h\"], flags)\n else:\n self._installed = self._system.test_library(\"Xm\", [\"Xm\/Xm.h\"])\n","new_contents":"#!\/usr\/bin\/env python\n#\n# Xm\n#\n# The xm library is hard to find and requires this special code.\n#\n# Author P G Jones - 11\/07\/2012 <p.g.jones@qmul.ac.uk> : First revision\n# Author P G Jones - 22\/09\/2012 <p.g.jones@qmul.ac.uk> : Major refactor of snoing.\n####################################################################################################\nimport systempackage\nimport system\nimport os\n\nclass Xm(systempackage.SystemPackage):\n \"\"\" Package for the Open Motif\/Xm library.\"\"\"\n def __init__(self, system):\n super(Xm, self).__init__(\"Xm\", system, \"Install Xm-dev (OpenMotif) on this system.\")\n def check_state(self):\n \"\"\" Check the Xm state, slightly more involved on macs.\"\"\"\n if self._system.get_os_type() == system.System.Mac:\n if os.path.exists(\"\/sw\/include\/Xm\"):\n flags = [ \"-I%s\" % \"\/sw\/include\", \"-L%s\" % \"\/sw\/lib\" ]\n elif os.path.exists(\"\/usr\/OpenMotif\"):\n flags = [ \"-I%s\" % \"\/usr\/OpenMotif\/include\", \"-L%s\" % \"\/usr\/OpenMotif\/lib\" ]\n self._installed = self._system._test_compile([\"Xm\/Xm.h\"], flags)\n else:\n self._installed = self._system.test_library(\"Xm\", [\"Xm\/Xm.h\"])\n","subject":"Fix Xm library location error.","message":"Fix Xm library location error.\n","lang":"Python","license":"mit","repos":"mjmottram\/snoing,mjmottram\/snoing"} {"commit":"026aa257bff85b897e8e3ef1999b8fc6f7e3cc30","old_file":"socketdjango\/socketdjango\/__init__.py","new_file":"socketdjango\/socketdjango\/__init__.py","old_contents":"\"\"\"\nSocketdjango Project Module\n\nInteresting Docstring goes here!\n\"\"\"\n\n__version__ = '0.0.1'\n","new_contents":"\"\"\"\nSocketdjango Project Module\n\nInteresting Docstring goes here!\n\"\"\"\n\n__version__ = '0.1.0'\n","subject":"Change Initial Version Number to '0.1.0'","message":"Change Initial Version Number to '0.1.0'\n\nChange __version__ to '0.1.0'\n","lang":"Python","license":"mit","repos":"bobbyrussell\/django-socketio,bobbyrussell\/django-socketio,bobbyrussell\/django-socketio"} {"commit":"502a5cb7179aaedf68f3f16bf8d2ef7eb1ad0032","old_file":"nsq\/sockets\/__init__.py","new_file":"nsq\/sockets\/__init__.py","old_contents":"'''Sockets that wrap different connection types'''\n\n# Not all platforms support all types of sockets provided here. For those that\n# are not available, the corresponding socket wrapper is imported as None.\n\nfrom .. import logger\n\n# Snappy support\ntry:\n from .snappy import SnappySocket\nexcept ImportError: # pragma: no cover\n logger.warn('Snappy compression not supported')\n SnappySocket = None\n\n\n# Deflate support\ntry:\n from .deflate import DeflateSocket\nexcept ImportError: # pragma: no cover\n logger.warn('Deflate compression not supported')\n DeflateSocket = None\n\n\n# The TLS socket\ntry:\n from .tls import TLSSocket\nexcept ImportError: # pragma: no cover\n logger.warn('TLS not supported')\n TLSSocket = None\n","new_contents":"'''Sockets that wrap different connection types'''\n\n# Not all platforms support all types of sockets provided here. For those that\n# are not available, the corresponding socket wrapper is imported as None.\n\nfrom .. import logger\n\n# Snappy support\ntry:\n from .snappy import SnappySocket\nexcept ImportError: # pragma: no cover\n logger.debug('Snappy compression not supported')\n SnappySocket = None\n\n\n# Deflate support\ntry:\n from .deflate import DeflateSocket\nexcept ImportError: # pragma: no cover\n logger.debug('Deflate compression not supported')\n DeflateSocket = None\n\n\n# The TLS socket\ntry:\n from .tls import TLSSocket\nexcept ImportError: # pragma: no cover\n logger.warn('TLS not supported')\n TLSSocket = None\n","subject":"Reduce log severity of socket import messages","message":"Reduce log severity of socket import messages","lang":"Python","license":"mit","repos":"dlecocq\/nsq-py,dlecocq\/nsq-py"} {"commit":"7dfe9c435b102eacddd9e0617540495f0af46416","old_file":"app\/config.py","new_file":"app\/config.py","old_contents":"import os\n\n\nif os.environ['DATABASE_URL'] is None:\n SQLALCHEMY_DATABASE_URI = 'sqlite:\/\/\/meetup.db'\nelse:\n SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']\n\nSQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning\n","new_contents":"import os\n\n\nif os.environ.get('DATABASE_URL') is None:\n SQLALCHEMY_DATABASE_URI = 'sqlite:\/\/\/meetup.db'\nelse:\n SQLALCHEMY_DATABASE_URI = os.environ['DATABASE_URL']\n\nSQLALCHEMY_TRACK_MODIFICATIONS = False # supress deprecation warning\n","subject":"Fix the SQLite URL problem","message":"Fix the SQLite URL problem\n","lang":"Python","license":"mit","repos":"Stark-Mountain\/meetup-facebook-bot,Stark-Mountain\/meetup-facebook-bot"} {"commit":"28e0a10925d866572cae86507a3ace845fbff6a9","old_file":"observers\/middleware.py","new_file":"observers\/middleware.py","old_contents":"from .models import Observer\n\n\nclass ObserverMiddleware(object):\n \"\"\"\n Attaches an observer instance to every request coming from an\n authenticated user.\n \"\"\"\n def process_request(self, request):\n assert hasattr(request, 'user'), \"ObserverMiddleware requires auth middleware to be installed.\"\n if request.user and request.user.is_authenticated():\n request.observer = Observer.objects.get(user=request.user)\n else:\n request.observer = None\n","new_contents":"from .models import Observer\n\n\nclass ObserverMiddleware(object):\n \"\"\"\n Attaches an observer instance to every request coming from an\n authenticated user.\n \"\"\"\n def process_request(self, request):\n assert hasattr(request, 'user'), \"ObserverMiddleware requires auth middleware to be installed.\"\n if request.user and request.user.is_authenticated:\n request.observer = Observer.objects.get(user=request.user)\n else:\n request.observer = None\n","subject":"Use is_authenticated as a property.","message":"Use is_authenticated as a property.\n","lang":"Python","license":"mit","repos":"zsiciarz\/variablestars.net,zsiciarz\/variablestars.net,zsiciarz\/variablestars.net"} {"commit":"a50a46ee26e5d7d325a228559bc701c86d1b392d","old_file":"arg-reader.py","new_file":"arg-reader.py","old_contents":"#!\/usr\/bin\/env python3\n\n# References:\n# http:\/\/docs.python.org\/3.3\/library\/argparse.html?highlight=argparse#argparse\n# http:\/\/bip.weizmann.ac.il\/course\/python\/PyMOTW\/PyMOTW\/docs\/argparse\/index.html\n\nimport argparse\n\ndef main():\n '''\n Read arguments from a file\n '''\n\n parser = argparse.ArgumentParser(description='Script reads arguments from a file. Type $ .\/arg-reader.py @argsfilename e.g. $ .\/arg-reader.py @args.txt',\n fromfile_prefix_chars='@',\n )\n\n parser.add_argument('-animalbig', action=\"store\", dest=\"animalbig\",\n help = 'name of a big animal')\n parser.add_argument('-animalsmall', action=\"store\", dest=\"animalsmall\",\n help = 'name of a small animal')\n\n arguments = parser.parse_args()\n print(arguments)\n print(arguments.animalbig)\n print(arguments.animalsmall)\n\nif __name__ == \"__main__\": main()\n\n","new_contents":"#!\/usr\/bin\/env python3\n\n# References:\n# http:\/\/docs.python.org\/3.3\/library\/argparse.html?highlight=argparse#argparse\n# http:\/\/bip.weizmann.ac.il\/course\/python\/PyMOTW\/PyMOTW\/docs\/argparse\/index.html\n\nimport argparse\n\ndef main():\n '''\n For help, use argument -h\n $ .\/arg-reader.py -h\n To specify an argument, prefix with -\n $ .\/arg-reader.py -animalbig hippo -animalsmall fly\n To read arguments from a file, prefix file name with @\n $ .\/arg-reader.py @args2.txt\n To specify arguments from command line and from a file\n $ .\/arg-reader.py @args.txt -animalbig hippo\n '''\n\n parser = argparse.ArgumentParser(description='To read arguments from a file, prefix file name with @ e.g. $ .\/arg-reader.py @args.txt -animalbig hippo',\n fromfile_prefix_chars='@',\n )\n\n parser.add_argument('-animalbig', action=\"store\", dest=\"animalbig\",\n help = 'name of a big animal')\n parser.add_argument('-animalsmall', action=\"store\", dest=\"animalsmall\",\n help = 'name of a small animal')\n\n arguments = parser.parse_args()\n print(arguments)\n print(arguments.animalbig)\n print(arguments.animalsmall)\n\nif __name__ == \"__main__\": main()\n\n","subject":"Add more comments about usage.","message":"Add more comments about usage.\n","lang":"Python","license":"mit","repos":"beepscore\/argparse"} {"commit":"da05390fa11a12d0491caff18d38e71a1e134b82","old_file":"spicedham\/sqlalchemywrapper\/models.py","new_file":"spicedham\/sqlalchemywrapper\/models.py","old_contents":"from sqlalchemy import Column, Integer, String\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.schema import UniqueConstraint\n\nBase = declarative_base()\n\nclass Store(Base):\n __tablename__ = 'store'\n id = Column(Integer, primary_key=True, autoincrement=True)\n key = Column(String)\n tag = Column(String)\n value = Column(String)\n __table_args__ = (UniqueConstraint('key', 'tag'),)\n\n def __unicode__(self):\n return unicode(key)\n","new_contents":"from sqlalchemy import Column, Integer, String\nfrom sqlalchemy.ext.declarative import declarative_base\nfrom sqlalchemy.schema import PrimaryKeyConstraint\n\nBase = declarative_base()\n\nclass Store(Base):\n __tablename__ = 'store'\n key = Column(String)\n tag = Column(String)\n value = Column(String)\n __table_args__ = (PrimaryKeyConstraint('key', 'tag'),)\n\n def __unicode__(self):\n return unicode(key)\n","subject":"Make tag and key be a composite primary key","message":"Make tag and key be a composite primary key\n","lang":"Python","license":"mpl-2.0","repos":"mozilla\/spicedham,mozilla\/spicedham"} {"commit":"63d1eb69fc614cb3f019e7b37dd4ec10896c644e","old_file":"chartflo\/views.py","new_file":"chartflo\/views.py","old_contents":"# -*- coding: utf-8 -*-\n\nfrom django.views.generic import TemplateView\nfrom chartflo.factory import ChartDataPack\n\n\nclass ChartsView(TemplateView):\n template_name = 'chartflo\/charts.html'\n graph_type = \"pie\"\n title = \"\"\n\n def get_data(self):\n return {}\n\n def get_context_data(self, **kwargs):\n context = super(ChartsView, self).get_context_data(**kwargs)\n # get data\n P = ChartDataPack()\n dataset = self.get_data()\n # package the data\n datapack = P.package(\"chart_id\", self.title, dataset)\n # options\n datapack['legend'] = True\n datapack['export'] = False\n context['datapack'] = datapack\n context[\"graph_type\"] = self.graph_type\n context[\"title\"] = context[\"label\"] = self.title\n context[\"chart_url\"] = self._get_chart_url()\n return context\n\n def _get_chart_url(self):\n url = \"chartflo\/charts\/\" + self.graph_type + \".html\"\n return url\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom django.views.generic import TemplateView\nfrom chartflo.factory import ChartDataPack\n\n\nclass ChartsView(TemplateView):\n template_name = 'chartflo\/charts.html'\n chart_type = \"pie\"\n title = \"\"\n\n def get_data(self):\n return {}\n\n def get_context_data(self, **kwargs):\n context = super(ChartsView, self).get_context_data(**kwargs)\n # get data\n P = ChartDataPack()\n dataset = self.get_data()\n # package the data\n datapack = P.package(\"chart_id\", self.title, dataset)\n # options\n datapack['legend'] = True\n datapack['export'] = False\n context['datapack'] = datapack\n context[\"title\"] = context[\"label\"] = self.title\n context[\"chart_url\"] = self._get_chart_url()\n return context\n\n def _get_chart_url(self):\n url = \"chartflo\/charts\/\" + self.chart_type + \".html\"\n return url\n","subject":"Change graph_type for chart_type and remove it from context","message":"Change graph_type for chart_type and remove it from context\n","lang":"Python","license":"mit","repos":"synw\/django-chartflo,synw\/django-chartflo,synw\/django-chartflo"} {"commit":"e66468faaf9c4885f13545329baa20fe4914f49c","old_file":"historia.py","new_file":"historia.py","old_contents":"from eve import Eve\nfrom eve_swagger import swagger\nfrom eve.auth import BasicAuth\nfrom config import *\nfrom hashlib import md5\n\nclass MyBasicAuth(BasicAuth):\n def check_auth(self, username, password, allowed_roles, resource,\n method):\n accounts = app.data.driver.db['accounts']\n account = accounts.find_one({'username': username})\n return account and password == account['password']\n\ndef set_reporter(request, lookup):\n print request\n \napp = Eve(auth=MyBasicAuth)\napp.on_pre_PUT_event += set_reporter\n\napp.register_blueprint(swagger)\napp.config['SWAGGER_INFO'] = SWAGGER_INFO\napp.config['SWAGGER_HOST'] = SWAGGER_HOST\n\nif __name__ == '__main__':\n app.run(host=LISTEN_IP, port=LISTEN_PORT)\n","new_contents":"from eve import Eve\nfrom eve_swagger import swagger\nfrom eve.auth import BasicAuth\nfrom config import *\nfrom hashlib import md5\n\nclass MyBasicAuth(BasicAuth):\n def check_auth(self, username, password, allowed_roles, resource,\n method):\n accounts = app.data.driver.db['accounts']\n account = accounts.find_one({'username': username})\n return account and md5(password).hexdigest() == account['password']\n\ndef set_reporter(request, lookup):\n print request\n\napp = Eve(auth=MyBasicAuth)\napp.on_pre_PUT_event += set_reporter\n\napp.register_blueprint(swagger)\napp.config['SWAGGER_INFO'] = SWAGGER_INFO\napp.config['SWAGGER_HOST'] = SWAGGER_HOST\n\nif __name__ == '__main__':\n app.run(host=LISTEN_IP, port=LISTEN_PORT)\n","subject":"Use MD5 to encode passwords","message":"Use MD5 to encode passwords\n","lang":"Python","license":"mit","repos":"waoliveros\/historia"} {"commit":"4f9e51ff45f6faf6d0be6a442b4b04c3301026fe","old_file":"cloudenvy\/commands\/envy_snapshot.py","new_file":"cloudenvy\/commands\/envy_snapshot.py","old_contents":"from cloudenvy.envy import Envy\n\n\nclass EnvySnapshot(object):\n \"\"\"Create a snapshot of an ENVy.\"\"\"\n\n def __init__(self, argparser):\n self._build_subparser(argparser)\n\n def _build_subparser(self, subparsers):\n subparser = subparsers.add_parser('snapshot', help='snapshot help')\n subparser.set_defaults(func=self.run)\n\n return subparser\n\n #TODO(jakedahn): The entire UX for this needs to be talked about, refer to\n # https:\/\/github.com\/bcwaldon\/cloudenvy\/issues\/27 for any\n # discussion, if you're curious.\n def run(self, config, args):\n envy = Envy(config)\n envy.snapshot('%s-snapshot' % envy.name)\n","new_contents":"from cloudenvy.envy import Envy\n\n\nclass EnvySnapshot(object):\n \"\"\"Create a snapshot of an ENVy.\"\"\"\n\n def __init__(self, argparser):\n self._build_subparser(argparser)\n\n def _build_subparser(self, subparsers):\n subparser = subparsers.add_parser('snapshot', help='snapshot help')\n subparser.set_defaults(func=self.run)\n\n subparser.add_argument('-n', '--name', action='store', default='',\n help='Specify custom name for an ENVy.')\n\n return subparser\n\n #TODO(jakedahn): The entire UX for this needs to be talked about, refer to\n # https:\/\/github.com\/bcwaldon\/cloudenvy\/issues\/27 for any\n # discussion, if you're curious.\n def run(self, config, args):\n envy = Envy(config)\n envy.snapshot('%s-snapshot' % envy.name)\n","subject":"Add missing --name flag to 'envy snapshot'","message":"Add missing --name flag to 'envy snapshot'\n","lang":"Python","license":"apache-2.0","repos":"cloudenvy\/cloudenvy"} {"commit":"68724546ba4f6063559ba14b8625c7e7ecdf9732","old_file":"src\/read_key.py","new_file":"src\/read_key.py","old_contents":"#!\/usr\/bin\/python\n\ndef readKey(keyFileName):\n return open(\"..\/options-and-settings\/api-keys\/\" + keyFileName, \"r\").readline()\n\n","new_contents":"#!\/usr\/bin\/python\n\ndef readKey(keyFileName):\n return open(\"..\/options-and-settings\/api-keys\/\" + keyFileName, \"r\").readline().rstrip('\\n').rstrip('\\r')\n\n","subject":"Remove newline and carraige return characters from key files so that API calls work","message":"Remove newline and carraige return characters from key files so that API calls work\n","lang":"Python","license":"mit","repos":"nilnullzip\/StalkerBot,nilnullzip\/StalkerBot"} {"commit":"8121465e891b0d01aa050c11a84bd23058ce678e","old_file":"packagename\/__init__.py","new_file":"packagename\/__init__.py","old_contents":"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\n# Packages may add whatever they like to this file, but\n# should keep this content at the top.\n# ----------------------------------------------------------------------------\nfrom ._astropy_init import * # noqa\n# ----------------------------------------------------------------------------\n\n# Enforce Python version check during package import.\n# This is the same check as the one at the top of setup.py\nimport sys\nfrom distutils.version import LooseVersion\n\n__minimum_python_version__ = \"3.6\"\n\n\nclass UnsupportedPythonError(Exception):\n pass\n\n\nif LooseVersion(sys.version) < LooseVersion(__minimum_python_version__):\n raise UnsupportedPythonError(\"packagename does not support Python < {}\"\n .format(__minimum_python_version__))\n\nif not _ASTROPY_SETUP_: # noqa\n # For egg_info test builds to pass, put package imports here.\n from .example_mod import * #noqa\n","new_contents":"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n\n# Packages may add whatever they like to this file, but\n# should keep this content at the top.\n# ----------------------------------------------------------------------------\nfrom ._astropy_init import * # noqa\n# ----------------------------------------------------------------------------\n\n# Enforce Python version check during package import.\n# This is the same check as the one at the top of setup.py\nimport sys\nfrom distutils.version import LooseVersion\n\n__minimum_python_version__ = \"3.6\"\n\n\nclass UnsupportedPythonError(Exception):\n pass\n\n\nif LooseVersion(sys.version) < LooseVersion(__minimum_python_version__):\n raise UnsupportedPythonError(\"packagename does not support Python < {}\"\n .format(__minimum_python_version__))\n\nif not _ASTROPY_SETUP_: # noqa\n # For egg_info test builds to pass, put package imports here.\n from .example_mod import * # noqa\n","subject":"Update rendered version to f92c94967b6faae518bc9edaf114e652b3c61cde","message":"Update rendered version to f92c94967b6faae518bc9edaf114e652b3c61cde\n","lang":"Python","license":"bsd-3-clause","repos":"alexrudy\/Zeeko,alexrudy\/Zeeko"} {"commit":"b362d4b898493a856a810880079d3f44fe7d5d41","old_file":"project\/members\/tests\/test_application.py","new_file":"project\/members\/tests\/test_application.py","old_contents":"# -*- coding: utf-8 -*-\nimport pytest\nfrom django.core.urlresolvers import reverse\nfrom members.tests.fixtures.memberlikes import MembershipApplicationFactory\nfrom members.tests.fixtures.types import MemberTypeFactory\nfrom members.models import Member\n\n@pytest.mark.django_db\ndef test_application_approve():\n mtypes = [MemberTypeFactory(label='Normal member')]\n application = MembershipApplicationFactory()\n email = application.email\n application.approve(set_mtypes=mtypes)\n Member.objects.get(email=email)\n\n@pytest.mark.django_db\ndef test_get_application_form(client):\n response = client.get(reverse('members-apply'))\n assert b'Apply for membership' in response.content\n\n# TODO: Figure out a good way to submitting the form\n","new_contents":"# -*- coding: utf-8 -*-\nimport pytest\nfrom django.core.urlresolvers import reverse\nfrom members.tests.fixtures.memberlikes import MembershipApplicationFactory, MemberFactory\nfrom members.tests.fixtures.types import MemberTypeFactory\nfrom members.models import Member\n\n@pytest.mark.django_db\ndef test_application_approve():\n mtypes = [MemberTypeFactory(label='Normal member')]\n application = MembershipApplicationFactory()\n email = application.email\n application.approve(set_mtypes=mtypes)\n Member.objects.get(email=email)\n\n@pytest.mark.django_db\ndef test_get_application_form(client):\n response = client.get(reverse('members-apply'))\n assert b'Apply for membership' in response.content\n\n# TODO: Figure out a good way to submitting the form\n\n@pytest.mark.django_db\ndef test_get_admin_members_list(admin_client):\n # Create a test member\n member = MemberFactory()\n response = admin_client.get('\/admin\/members\/member\/')\n assert member.email in response.content.decode('utf-8')\n\n@pytest.mark.django_db\ndef test_get_admin_applications_list(admin_client):\n application = MembershipApplicationFactory()\n response = admin_client.get('\/admin\/members\/membershipapplication\/')\n assert application.email in response.content.decode('utf-8')\n","subject":"Add quick admin-site tests too","message":"Add quick admin-site tests too\n","lang":"Python","license":"mit","repos":"jautero\/asylum,jautero\/asylum,rambo\/asylum,hacklab-fi\/asylum,hacklab-fi\/asylum,HelsinkiHacklab\/asylum,jautero\/asylum,hacklab-fi\/asylum,rambo\/asylum,HelsinkiHacklab\/asylum,HelsinkiHacklab\/asylum,HelsinkiHacklab\/asylum,jautero\/asylum,rambo\/asylum,rambo\/asylum,hacklab-fi\/asylum"} {"commit":"ee12a7ee66600524cb56b779c217962e4974aead","old_file":"zappa\/__init__.py","new_file":"zappa\/__init__.py","old_contents":"import sys\n\nSUPPORTED_VERSIONS = [(2, 7), (3, 6)]\n\npython_major_version = sys.version_info[0]\npython_minor_version = sys.version_info[1]\n\nif (python_major_version, python_minor_version) not in SUPPORTED_VERSIONS:\n formatted_supported_versions = ['{}.{}'.format(mav, miv) for mav, miv in SUPPORTED_VERSIONS]\n err_msg = 'This version of Python ({}.{}) is not supported!\\n'.format(python_major_version, python_minor_version) +\\\n 'Zappa (and AWS Lambda) support the following versions of Python: {}'.format(formatted_supported_versions)\n raise RuntimeError(err_msg)\n\n__version__ = '0.46.1'\n","new_contents":"import sys\n\nSUPPORTED_VERSIONS = [(2, 7), (3, 6)]\n\npython_major_version = sys.version_info[0]\npython_minor_version = sys.version_info[1]\n\nif (python_major_version, python_minor_version) not in SUPPORTED_VERSIONS:\n formatted_supported_versions = ['{}.{}'.format(mav, miv) for mav, miv in SUPPORTED_VERSIONS]\n err_msg = 'This version of Python ({}.{}) is not supported!\\n'.format(python_major_version, python_minor_version) +\\\n 'Zappa (and AWS Lambda) support the following versions of Python: {}'.format(formatted_supported_versions)\n raise RuntimeError(err_msg)\n\n__version__ = '0.46.2'\n","subject":"Update version to 0.46.2 to match pypi release","message":"Update version to 0.46.2 to match pypi release","lang":"Python","license":"mit","repos":"Miserlou\/Zappa,scoates\/Zappa,pjz\/Zappa,pjz\/Zappa,Miserlou\/Zappa,scoates\/Zappa"} {"commit":"5456bee257cb36e4d1400da7e27480beadbf21fd","old_file":"examples\/arabic.py","new_file":"examples\/arabic.py","old_contents":"#!\/usr\/bin\/env python\n\"\"\"\nExample using Arabic\n===============\nGenerating a wordcloud from Arabic text\nOther dependencies: bidi.algorithm, arabic_reshaper\n\"\"\"\n\nfrom os import path\nimport codecs\nfrom wordcloud import WordCloud\nimport arabic_reshaper\nfrom bidi.algorithm import get_display\n\nd = path.dirname(__file__)\n\n# Read the whole text.\nf = codecs.open(path.join(d, 'arabicwords.txt'), 'r', 'utf-8')\n\n# Make text readable for a non-Arabic library like wordcloud\ntext = arabic_reshaper.reshape(f.read())\ntext = get_display(text)\n\n# Generate a word cloud image\nwordcloud = WordCloud(font_path='fonts\/NotoNaskhArabic\/NotoNaskhArabic-Regular.ttf').generate(text)\n\n# Export to an image\nwordcloud.to_file(\"arabic_example.png\")\n","new_contents":"#!\/usr\/bin\/env python\n\"\"\"\nCreate wordcloud with Arabic\n===============\nGenerating a wordcloud from Arabic text\nOther dependencies: bidi.algorithm, arabic_reshaper\n\"\"\"\n\nfrom os import path\nimport codecs\nfrom wordcloud import WordCloud\nimport arabic_reshaper\nfrom bidi.algorithm import get_display\n\nd = path.dirname(__file__)\n\n# Read the whole text.\nf = codecs.open(path.join(d, 'arabicwords.txt'), 'r', 'utf-8')\n\n# Make text readable for a non-Arabic library like wordcloud\ntext = arabic_reshaper.reshape(f.read())\ntext = get_display(text)\n\n# Generate a word cloud image\nwordcloud = WordCloud(font_path='fonts\/NotoNaskhArabic\/NotoNaskhArabic-Regular.ttf').generate(text)\n\n# Export to an image\nwordcloud.to_file(\"arabic_example.png\")\n","subject":"Change the title of the example","message":"Change the title of the example\n","lang":"Python","license":"mit","repos":"amueller\/word_cloud"} {"commit":"9e783b39e89e34ded032dc550bc8cc9016f1eded","old_file":"cacheops\/__init__.py","new_file":"cacheops\/__init__.py","old_contents":"VERSION = (3, 0, 1)\n__version__ = '.'.join(map(str, VERSION if VERSION[-1] else VERSION[:2]))\n\n\nfrom django.apps import AppConfig\n\nfrom .simple import *\nfrom .query import *\nfrom .invalidation import *\nfrom .templatetags.cacheops import *\nfrom .transaction import install_cacheops_transaction_support\n\n\nclass CacheopsConfig(AppConfig):\n name = 'cacheops'\n\n def ready(self):\n install_cacheops()\n install_cacheops_transaction_support()\n\ndefault_app_config = 'cacheops.CacheopsConfig'\n","new_contents":"VERSION = (3, 0, 1)\n__version__ = '.'.join(map(str, VERSION if VERSION[-1] else VERSION[:2]))\n\n\nfrom django.apps import AppConfig\n\nfrom .simple import *\nfrom .query import *\nfrom .invalidation import *\nfrom .templatetags.cacheops import *\nfrom .transaction import install_cacheops_transaction_support\nfrom .utils import debug_cache_key # noqa\n\n\nclass CacheopsConfig(AppConfig):\n name = 'cacheops'\n\n def ready(self):\n install_cacheops()\n install_cacheops_transaction_support()\n\ndefault_app_config = 'cacheops.CacheopsConfig'\n","subject":"Make debug_cache_key a part of API","message":"Make debug_cache_key a part of API\n","lang":"Python","license":"bsd-3-clause","repos":"LPgenerator\/django-cacheops,Suor\/django-cacheops"} {"commit":"843f84877d06329179f326600980eff0558e37e0","old_file":"report_qweb_pdf_watermark\/__manifest__.py","new_file":"report_qweb_pdf_watermark\/__manifest__.py","old_contents":"# © 2016 Therp BV <http:\/\/therp.nl>\n# License AGPL-3.0 or later (http:\/\/www.gnu.org\/licenses\/agpl.html).\n{\n \"name\": \"Pdf watermark\",\n \"version\": \"12.0.1.0.0\",\n \"author\": \"Therp BV, \"\n \"Odoo Community Association (OCA)\",\n \"license\": \"AGPL-3\",\n \"category\": \"Technical Settings\",\n \"summary\": \"Add watermarks to your QWEB PDF reports\",\n \"website\": \"https:\/\/github.com\/oca\/reporting-engine\",\n \"depends\": [\n 'web',\n ],\n \"data\": [\n \"demo\/report.xml\",\n \"views\/ir_actions_report_xml.xml\",\n \"views\/layout_templates.xml\",\n ],\n \"demo\": [\n \"demo\/report.xml\"\n ],\n \"intallable\": True,\n 'external_dependencies': {\n 'python': [\n 'PyPDF2',\n ],\n },\n}\n","new_contents":"# © 2016 Therp BV <http:\/\/therp.nl>\n# License AGPL-3.0 or later (http:\/\/www.gnu.org\/licenses\/agpl.html).\n{\n \"name\": \"Pdf watermark\",\n \"version\": \"12.0.1.0.0\",\n \"author\": \"Therp BV, \"\n \"Odoo Community Association (OCA)\",\n \"license\": \"AGPL-3\",\n \"category\": \"Technical Settings\",\n \"summary\": \"Add watermarks to your QWEB PDF reports\",\n \"website\": \"https:\/\/github.com\/oca\/reporting-engine\",\n \"depends\": [\n 'web',\n ],\n \"data\": [\n \"demo\/report.xml\",\n \"views\/ir_actions_report_xml.xml\",\n \"views\/layout_templates.xml\",\n ],\n \"demo\": [\n \"demo\/report.xml\"\n ],\n \"installable\": True,\n 'external_dependencies': {\n 'python': [\n 'PyPDF2',\n ],\n },\n}\n","subject":"Fix 'installable' syntax in manifest file","message":"[FIX] Fix 'installable' syntax in manifest file\n","lang":"Python","license":"agpl-3.0","repos":"OCA\/reporting-engine,OCA\/reporting-engine,OCA\/reporting-engine,OCA\/reporting-engine"} {"commit":"e35d55f46ffb9d42736ad4e57ae2a6c29838b054","old_file":"board\/tests.py","new_file":"board\/tests.py","old_contents":"from django.test import TestCase\n\n# Create your tests here.\n","new_contents":"from test_plus.test import TestCase\n\n\nclass BoardTest(TestCase):\n def test_get_board_list(self):\n board_list_url = self.reverse(\"board:list\")\n self.get_check_200(board_list_url)\n\n","subject":"Add board list test code.","message":"Add board list test code.\n","lang":"Python","license":"mit","repos":"9XD\/9XD,9XD\/9XD,9XD\/9XD,9XD\/9XD"} {"commit":"31fedddedc5ece0b7e68762269730e2cce110cb9","old_file":"pnnl\/models\/__init__.py","new_file":"pnnl\/models\/__init__.py","old_contents":"import importlib\nimport logging\nfrom volttron.platform.agent import utils\n\n_log = logging.getLogger(__name__)\nutils.setup_logging()\n__version__ = \"0.1\"\n\n__all__ = ['Model']\n\n\nclass Model(object):\n def __init__(self, config, **kwargs):\n self.model = None\n config = self.store_model_config(config)\n if not config:\n return\n base_module = \"volttron.pnnl.models.\"\n try:\n model_type = config[\"model_type\"]\n except KeyError as e:\n _log.exception(\"Missing Model Type key: {}\".format(e))\n raise e\n _file, model_type = model_type.split(\".\")\n module = importlib.import_module(base_module + _file)\n self.model_class = getattr(module, model_type)\n\n def get_q(self, _set, sched_index, market_index, occupied):\n q = self.model.predict(_set, sched_index, market_index, occupied)\n return q\n\n def store_model_config(self, _config):\n try:\n config = self.vip.config.get(\"model\")\n except KeyError:\n config = {}\n try:\n self.vip.config.set(\"model\", _config, send_update=False)\n except RuntimeError:\n _log.debug(\"Cannot change config store on config callback!\")\n _config.update(config)\n return _config\n\n","new_contents":"import importlib\nimport logging\nfrom volttron.platform.agent import utils\n\n_log = logging.getLogger(__name__)\nutils.setup_logging()\n__version__ = \"0.1\"\n\n__all__ = ['Model']\n\n\nclass Model(object):\n def __init__(self, config, **kwargs):\n self.model = None\n config = self.store_model_config(config)\n if not config:\n return\n base_module = \"volttron.pnnl.models.\"\n try:\n model_type = config[\"model_type\"]\n except KeyError as e:\n _log.exception(\"Missing Model Type key: {}\".format(e))\n raise e\n _file, model_type = model_type.split(\".\")\n module = importlib.import_module(base_module + _file)\n self.model_class = getattr(module, model_type)\n self.model = self.model_class(config, self)\n\n def get_q(self, _set, sched_index, market_index, occupied):\n q = self.model.predict(_set, sched_index, market_index, occupied)\n return q\n\n def store_model_config(self, _config):\n try:\n config = self.vip.config.get(\"model\")\n except KeyError:\n config = {}\n try:\n self.vip.config.set(\"model\", _config, send_update=False)\n except RuntimeError:\n _log.debug(\"Cannot change config store on config callback!\")\n _config.update(config)\n return _config\n\n","subject":"Fix self.model is not set.","message":"Fix self.model is not set.\n","lang":"Python","license":"bsd-3-clause","repos":"VOLTTRON\/volttron-applications,VOLTTRON\/volttron-applications,VOLTTRON\/volttron-applications,VOLTTRON\/volttron-applications,VOLTTRON\/volttron-applications"} {"commit":"9cd74ee3a8a162ff974c34321a557bf108476c0d","old_file":"runserver.py","new_file":"runserver.py","old_contents":"\"\"\"\nStarts up a development run of the codegolf server.\n\"\"\"\nfrom codegolf import app\n\napp.run(debug=True, host=\"localhost\")\n","new_contents":"\"\"\"\nStarts up a development run of the codegolf server.\n\"\"\"\nfrom codegolf import app\n\napp.run(debug=True, host=\"localhost\", port=9000)\n","subject":"Change port because I cbf to do anything nicer","message":"Change port because I cbf to do anything nicer\n","lang":"Python","license":"mit","repos":"UQComputingSociety\/codegolf,UQComputingSociety\/codegolf,UQComputingSociety\/codegolf"} {"commit":"cabc914567ee3d3b313e31282e8f20dd3ce15f9a","old_file":"derrida\/__init__.py","new_file":"derrida\/__init__.py","old_contents":"__version_info__ = (0, 9, 0)\n\n\n# Dot-connect all but the last. Last is dash-connected if not None.\n__version__ = '.'.join([str(i) for i in __version_info__[:-1]])\nif __version_info__[-1] is not None:\n __version__ += ('-%s' % (__version_info__[-1],))\n\n\n# context processor to add version to the template environment\ndef context_extras(request):\n return {\n # software version\n 'SW_VERSION': __version__\n }\n","new_contents":"__version_info__ = (1, 0, 0, 'dev')\n\n\n# Dot-connect all but the last. Last is dash-connected if not None.\n__version__ = '.'.join([str(i) for i in __version_info__[:-1]])\nif __version_info__[-1] is not None:\n __version__ += ('-%s' % (__version_info__[-1],))\n\n\n# context processor to add version to the template environment\ndef context_extras(request):\n return {\n # software version\n 'SW_VERSION': __version__\n }\n","subject":"Bump version to 1.0 dev","message":"Bump version to 1.0 dev\n","lang":"Python","license":"apache-2.0","repos":"Princeton-CDH\/derrida-django,Princeton-CDH\/derrida-django,Princeton-CDH\/derrida-django,Princeton-CDH\/derrida-django"} {"commit":"3d2b4536803df4a202d8c1c9b5d0e689f1053378","old_file":"tests\/config.py","new_file":"tests\/config.py","old_contents":"import sys\nsys.path.append('..\/ideascaly')\n\nfrom ideascaly.auth import AuthNonSSO\nfrom ideascaly.api import API\n\nimport unittest\n\ntesting_community = 'fiveheads.ideascale.com'\ntesting_token = '5b3326f8-50a5-419d-8f02-eef6a42fd61a'\n\n\nclass IdeascalyTestCase(unittest.TestCase):\n\n def setUp(self):\n self.auth = create_auth()\n self.api = API(self.auth)\n self.api.community_url = testing_community\n\n\ndef create_auth():\n auth = AuthNonSSO(testing_token)\n return auth","new_contents":"import os\nimport sys\nimport unittest\nsys.path.append('..\/ideascaly')\n\nfrom ideascaly.auth import AuthNonSSO\nfrom ideascaly.api import API\n\ntesting_community = 'fiveheads.ideascale.com'\ntesting_token = os.environ.get('TOKEN', '')\n\n\nclass IdeascalyTestCase(unittest.TestCase):\n\n def setUp(self):\n self.auth = create_auth()\n self.api = API(self.auth)\n self.api.community_url = testing_community\n\n\ndef create_auth():\n auth = AuthNonSSO(testing_token)\n return auth","subject":"Read token from environment variable","message":"Read token from environment variable\n","lang":"Python","license":"mit","repos":"joausaga\/ideascaly"} {"commit":"7845e017b264a38472d0dc94988a0afe6938132f","old_file":"tests\/acceptance\/conftest.py","new_file":"tests\/acceptance\/conftest.py","old_contents":"# -*- coding: utf-8 -*-\nimport mock\nimport pytest\n\n\n@pytest.fixture\ndef default_trace_id_generator(dummy_request):\n return lambda dummy_request: '17133d482ba4f605'\n\n\n@pytest.fixture\ndef settings():\n return {\n 'zipkin.tracing_percent': 100,\n 'zipkin.trace_id_generator': default_trace_id_generator,\n }\n\n\n@pytest.fixture\ndef get_span():\n return {\n 'id': '1',\n 'tags': {\n 'http.uri': '\/sample',\n 'http.uri.qs': '\/sample',\n 'http.route': '\/sample',\n 'response_status_code': '200',\n },\n 'name': 'GET \/sample',\n 'traceId': '17133d482ba4f605',\n 'localEndpoint': {\n 'ipv4': '127.0.0.1',\n 'port': 80,\n 'serviceName': 'acceptance_service',\n },\n 'kind': 'SERVER',\n 'timestamp': mock.ANY,\n 'duration': mock.ANY,\n }\n","new_contents":"# -*- coding: utf-8 -*-\nimport mock\nimport pytest\n\n\n@pytest.fixture\ndef default_trace_id_generator(dummy_request):\n return lambda dummy_request: '17133d482ba4f605'\n\n\n@pytest.fixture\ndef settings():\n return {\n 'zipkin.tracing_percent': 100,\n 'zipkin.trace_id_generator': default_trace_id_generator,\n }\n\n\n@pytest.fixture\ndef get_span():\n return {\n 'id': '1',\n 'tags': {\n 'http.uri': '\/sample',\n 'http.uri.qs': '\/sample',\n 'http.route': '\/sample',\n 'response_status_code': '200',\n },\n 'name': 'GET \/sample',\n 'traceId': '17133d482ba4f605',\n 'localEndpoint': {\n 'ipv4': mock.ANY,\n 'port': 80,\n 'serviceName': 'acceptance_service',\n },\n 'kind': 'SERVER',\n 'timestamp': mock.ANY,\n 'duration': mock.ANY,\n }\n","subject":"Allow any ip in the get_span expected span since it's not deterministic","message":"Allow any ip in the get_span expected span since it's not deterministic\n","lang":"Python","license":"apache-2.0","repos":"Yelp\/pyramid_zipkin"} {"commit":"c96e82caaa3fd560263c54db71772b44e9cd78d7","old_file":"examples\/upgrade_local_charm_k8s.py","new_file":"examples\/upgrade_local_charm_k8s.py","old_contents":"\"\"\"\nThis example:\n\n1. Connects to the current model\n2. Deploy a bundle and waits until it reports itself active\n3. Destroys the units and applications\n\n\"\"\"\nfrom juju import jasyncio\nfrom juju.model import Model\n\n\nasync def main():\n model = Model()\n print('Connecting to model')\n # Connect to current model with current user, per Juju CLI\n await model.connect()\n\n try:\n print('Deploying bundle')\n applications = await model.deploy(\n '.\/examples\/k8s-local-bundle\/bundle.yaml',\n )\n\n print('Waiting for active')\n await model.wait_for_idle(status='active')\n print(\"Successfully deployed!\")\n\n await applications[0].upgrade_charm(path='.\/examples\/charms\/onos.charm')\n\n await model.wait_for_idle(status='active')\n\n print('Removing bundle')\n for application in applications:\n await application.remove()\n finally:\n print('Disconnecting from model')\n await model.disconnect()\n print(\"Success\")\n\n\nif __name__ == '__main__':\n jasyncio.run(main())\n","new_contents":"\"\"\"\nThis example:\n\n1. Connects to the current model\n2. Deploy a bundle and waits until it reports itself active\n3. Upgrades the charm with a local path\n4. Destroys the units and applications\n\n\"\"\"\nfrom juju import jasyncio\nfrom juju.model import Model\n\n\nasync def main():\n model = Model()\n print('Connecting to model')\n # Connect to current model with current user, per Juju CLI\n await model.connect()\n\n try:\n print('Deploying bundle')\n applications = await model.deploy(\n '.\/examples\/k8s-local-bundle\/bundle.yaml',\n )\n\n print('Waiting for active')\n await model.wait_for_idle(status='active')\n print(\"Successfully deployed!\")\n\n local_path = '.\/examples\/charms\/onos.charm'\n print('Upgrading charm with %s' % local_path)\n await applications[0].upgrade_charm(path=local_path)\n\n await model.wait_for_idle(status='active')\n\n print('Removing bundle')\n for application in applications:\n await application.remove()\n finally:\n print('Disconnecting from model')\n await model.disconnect()\n print(\"Success\")\n\n\nif __name__ == '__main__':\n jasyncio.run(main())\n","subject":"Make the example more informative","message":"Make the example more informative\n","lang":"Python","license":"apache-2.0","repos":"juju\/python-libjuju,juju\/python-libjuju"} {"commit":"1b9b4365a46cdbfbfe88e2f5e271ba387fe4274f","old_file":"var_log_dieta\/constants.py","new_file":"var_log_dieta\/constants.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import, unicode_literals, division\n\nimport logging\n\nlogger = logging.getLogger(__name__) # pylint: disable=invalid-name\n\nDATA_DIR = 'data'\n\nDEFAULT_CONVERSIONS = {\n 'kg': {'g': 1000},\n 'l': {'ml': 1000},\n} # yapf: disable\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import, unicode_literals, division\n\nimport logging\n\nlogger = logging.getLogger(__name__) # pylint: disable=invalid-name\n\nDATA_DIR = 'data'\n\nDEFAULT_CONVERSIONS = {\n 'kg': {'g': 1000},\n 'l': {'ml': 1000},\n 'taza': {'ml': 250},\n 'tazon': {'ml': 350},\n 'vaso': {'ml': 300},\n} # yapf: disable\n","subject":"Add taza, tazon and vaso global conversions","message":"Add taza, tazon and vaso global conversions\n","lang":"Python","license":"bsd-3-clause","repos":"pignacio\/vld"} {"commit":"18da33bd5524a7e9a043de90fb9b7aa78a26412d","old_file":"addons\/meme.py","new_file":"addons\/meme.py","old_contents":"import discord\nimport random\nfrom discord.ext import commands\n\n\nclass Meme:\n \"\"\"\n Meme commands.\n \"\"\"\n\n def __init__(self, bot):\n self.bot = bot\n print('Addon \"{}\" loaded'.format(self.__class__.__name__))\n\n @commands.has_permissions(kick_members=True)\n @commands.command(pass_context=True, name=\"bam\")\n async def bam_member(self, ctx, user: discord.Member, *, reason=\"\"):\n \"\"\"Bams a user. Staff only.\"\"\"\n await self.bot.say(\"{} is ̶n͢ow b̕&̡.̷ 👍̡\".format(self.bot.escape_name(user)))\n\n @commands.has_permissions(kick_members=True)\n @commands.command(pass_context=True, name=\"warm\")\n async def warm_member(self, ctx, user: discord.Member, *, reason=\"\"):\n \"\"\"Warms a user :3. Staff only.\"\"\"\n await self.bot.say(\"{} warmed. User is now {}°C.\".format(user.mention, str(random.randint(0, 100))))\n\n\ndef setup(bot):\n bot.add_cog(Meme(bot))\n","new_contents":"import discord\nimport random\nfrom discord.ext import commands\n\n\nclass Meme:\n \"\"\"\n Meme commands.\n \"\"\"\n\n def __init__(self, bot):\n self.bot = bot\n print('Addon \"{}\" loaded'.format(self.__class__.__name__))\n\n @commands.command(pass_context=True, hidden=True, name=\"bam\")\n async def bam_member(self, ctx, user: discord.Member, *, reason=\"\"):\n \"\"\"Bams a user owo\"\"\"\n await self.bot.say(\"{} is ̶n͢ow b̕&̡.̷ 👍̡\".format(self.bot.escape_name(user)))\n\n @commands.command(pass_context=True, hidden=True, name=\"warm\")\n async def warm_member(self, ctx, user: discord.Member, *, reason=\"\"):\n \"\"\"Warms a user :3\"\"\"\n await self.bot.say(\"{} warmed. User is now {}°C.\".format(user.mention, str(random.randint(0, 100))))\n\n\ndef setup(bot):\n bot.add_cog(Meme(bot))\n","subject":"Allow everyone to bam and warm, hide commands","message":"Allow everyone to bam and warm, hide commands","lang":"Python","license":"apache-2.0","repos":"916253\/Kurisu-Reswitched"} {"commit":"b82fc6f21245cba7fadb35a6676433f015aad516","old_file":"tripleo_common\/utils\/tarball.py","new_file":"tripleo_common\/utils\/tarball.py","old_contents":"# Copyright 2016 Red Hat, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nimport logging\n\nfrom oslo_concurrency import processutils\n\nLOG = logging.getLogger(__name__)\n\n\ndef create_tarball(directory, filename, options='-czf'):\n \"\"\"Create a tarball of a directory.\"\"\"\n LOG.debug('Creating tarball of %s at location %s' % (directory, filename))\n processutils.execute('\/usr\/bin\/tar', '-C', directory, options, filename,\n '--exclude', '.git', '--exclude', '.tox', '.')\n\n\ndef tarball_extract_to_swift_container(object_client, filename, container):\n LOG.debug('Uploading filename %s to Swift container %s' % (filename,\n container))\n with open(filename, 'r') as f:\n object_client.put_object(\n container=container,\n obj='',\n contents=f,\n query_string='extract-archive=tar.gz',\n headers={'X-Detect-Content-Type': 'true'}\n )\n","new_contents":"# Copyright 2016 Red Hat, Inc.\n# All Rights Reserved.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\nimport logging\n\nfrom oslo_concurrency import processutils\n\nLOG = logging.getLogger(__name__)\n\n\ndef create_tarball(directory, filename, options='-czf'):\n \"\"\"Create a tarball of a directory.\"\"\"\n LOG.debug('Creating tarball of %s at location %s' % (directory, filename))\n processutils.execute('\/usr\/bin\/tar', '-C', directory, options, filename,\n '--exclude', '.git', '--exclude', '.tox',\n '--exclude', '*.pyc', '--exclude', '*.pyo', '.')\n\n\ndef tarball_extract_to_swift_container(object_client, filename, container):\n LOG.debug('Uploading filename %s to Swift container %s' % (filename,\n container))\n with open(filename, 'r') as f:\n object_client.put_object(\n container=container,\n obj='',\n contents=f,\n query_string='extract-archive=tar.gz',\n headers={'X-Detect-Content-Type': 'true'}\n )\n","subject":"Exclude more unneeded files from default plan","message":"Exclude more unneeded files from default plan\n\nThis patch exludes more file types from the tarball uploaded to swift as\nthe default deployment plan.\n\nChange-Id: I8b6d8de8d7662604cdb871fa6a4fb872c7937e25\nCloses-Bug: #1613286\n","lang":"Python","license":"apache-2.0","repos":"openstack\/tripleo-common,openstack\/tripleo-common"} {"commit":"7ac384be36e22919a15fc7d25de25aa7afcd9382","old_file":"statscache\/consumer.py","new_file":"statscache\/consumer.py","old_contents":"import copy\n\nimport fedmsg.consumers\n\nimport logging\nlog = logging.getLogger(\"fedmsg\")\n\nclass StatsConsumer(fedmsg.consumers.FedmsgConsumer):\n \"\"\"\n The actual 'cache' of statscache that accumulates messages to be processed.\n \"\"\"\n topic = '*'\n config_key = 'statscache.consumer.enabled'\n\n def __init__(self, *args, **kwargs):\n \"\"\" Instantiate the consumer and a default list of buckets \"\"\"\n log.debug(\"statscache consumer initializing\")\n super(StatsConsumer, self).__init__(*args, **kwargs)\n log.debug(\"statscache consumer initialized\")\n self.buckets = {\n 'OneSecond': [],\n 'FiveSecond': [],\n 'OneMinute': [],\n }\n\n def consume(self, raw_msg):\n \"\"\" Receive a message and enqueue it onto each bucket \"\"\"\n topic, msg = raw_msg['topic'], raw_msg['body']\n log.info(\"Got message %r\", topic)\n for name, bucket in self.buckets.items():\n bucket.append(copy.deepcopy(msg))\n\n def stop(self):\n log.info(\"Cleaning up StatsConsumer.\")\n super(StatsConsumer, self).stop()\n","new_contents":"import copy\n\nimport fedmsg.consumers\n\nimport logging\nlog = logging.getLogger(\"fedmsg\")\n\nclass StatsConsumer(fedmsg.consumers.FedmsgConsumer):\n \"\"\"\n The actual 'cache' of statscache that accumulates messages to be processed.\n \"\"\"\n topic = '*'\n config_key = 'statscache.consumer.enabled'\n\n def __init__(self, *args, **kwargs):\n \"\"\" Instantiate the consumer and a default list of buckets \"\"\"\n log.debug(\"statscache consumer initializing\")\n super(StatsConsumer, self).__init__(*args, **kwargs)\n log.debug(\"statscache consumer initialized\")\n self.buckets = {\n 'OneSecond': [],\n 'FiveSecond': [],\n 'OneMinute': [],\n 'OneDay': [],\n }\n\n def consume(self, raw_msg):\n \"\"\" Receive a message and enqueue it onto each bucket \"\"\"\n topic, msg = raw_msg['topic'], raw_msg['body']\n log.info(\"Got message %r\", topic)\n for name, bucket in self.buckets.items():\n bucket.append(copy.deepcopy(msg))\n\n def stop(self):\n log.info(\"Cleaning up StatsConsumer.\")\n super(StatsConsumer, self).stop()\n","subject":"Create missing bucket for one-day frequency","message":"Create missing bucket for one-day frequency\n","lang":"Python","license":"lgpl-2.1","repos":"yazman\/statscache,yazman\/statscache,yazman\/statscache"} {"commit":"af5a9e3f1e5f89836c8af6aab274d726170deb8b","old_file":"utils\/lit\/lit\/LitTestCase.py","new_file":"utils\/lit\/lit\/LitTestCase.py","old_contents":"import unittest\nimport Test\n\n\"\"\"\nTestCase adaptor for providing a 'unittest' compatible interface to 'lit' tests.\n\"\"\"\n\nclass UnresolvedError(RuntimeError):\n pass\n \nclass LitTestCase(unittest.TestCase):\n def __init__(self, test, lit_config):\n unittest.TestCase.__init__(self)\n self._test = test\n self._lit_config = lit_config\n\n def id(self):\n return self._test.getFullName()\n\n def shortDescription(self):\n return self._test.getFullName()\n\n def runTest(self):\n tr, output = self._test.config.test_format.execute(\n self._test, self._lit_config)\n\n if tr is Test.UNRESOLVED:\n raise UnresolvedError(output)\n elif tr.isFailure:\n self.fail(output)\n","new_contents":"from __future__ import absolute_import\nimport unittest\n\nimport lit.Test\n\n\"\"\"\nTestCase adaptor for providing a 'unittest' compatible interface to 'lit' tests.\n\"\"\"\n\nclass UnresolvedError(RuntimeError):\n pass\n \nclass LitTestCase(unittest.TestCase):\n def __init__(self, test, lit_config):\n unittest.TestCase.__init__(self)\n self._test = test\n self._lit_config = lit_config\n\n def id(self):\n return self._test.getFullName()\n\n def shortDescription(self):\n return self._test.getFullName()\n\n def runTest(self):\n tr, output = self._test.config.test_format.execute(\n self._test, self._lit_config)\n\n if tr is lit.Test.UNRESOLVED:\n raise UnresolvedError(output)\n elif tr.isFailure:\n self.fail(output)\n","subject":"Fix a relative import issue I missed earlier.","message":"[lit] Fix a relative import issue I missed earlier.\n\ngit-svn-id: 0ff597fd157e6f4fc38580e8d64ab130330d2411@188360 91177308-0d34-0410-b5e6-96231b3b80d8\n","lang":"Python","license":"bsd-2-clause","repos":"dslab-epfl\/asap,apple\/swift-llvm,chubbymaggie\/asap,llvm-mirror\/llvm,GPUOpen-Drivers\/llvm,chubbymaggie\/asap,apple\/swift-llvm,dslab-epfl\/asap,chubbymaggie\/asap,llvm-mirror\/llvm,llvm-mirror\/llvm,chubbymaggie\/asap,llvm-mirror\/llvm,dslab-epfl\/asap,apple\/swift-llvm,apple\/swift-llvm,llvm-mirror\/llvm,llvm-mirror\/llvm,apple\/swift-llvm,dslab-epfl\/asap,dslab-epfl\/asap,GPUOpen-Drivers\/llvm,GPUOpen-Drivers\/llvm,chubbymaggie\/asap,GPUOpen-Drivers\/llvm,llvm-mirror\/llvm,apple\/swift-llvm,chubbymaggie\/asap,dslab-epfl\/asap,llvm-mirror\/llvm,GPUOpen-Drivers\/llvm,GPUOpen-Drivers\/llvm,dslab-epfl\/asap,GPUOpen-Drivers\/llvm,GPUOpen-Drivers\/llvm,apple\/swift-llvm,llvm-mirror\/llvm,apple\/swift-llvm"} {"commit":"8cbc55794d67571831ccc22b1ccdcf716362d814","old_file":"tests\/test_hmmsearch3.py","new_file":"tests\/test_hmmsearch3.py","old_contents":"import os\nimport unittest\nimport sys\n\n# hack to allow tests to find inmembrane in directory above\nmodule_dir = os.path.abspath(os.path.dirname(__file__))\nsys.path.insert(0, os.path.join(module_dir, '..'))\n\nimport inmembrane \n\n\nclass TestHmmsearch3(unittest.TestCase):\n def setUp(self):\n self.dir = os.path.join(module_dir, 'hmmsearch3')\n\n def test_hmmsearch3(self):\n save_dir = os.getcwd()\n os.chdir(self.dir)\n\n inmembrane.silence_log(True)\n \n self.params = inmembrane.get_params()\n self.params['fasta'] = \"hmmsearch3.fasta\"\n self.params['hmm_profiles_dir'] = \"..\/..\/protocols\/gram_neg_profiles\"\n self.seqids, self.proteins = \\\n inmembrane.create_protein_data_structure(self.params['fasta'])\n inmembrane.hmmsearch3(self.params, self.proteins)\n\n self.expected_output = {\n u'SPy_0128': ['LPxTG'], \n u'SPy_0191a': ['SLH_ls'], \n }\n for seqid in self.expected_output:\n for motif in self.expected_output[seqid]:\n self.assertTrue(motif in self.proteins[seqid]['hmmsearch'])\n\n os.chdir(save_dir)\n\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"import os\nimport unittest\nimport sys\n\n# hack to allow tests to find inmembrane in directory above\nmodule_dir = os.path.abspath(os.path.dirname(__file__))\nsys.path.insert(0, os.path.join(module_dir, '..'))\n\nimport inmembrane \n\n\nclass TestHmmsearch3(unittest.TestCase):\n def setUp(self):\n self.dir = os.path.join(module_dir, 'hmmsearch3')\n\n def test_hmmsearch3(self):\n save_dir = os.getcwd()\n os.chdir(self.dir)\n\n inmembrane.silence_log(True)\n \n self.params = inmembrane.get_params()\n self.params['fasta'] = \"hmmsearch3.fasta\"\n self.params['hmm_profiles_dir'] = \"..\/..\/protocols\/gram_pos_profiles\"\n self.seqids, self.proteins = \\\n inmembrane.create_protein_data_structure(self.params['fasta'])\n inmembrane.hmmsearch3(self.params, self.proteins)\n\n self.expected_output = {\n u'SPy_0128': ['LPxTG'], \n u'SPy_0191a': ['SLH_ls'], \n }\n for seqid in self.expected_output:\n for motif in self.expected_output[seqid]:\n self.assertTrue(motif in self.proteins[seqid]['hmmsearch'])\n\n os.chdir(save_dir)\n\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Put correct directory for profiles in test_hmmsearch","message":"Put correct directory for profiles in test_hmmsearch\n","lang":"Python","license":"bsd-2-clause","repos":"boscoh\/inmembrane"} {"commit":"cbadf5c564d7f5f701499409e2ae77ff90ba477c","old_file":"tests\/test_tensorflow.py","new_file":"tests\/test_tensorflow.py","old_contents":"import unittest\n\nimport numpy as np\nimport tensorflow as tf\n\nfrom common import gpu_test\n\n\nclass TestTensorflow(unittest.TestCase):\n def test_addition(self): \n op = tf.add(2, 3) \n sess = tf.Session()\n\n result = sess.run(op)\n\n self.assertEqual(5, result)\n \n @gpu_test\n def test_gpu(self):\n with tf.device('\/gpu:0'):\n m1 = tf.constant([2.0, 3.0], shape=[1, 2], name='a')\n m2 = tf.constant([3.0, 4.0], shape=[2, 1], name='b')\n op = tf.matmul(m1, m2)\n\n sess = tf.Session()\n result = sess.run(op)\n\n self.assertEqual(np.array(18, dtype=np.float32, ndmin=2), result)\n","new_contents":"import unittest\n\nimport numpy as np\nimport tensorflow as tf\n\nfrom common import gpu_test\n\n\nclass TestTensorflow(unittest.TestCase):\n def test_addition(self): \n op = tf.add(2, 3) \n sess = tf.Session()\n\n result = sess.run(op)\n\n self.assertEqual(5, result)\n\n def test_conv2d(self):\n input = tf.random_normal([1,2,2,1])\n filter = tf.random_normal([1,1,1,1])\n\n op = tf.nn.conv2d(input, filter, strides=[1, 1, 1, 1], padding='SAME')\n with tf.Session() as sess:\n result = sess.run(op)\n self.assertEqual(4, len(result.shape))\n \n @gpu_test\n def test_gpu(self):\n with tf.device('\/gpu:0'):\n m1 = tf.constant([2.0, 3.0], shape=[1, 2], name='a')\n m2 = tf.constant([3.0, 4.0], shape=[2, 1], name='b')\n op = tf.matmul(m1, m2)\n\n sess = tf.Session()\n result = sess.run(op)\n\n self.assertEqual(np.array(18, dtype=np.float32, ndmin=2), result)\n","subject":"Add conv2d test for tensorflow","message":"Add conv2d test for tensorflow\n","lang":"Python","license":"apache-2.0","repos":"Kaggle\/docker-python,Kaggle\/docker-python"} {"commit":"dfd4a6f6b23447538b2b22da11666f5218d791db","old_file":"mots_vides\/constants.py","new_file":"mots_vides\/constants.py","old_contents":"\"\"\"\nConstants for mots-vides\n\"\"\"\nimport os\n\nDATA_DIRECTORY = os.path.join(\n os.path.dirname(\n os.path.abspath(__file__)),\n 'datas\/'\n)\n","new_contents":"\"\"\"\nConstants for mots-vides\n\"\"\"\nimport os\n\nDATA_DIRECTORY = os.path.join(\n os.path.dirname(\n os.path.abspath(__file__)),\n 'datas\/'\n)\n\nLANGUAGE_CODES = {\n 'af': 'afrikaans',\n 'ar': 'arabic',\n 'az': 'azerbaijani',\n 'bg': 'bulgarian',\n 'be': 'belarusian',\n 'bn': 'bengali',\n 'br': 'breton',\n 'bs': 'bosnian',\n 'ca': 'catalan',\n 'cs': 'czech',\n 'cy': 'welsh',\n 'da': 'danish',\n 'de': 'german',\n 'el': 'greek',\n 'en': 'english',\n 'eo': 'esperanto',\n 'es': 'spanish',\n 'et': 'estonian',\n 'eu': 'basque',\n 'fa': 'persian',\n 'fi': 'finnish',\n 'fr': 'french',\n 'fy': 'frisian',\n 'ga': 'irish',\n 'gl': 'galician',\n 'he': 'hebrew',\n 'hi': 'hindi',\n 'hr': 'croatian',\n 'hu': 'hungarian',\n 'ia': 'interlingua',\n 'id': 'indonesian',\n 'io': 'ido',\n 'is': 'icelandic',\n 'it': 'italian',\n 'ja': 'japanese',\n 'ka': 'georgian',\n 'kk': 'kazakh',\n 'km': 'khmer',\n 'kn': 'kannada',\n 'ko': 'korean',\n 'lb': 'luxembourgish',\n 'lt': 'lithuanian',\n 'lv': 'latvian',\n 'mk': 'macedonian',\n 'ml': 'malayalam',\n 'mn': 'mongolian',\n 'mr': 'marathi',\n 'my': 'burmese',\n 'nb': 'norwegian',\n 'ne': 'nepali',\n 'nl': 'dutch',\n 'os': 'ossetic',\n 'pa': 'punjabi',\n 'pl': 'polish',\n 'pt': 'portuguese',\n 'ro': 'romanian',\n 'ru': 'russian',\n 'sk': 'slovak',\n 'sl': 'slovenian',\n 'sq': 'albanian',\n 'sr': 'serbian',\n 'sv': 'swedish',\n 'sw': 'swahili',\n 'ta': 'tamil',\n 'te': 'telugu',\n 'th': 'thai',\n 'tr': 'turkish',\n 'tt': 'tatar',\n 'uk': 'ukrainian',\n 'ur': 'urdu',\n 'vi': 'vietnamese',\n 'zh': 'chinese',\n}\n","subject":"Define a complete list of language code, for easy future maintenance","message":"Define a complete list of language code, for easy future maintenance\n","lang":"Python","license":"bsd-3-clause","repos":"Fantomas42\/mots-vides,Fantomas42\/mots-vides"} {"commit":"8dc6c7567f9bc94dc1b4a96b80d059f1231039bc","old_file":"st2auth_flat_file_backend\/__init__.py","new_file":"st2auth_flat_file_backend\/__init__.py","old_contents":"# Licensed to the StackStorm, Inc ('StackStorm') under one or more\n# contributor license agreements. See the NOTICE file distributed with\n# this work for additional information regarding copyright ownership.\n# The ASF licenses this file to You under the Apache License, Version 2.0\n# (the \"License\"); you may not use this file except in compliance with\n# the License. You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom flat_file import FlatFileAuthenticationBackend\n\n__all__ = [\n 'FlatFileAuthenticationBackend'\n]\n\n__version__ = '0.1.0'\n","new_contents":"# Licensed to the StackStorm, Inc ('StackStorm') under one or more\n# contributor license agreements. See the NOTICE file distributed with\n# this work for additional information regarding copyright ownership.\n# The ASF licenses this file to You under the Apache License, Version 2.0\n# (the \"License\"); you may not use this file except in compliance with\n# the License. You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import absolute_import\n\nfrom .flat_file import FlatFileAuthenticationBackend\n\n__all__ = [\n 'FlatFileAuthenticationBackend'\n]\n\n__version__ = '0.1.0'\n","subject":"Fix code so it also works under Python 3.","message":"Fix code so it also works under Python 3.\n","lang":"Python","license":"apache-2.0","repos":"StackStorm\/st2-auth-backend-flat-file"} {"commit":"4abd7baafcd982993471d5c0137d4b506ea49e8b","old_file":"src\/runcommands\/util\/enums.py","new_file":"src\/runcommands\/util\/enums.py","old_contents":"import enum\nimport os\nimport subprocess\nimport sys\n\nimport blessings\n\nfrom .misc import isatty\n\n\nif isatty(sys.stdout) and os.getenv(\"TERM\"):\n Terminal = blessings.Terminal\nelse:\n\n class Terminal:\n def __getattr__(self, name):\n return \"\"\n\n\nTERM = Terminal()\n\n\nclass Color(enum.Enum):\n\n none = \"\"\n reset = TERM.normal\n black = TERM.black\n red = TERM.red\n green = TERM.green\n yellow = TERM.yellow\n blue = TERM.blue\n magenta = TERM.magenta\n cyan = TERM.cyan\n white = TERM.white\n\n def __str__(self):\n return self.value\n\n\nclass StreamOptions(enum.Enum):\n\n \"\"\"Choices for stream handling.\"\"\"\n\n capture = \"capture\"\n hide = \"hide\"\n none = \"none\"\n\n @property\n def option(self):\n return {\n \"capture\": subprocess.PIPE,\n \"hide\": subprocess.DEVNULL,\n \"none\": None,\n }[self.value]\n","new_contents":"import enum\nimport os\nimport subprocess\nimport sys\n\nimport blessings\n\nfrom .misc import isatty\n\n\nif isatty(sys.stdout) and os.getenv(\"TERM\"):\n Terminal = blessings.Terminal\nelse:\n # XXX: Mock terminal that returns \"\" for all attributes\n class TerminalValue:\n registry = {}\n\n @classmethod\n def get(cls, name):\n if name not in cls.registry:\n cls.registry[name] = cls(name)\n return cls.registry[name]\n\n def __init__(self, name):\n self.name = name\n\n def __repr__(self):\n return f\"{self.__class__.__name__}({self.name})\"\n\n def __str__(self):\n return \"\"\n\n class Terminal:\n def __getattr__(self, name):\n return TerminalValue.get(name)\n\n\nTERM = Terminal()\n\n\nclass Color(enum.Enum):\n\n none = \"\"\n reset = TERM.normal\n black = TERM.black\n red = TERM.red\n green = TERM.green\n yellow = TERM.yellow\n blue = TERM.blue\n magenta = TERM.magenta\n cyan = TERM.cyan\n white = TERM.white\n\n def __str__(self):\n return str(self.value)\n\n\nclass StreamOptions(enum.Enum):\n\n \"\"\"Choices for stream handling.\"\"\"\n\n capture = \"capture\"\n hide = \"hide\"\n none = \"none\"\n\n @property\n def option(self):\n return {\n \"capture\": subprocess.PIPE,\n \"hide\": subprocess.DEVNULL,\n \"none\": None,\n }[self.value]\n","subject":"Fix Color enum setup when TERM isn't set","message":"Fix Color enum setup when TERM isn't set\n\nThe previous version of this didn't work right because all the values\nwere the same empty string.\n\nThis works around that by creating distinct values that evaluate to \"\".\n\nAmends 94b55ead63523f7f5677989f1a4999994b205cdf\n","lang":"Python","license":"mit","repos":"wylee\/runcommands,wylee\/runcommands"} {"commit":"c4ee061f62e34c70cc67286ed0291423353cbcbe","old_file":"imgur_cli\/utils.py","new_file":"imgur_cli\/utils.py","old_contents":"import json\n\n\ndef cli_arg(*args, **kwargs):\n \"\"\"Decorator for CLI args\"\"\"\n def _decorator(func):\n add_arg(func, *args, **kwargs)\n return func\n return _decorator\n\n\ndef add_arg(func, *args, **kwargs):\n \"\"\"Bind CLI arguments a 'cmd_' format function\"\"\"\n if not hasattr(func, 'arguments'):\n func.arguments = []\n\n if (args, kwargs) not in func.arguments:\n # Because of the semantics of decorator composition if we just append\n # to the options list positional options will appear to be backwards.\n func.arguments.insert(0, (args, kwargs))\n\n\ndef generate_output(result, output_filename=None):\n if output_filename:\n with open(output_filename, 'w') as json_file:\n data = json.dumps(result, json_file, indent=4, separators=(',', ': '))\n json_file.write(data)\n else:\n print(json.dumps(result, indent=4, separators=(',', ': ')))\n","new_contents":"import json\n\n\ndef cli_arg(*args, **kwargs):\n \"\"\"Decorator for CLI args\"\"\"\n def _decorator(func):\n add_arg(func, *args, **kwargs)\n return func\n return _decorator\n\n\ndef add_arg(func, *args, **kwargs):\n \"\"\"Bind CLI arguments to a 'cmd_' format function\"\"\"\n if not hasattr(func, 'arguments'):\n func.arguments = []\n\n if (args, kwargs) not in func.arguments:\n # Because of the semantics of decorator composition if we just append\n # to the options list positional options will appear to be backwards.\n func.arguments.insert(0, (args, kwargs))\n\n\ndef cli_subparser(*args, **kwargs):\n \"\"\"Decorator for CLI subparsers\"\"\"\n def _decorator(func):\n add_subparser(func, *args, **kwargs)\n return func\n return _decorator\n\n\ndef add_subparser(func, *args, **kwargs):\n \"\"\"Bind CLI subparsers to a 'subparser_' format function\"\"\"\n if not hasattr(func, 'subparser'):\n func.subparser = args[0]\n\n\ndef generate_output(result, output_filename=None):\n if output_filename:\n with open(output_filename, 'w') as json_file:\n data = json.dumps(result, json_file, indent=4, separators=(',', ': '))\n json_file.write(data)\n else:\n print(json.dumps(result, indent=4, separators=(',', ': ')))\n","subject":"Define function and decorators for subparsers","message":"Define function and decorators for subparsers\n","lang":"Python","license":"mit","repos":"ueg1990\/imgur-cli"} {"commit":"5ff58311b6cf2dc8ad03351e818d05fca9e33e1b","old_file":"hastexo\/migrations\/0010_add_user_foreign_key.py","new_file":"hastexo\/migrations\/0010_add_user_foreign_key.py","old_contents":"from django.conf import settings\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.db import migrations, models\nimport django.db.models.deletion\n\nimport logging\nlogger = logging.getLogger(__name__)\n\n\nclass Migration(migrations.Migration):\n\n def backfill_learner(apps, schema_editor):\n \"\"\"\n Use the 'student_id' to link stacks to the User model.\n \"\"\"\n Stack = apps.get_model(\"hastexo\", \"Stack\")\n AnonymousUserId = apps.get_model(\"student\", \"AnonymousUserId\")\n for stack in Stack.objects.all():\n try:\n stack.learner = AnonymousUserId.objects.get(\n anonymous_user_id=stack.student_id).user\n stack.save(update_fields=['learner'])\n except ObjectDoesNotExist:\n logger.warning('Unable to link stack to user: '\n f'{stack.name}')\n\n dependencies = [\n migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n ('hastexo', '0009_add_null_true_for_key_and_password'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='stack',\n name='learner',\n field=models.ForeignKey(\n db_constraint=False,\n null=True,\n on_delete=django.db.models.deletion.PROTECT,\n to=settings.AUTH_USER_MODEL),\n ),\n migrations.RunPython(backfill_learner),\n ]\n","new_contents":"from django.conf import settings\nfrom django.core.exceptions import ObjectDoesNotExist\nfrom django.db import migrations, models\nimport django.db.models.deletion\n\nimport logging\nlogger = logging.getLogger(__name__)\n\n\nclass Migration(migrations.Migration):\n\n def backfill_learner(apps, schema_editor):\n \"\"\"\n Use the 'student_id' to link stacks to the User model.\n \"\"\"\n Stack = apps.get_model(\"hastexo\", \"Stack\")\n AnonymousUserId = apps.get_model(\"student\", \"AnonymousUserId\")\n for stack in Stack.objects.all():\n try:\n stack.learner = AnonymousUserId.objects.get(\n anonymous_user_id=stack.student_id).user\n stack.save(update_fields=['learner'])\n except ObjectDoesNotExist:\n logger.warning('Unable to link stack to user: '\n f'{stack.name}')\n\n dependencies = [\n migrations.swappable_dependency(settings.AUTH_USER_MODEL),\n ('hastexo', '0009_add_null_true_for_key_and_password'),\n ]\n\n operations = [\n migrations.AddField(\n model_name='stack',\n name='learner',\n field=models.ForeignKey(\n db_constraint=False,\n null=True,\n on_delete=django.db.models.deletion.PROTECT,\n to=settings.AUTH_USER_MODEL),\n ),\n migrations.RunPython(backfill_learner),\n migrations.AlterField(\n model_name='stack',\n name='learner',\n field=models.ForeignKey(\n db_constraint=True,\n null=True,\n on_delete=django.db.models.deletion.PROTECT,\n to=settings.AUTH_USER_MODEL),\n ),\n ]\n","subject":"Apply additional fix to add_user_foreign_key migration","message":"Apply additional fix to add_user_foreign_key migration\n\nThe hack in 583fb729b1e201c830579345dca5beca4b131006 modified\n0010_add_user_foreign_key in such a way that it ended up *not* setting\na database constraint when it should have.\n\nEnable the database-enforced constraint in the right place.\n\nCo-authored-by: Florian Haas <73262ad0334ab37227b2f7a0205f51db1e606681@cleura.com>\n","lang":"Python","license":"agpl-3.0","repos":"hastexo\/hastexo-xblock,hastexo\/hastexo-xblock,hastexo\/hastexo-xblock,hastexo\/hastexo-xblock"} {"commit":"ea73cd99b6ff67d65c0784471603d8734b6b3d75","old_file":"scripts\/plot_example.py","new_file":"scripts\/plot_example.py","old_contents":"import os\n\nimport matplotlib.pyplot as plt\nplt.style.use(\"ggplot\")\nplt.rcParams[\"figure.figsize\"] = 10, 5\nplt.rcParams[\"font.family\"] = \"serif\"\nplt.rcParams[\"font.size\"] = 12\n\nimport pyhector\nfrom pyhector import rcp26, rcp45, rcp60, rcp85\n\npath = os.path.join(os.path.dirname(__file__),\n '.\/example-plot.png')\n\nfor rcp in [rcp26, rcp45, rcp60, rcp85]:\n output, _ = pyhector.run(rcp, {\"core\": {\"endDate\": 2100}})\n temp = output[\"temperature.Tgav\"]\n temp = temp.loc[1850:] - temp.loc[1850:1900].mean()\n temp.plot(label=rcp.name.split(\"_\")[0])\nplt.title(\"Global mean temperature\")\nplt.ylabel(\"°C over pre-industrial (1850-1900 mean)\")\nplt.legend(loc=\"best\")\n\nplt.savefig(path, dpi=96)\n","new_contents":"import os\n\nimport matplotlib.pyplot as plt\nplt.style.use(\"ggplot\")\nplt.rcParams[\"figure.figsize\"] = 10, 5\nplt.rcParams[\"font.family\"] = \"serif\"\nplt.rcParams[\"font.size\"] = 12\n\nimport pyhector\nfrom pyhector import rcp26, rcp45, rcp60, rcp85\n\npath = os.path.join(os.path.dirname(__file__),\n '.\/example-plot.png')\n\nfor rcp in [rcp26, rcp45, rcp60, rcp85]:\n output = pyhector.run(rcp, {\"core\": {\"endDate\": 2100}})\n temp = output[\"temperature.Tgav\"]\n temp = temp.loc[1850:] - temp.loc[1850:1900].mean()\n temp.plot(label=rcp.name.split(\"_\")[0])\nplt.title(\"Global mean temperature\")\nplt.ylabel(\"°C over pre-industrial (1850-1900 mean)\")\nplt.legend(loc=\"best\")\n\nplt.savefig(path, dpi=96)\n","subject":"Update example plot script with new API","message":"Update example plot script with new API\n","lang":"Python","license":"agpl-3.0","repos":"openclimatedata\/pyhector,openclimatedata\/pyhector,openclimatedata\/pyhector"} {"commit":"7519bebe1d9d87930275858a537dcc0a0a64f007","old_file":"tools\/strip_filenames.py","new_file":"tools\/strip_filenames.py","old_contents":"#!\/bin\/python\nimport os\ndirectory = os.listdir()\nillegal_characters = \"%?_'*+$!\\\"\"\ntolowercase=True\n\nfor a in range(len(directory)):\n newname=\"\"\n for c in directory[a]:\n if c in illegal_characters:\n continue\n if c.isalnum() or c == '.':\n newname=newname+c.lower()\n print(\"convert {} to {}\".format(directory[a],newname))\n os.rename(directory[a], newname)\n\n","new_contents":"#!\/bin\/env python3\n\"\"\"\nUse only legal characters from files or current directory\nUsage:\n strip_filenames.py [<filename>...]\nOptions:\n -l, --lowercase Only lowercase\n -h, --help Show this screen and exit.\n\"\"\"\n\nimport sys\nimport os\nfrom docopt import docopt\n\n# docopt(doc, argv=None, help=True, version=None, options_first=False))\n\ndef main():\n opt = docopt(__doc__, sys.argv[1:])\n directory = opt.get(\"filename\", os.listdir())\n legal_characters = \"\"\n list_N010 = list(range(size))\n list_alpha = [ chr(x+97) for x in range(26) ]\n list_ALPHA = [ chr(x+65) for x in range(26) ]\n\n legal_characters += \"\".join(list_N010)\n legal_characters += \"\".join(list_alpha)\n if not opt.get(\"--lowercase\", False):\n legal_characters += \"\".join(list_N010)\n\n\n for a in range(len(directory)):\n newname=\"\"\n for c in directory[a]:\n if c not in legal_characters:\n continue\n newname += c\n print(\"convert {} to {}\".format(directory[a],newname))\n os.rename(directory[a], newname)\n\nif __name__ == \"__main__\":\n main()\n","subject":"Use legal characters for stripping filenames","message":"Use legal characters for stripping filenames\n","lang":"Python","license":"mit","repos":"dgengtek\/scripts,dgengtek\/scripts"} {"commit":"0fac3d59a34a861c7a826b0d1fa2f3002356e04c","old_file":"src\/shared.py","new_file":"src\/shared.py","old_contents":"# -*- coding: utf-8 -*-\nimport logging\nimport os\nimport queue\nimport threading\n\nlistening_port = 8444\nsend_outgoing_connections = True\nlisten_for_connections = True\ndata_directory = 'minode_data\/'\nsource_directory = os.path.dirname(os.path.realpath(__file__))\ntrusted_peer = None\n# trusted_peer = ('127.0.0.1', 8444)\n\nlog_level = logging.INFO\n\nmagic_bytes = b'\\xe9\\xbe\\xb4\\xd9'\nprotocol_version = 3\nservices = 3 # NODE_NETWORK, NODE_SSL\nstream = 1\nnonce = os.urandom(8)\nuser_agent = b'MiNode-v0.2.0'\ntimeout = 600\nheader_length = 24\n\nnonce_trials_per_byte = 1000\npayload_length_extra_bytes = 1000\n\nshutting_down = False\n\nvector_advertise_queue = queue.Queue()\naddress_advertise_queue = queue.Queue()\n\nconnections = set()\nconnections_lock = threading.Lock()\n\nhosts = set()\n\ncore_nodes = set()\n\nnode_pool = set()\nunchecked_node_pool = set()\n\noutgoing_connections = 8\nconnection_limit = 150\n\nobjects = {}\nobjects_lock = threading.Lock()\n","new_contents":"# -*- coding: utf-8 -*-\nimport logging\nimport os\nimport queue\nimport threading\n\nlistening_port = 8444\nsend_outgoing_connections = True\nlisten_for_connections = True\ndata_directory = 'minode_data\/'\nsource_directory = os.path.dirname(os.path.realpath(__file__))\ntrusted_peer = None\n# trusted_peer = ('127.0.0.1', 8444)\n\nlog_level = logging.INFO\n\nmagic_bytes = b'\\xe9\\xbe\\xb4\\xd9'\nprotocol_version = 3\nservices = 3 # NODE_NETWORK, NODE_SSL\nstream = 1\nnonce = os.urandom(8)\nuser_agent = b'\/MiNode:0.2.1\/'\ntimeout = 600\nheader_length = 24\n\nnonce_trials_per_byte = 1000\npayload_length_extra_bytes = 1000\n\nshutting_down = False\n\nvector_advertise_queue = queue.Queue()\naddress_advertise_queue = queue.Queue()\n\nconnections = set()\nconnections_lock = threading.Lock()\n\nhosts = set()\n\ncore_nodes = set()\n\nnode_pool = set()\nunchecked_node_pool = set()\n\noutgoing_connections = 8\nconnection_limit = 150\n\nobjects = {}\nobjects_lock = threading.Lock()\n","subject":"Change User Agent to comply with specification","message":"Change User Agent to comply with specification\n","lang":"Python","license":"mit","repos":"TheKysek\/MiNode,TheKysek\/MiNode"} {"commit":"c0b76d401b305c1bcd2ed5814a89719d4c6a3d83","old_file":"heat_cfnclient\/tests\/test_cli.py","new_file":"heat_cfnclient\/tests\/test_cli.py","old_contents":"# vim: tabstop=4 shiftwidth=4 softtabstop=4\n\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\nimport testtools\nimport heat_cfnclient\nimport os\nimport subprocess\n\nbasepath = os.path.join(heat_cfnclient.__path__[0], os.path.pardir)\n\n\nclass CliTest(testtools.TestCase):\n\n def test_heat_cfn(self):\n self.bin_run('heat-cfn')\n\n def test_heat_boto(self):\n self.bin_run('heat-boto')\n\n def test_heat_watch(self):\n self.bin_run('heat-watch')\n\n def bin_run(self, bin):\n fullpath = basepath + '\/bin\/' + bin\n\n proc = subprocess.Popen(fullpath,\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n stdout, stderr = proc.communicate()\n\n if proc.returncode:\n print('Error executing %s:\\n %s %s ' % (bin, stdout, stderr))\n raise subprocess.CalledProcessError(proc.returncode, bin)\n","new_contents":"# vim: tabstop=4 shiftwidth=4 softtabstop=4\n\n# Licensed under the Apache License, Version 2.0 (the \"License\"); you may\n# not use this file except in compliance with the License. You may obtain\n# a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS, WITHOUT\n# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the\n# License for the specific language governing permissions and limitations\n# under the License.\n\n\nimport testtools\nimport heat_cfnclient\nimport os\nimport subprocess\n\nbasepath = os.path.join(heat_cfnclient.__path__[0], os.path.pardir)\n\n\n@testtools.skip\nclass CliTest(testtools.TestCase):\n\n def test_heat_cfn(self):\n self.bin_run('heat-cfn')\n\n def test_heat_boto(self):\n self.bin_run('heat-boto')\n\n def test_heat_watch(self):\n self.bin_run('heat-watch')\n\n def bin_run(self, bin):\n fullpath = basepath + '\/bin\/' + bin\n\n proc = subprocess.Popen(fullpath,\n stdout=subprocess.PIPE,\n stderr=subprocess.PIPE)\n stdout, stderr = proc.communicate()\n\n if proc.returncode:\n print('Error executing %s:\\n %s %s ' % (bin, stdout, stderr))\n raise subprocess.CalledProcessError(proc.returncode, bin)\n","subject":"Disable tests until new repo is stable","message":"Disable tests until new repo is stable\n\nChange-Id: Ic6932c1028c72b5600d03ab59102d1c1cff1b36c\n","lang":"Python","license":"apache-2.0","repos":"openstack-dev\/heat-cfnclient"} {"commit":"82ad6bf164000940e17dcb01b27b22b97c69beba","old_file":"questionnaire\/urls.py","new_file":"questionnaire\/urls.py","old_contents":"# vim: set fileencoding=utf-8\n\nfrom django.conf.urls.defaults import *\nfrom views import *\n\nurlpatterns = patterns('',\n url(r'^$',\n questionnaire, name='questionnaire_noargs'),\n url(r'^csv\/(?P<qid>\\d+)\/',\n export_csv, name='export_csv'),\n url(r'^(?P<runcode>[^\/]+)\/(?P<qs>\\d+)\/$',\n questionnaire, name='questionset'),\n url(r'^(?P<runcode>[^\/]+)\/',\n questionnaire, name='questionnaire'),\n)\n","new_contents":"# vim: set fileencoding=utf-8\n\nfrom django.conf.urls.defaults import *\nfrom views import *\n\nurlpatterns = patterns('',\n url(r'^$',\n questionnaire, name='questionnaire_noargs'),\n url(r'^csv\/(?P<qid>\\d+)\/',\n export_csv, name='export_csv'),\n url(r'^(?P<runcode>[^\/]+)\/(?P<qs>[-]{0,1}\\d+)\/$',\n questionnaire, name='questionset'),\n url(r'^(?P<runcode>[^\/]+)\/',\n questionnaire, name='questionnaire'),\n)\n","subject":"Enable questionsets with negative sortids","message":"Enable questionsets with negative sortids\n","lang":"Python","license":"bsd-3-clause","repos":"JanOosting\/ed-questionnaire,affan2\/ed-questionnaire,seantis\/seantis-questionnaire,n3storm\/seantis-questionnaire,affan2\/ed-questionnaire,daniboy\/seantis-questionnaire,eugena\/ed-questionnaire,eugena\/seantis-questionnaire,JanOosting\/ed-questionnaire,eugena\/seantis-questionnaire,trantu\/seantis-questionnaire,daniboy\/seantis-questionnaire,JanOosting\/ed-questionnaire,n3storm\/seantis-questionnaire,seantis\/seantis-questionnaire,eugena\/ed-questionnaire,daniboy\/seantis-questionnaire,eugena\/ed-questionnaire,trantu\/seantis-questionnaire,eugena\/seantis-questionnaire,seantis\/seantis-questionnaire,trantu\/seantis-questionnaire,affan2\/ed-questionnaire"} {"commit":"c3bb58fbcbd7c1699571859af736952c36f3029a","old_file":"project\/library\/urls.py","new_file":"project\/library\/urls.py","old_contents":"from django.conf.urls.defaults import patterns, include, url\nfrom django.conf import settings\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\nurlpatterns = patterns('library',\n url(r'^all$',\n view='views.listing',\n kwargs={'template':'book_listing.html'},\n name='listing'\n ),\n url(r'^library\/(?P<id>[-\\w]+)\/$',\n view='views.book',\n kwargs={'template':'book.html'},\n name='book'\n )\n)\n","new_contents":"from django.conf.urls.defaults import patterns, include, url\nfrom django.conf import settings\nfrom django.contrib import admin\n\nadmin.autodiscover()\n\nurlpatterns = patterns('library',\n url(r'^all$',\n view='views.listing',\n kwargs={'template':'book_listing.html'},\n name='listing'\n ),\n url(r'^book\/(?P<id>[-\\w]+)\/$',\n view='views.book',\n kwargs={'template':'book.html'},\n name='book'\n )\n)\n","subject":"Update url for books to be more semantic","message":"Update url for books to be more semantic\n","lang":"Python","license":"mit","repos":"DUCSS\/ducss-site-old,DUCSS\/ducss-site-old,DUCSS\/ducss-site-old"} {"commit":"a15518111b6d03a4b67a2dbaa759afff15fe3302","old_file":"spec\/Report_S52_spec.py","new_file":"spec\/Report_S52_spec.py","old_contents":"from expects import expect, equal\nfrom primestg.report import Report\n\n\nwith fdescription('Report S52 example'):\n with before.all:\n\n self.data_filename = 'spec\/data\/MRTR000000822522_0_S52_1_20200929001048'\n\n self.report = {}\n with open(self.data_filename) as data_file:\n self.report = Report(data_file)\n\n with it('generates expected results for a value of the first line of '\n 'first remote terminal unit'):\n\n expected_first_value = dict(\n ae=0.0,\n bc='00',\n ai=8717.0,\n r1=43.0,\n r2=0.0,\n r3=0.0,\n r4=142.0,\n timestamp='2020-09-14 01:00:00',\n rt_unit_name='MRTR000000822522',\n name='MRTL000006609121',\n magn=1\n )\n\n rt_unit = list(self.report.rt_units)[0]\n line = list(rt_unit.lines)[0]\n values = line.values\n\n first_value_first_line = {}\n for x in values:\n if x['timestamp'] == expected_first_value['timestamp']:\n first_value_first_line = x\n\n expect(first_value_first_line)\\\n .to(equal(expected_first_value))\n","new_contents":"from expects import expect, equal\nfrom primestg.report import Report\n\n\nwith description('Report S52 example'):\n with before.all:\n\n self.data_filename = 'spec\/data\/MRTR000000822522_0_S52_1_20200929001048'\n\n self.report = {}\n with open(self.data_filename) as data_file:\n self.report = Report(data_file)\n\n with it('generates expected results for a value of the first line of '\n 'first remote terminal unit'):\n\n expected_first_value = dict(\n ae=0.0,\n bc='00',\n ai=8717.0,\n r1=43.0,\n r2=0.0,\n r3=0.0,\n r4=142.0,\n timestamp='2020-09-14 01:00:00',\n rt_unit_name='MRTR000000822522',\n name='MRTL000006609121',\n magn=1\n )\n\n rt_unit = list(self.report.rt_units)[0]\n line = list(rt_unit.lines)[0]\n values = line.values\n\n first_value_first_line = {}\n for x in values:\n if x['timestamp'] == expected_first_value['timestamp']:\n first_value_first_line = x\n\n expect(first_value_first_line)\\\n .to(equal(expected_first_value))\n","subject":"FIX only pass S52 test","message":"FIX only pass S52 test\n","lang":"Python","license":"agpl-3.0","repos":"gisce\/primestg"} {"commit":"8d5d45f3a04235a9ee4fd1cadd39cc0010775ac9","old_file":"humbug\/ratelimit.py","new_file":"humbug\/ratelimit.py","old_contents":"import traceback\nfrom hashlib import sha256\nfrom datetime import datetime, timedelta\n\n# Adapted http:\/\/djangosnippets.org\/snippets\/2242\/ by user s29 (October 25, 2010)\n\nclass _RateLimitFilter(object):\n last_error = 0\n\n def filter(self, record):\n from django.conf import settings\n from django.core.cache import cache\n\n # Track duplicate errors\n duplicate = False\n rate = getattr(settings, '%s_LIMIT' % self.__class__.__name__.upper(),\n 600) # seconds\n if rate > 0:\n # Test if the cache works\n try:\n cache.set('RLF_TEST_KEY', 1, 1)\n use_cache = cache.get('RLF_TEST_KEY') == 1\n except:\n use_cache = False\n\n if use_cache:\n key = self.__class__.__name__.upper()\n duplicate = cache.get(key) == 1\n cache.set(key, 1, rate)\n else:\n min_date = datetime.now() - timedelta(seconds=rate)\n duplicate = (self.last_error >= min_date)\n if not duplicate:\n self.last_error = datetime.now()\n\n return not duplicate\n\nclass HumbugLimiter(_RateLimitFilter):\n pass\n\nclass EmailLimiter(_RateLimitFilter):\n pass\n","new_contents":"import traceback\nfrom hashlib import sha256\nfrom datetime import datetime, timedelta\n\n# Adapted http:\/\/djangosnippets.org\/snippets\/2242\/ by user s29 (October 25, 2010)\n\nclass _RateLimitFilter(object):\n last_error = datetime.min\n\n def filter(self, record):\n from django.conf import settings\n from django.core.cache import cache\n\n # Track duplicate errors\n duplicate = False\n rate = getattr(settings, '%s_LIMIT' % self.__class__.__name__.upper(),\n 600) # seconds\n if rate > 0:\n # Test if the cache works\n try:\n cache.set('RLF_TEST_KEY', 1, 1)\n use_cache = cache.get('RLF_TEST_KEY') == 1\n except:\n use_cache = False\n\n if use_cache:\n key = self.__class__.__name__.upper()\n duplicate = cache.get(key) == 1\n cache.set(key, 1, rate)\n else:\n min_date = datetime.now() - timedelta(seconds=rate)\n duplicate = (self.last_error >= min_date)\n if not duplicate:\n self.last_error = datetime.now()\n\n return not duplicate\n\nclass HumbugLimiter(_RateLimitFilter):\n pass\n\nclass EmailLimiter(_RateLimitFilter):\n pass\n","subject":"Use datetime.min for initial last_error rather than int 0.","message":"Use datetime.min for initial last_error rather than int 0.\n\nOtherwise, code may break when it encounters a comparison against\nlast_error.\n\n(imported from commit 301f256fba065ae9704b1d7f6e91e69ec54f1aa1)\n","lang":"Python","license":"apache-2.0","repos":"levixie\/zulip,zwily\/zulip,reyha\/zulip,rht\/zulip,jrowan\/zulip,praveenaki\/zulip,esander91\/zulip,jeffcao\/zulip,Juanvulcano\/zulip,zachallaun\/zulip,Batterfii\/zulip,KingxBanana\/zulip,krtkmj\/zulip,zorojean\/zulip,christi3k\/zulip,easyfmxu\/zulip,arpitpanwar\/zulip,glovebx\/zulip,yuvipanda\/zulip,ashwinirudrappa\/zulip,johnnygaddarr\/zulip,jackrzhang\/zulip,yuvipanda\/zulip,jerryge\/zulip,bastianh\/zulip,zachallaun\/zulip,hengqujushi\/zulip,JPJPJPOPOP\/zulip,dattatreya303\/zulip,babbage\/zulip,JanzTam\/zulip,fw1121\/zulip,sup95\/zulip,udxxabp\/zulip,so0k\/zulip,JanzTam\/zulip,zhaoweigg\/zulip,technicalpickles\/zulip,moria\/zulip,eeshangarg\/zulip,moria\/zulip,wangdeshui\/zulip,hackerkid\/zulip,bssrdf\/zulip,ryanbackman\/zulip,bitemyapp\/zulip,Cheppers\/zulip,gkotian\/zulip,punchagan\/zulip,tdr130\/zulip,joshisa\/zulip,synicalsyntax\/zulip,j831\/zulip,RobotCaleb\/zulip,amanharitsh123\/zulip,dawran6\/zulip,lfranchi\/zulip,amyliu345\/zulip,Cheppers\/zulip,aakash-cr7\/zulip,tbutter\/zulip,dattatreya303\/zulip,jonesgithub\/zulip,LAndreas\/zulip,thomasboyt\/zulip,MayB\/zulip,kokoar\/zulip,jerryge\/zulip,wangdeshui\/zulip,qq1012803704\/zulip,themass\/zulip,Vallher\/zulip,calvinleenyc\/zulip,zwily\/zulip,vabs22\/zulip,j831\/zulip,hj3938\/zulip,ufosky-server\/zulip,gigawhitlocks\/zulip,dnmfarrell\/zulip,SmartPeople\/zulip,itnihao\/zulip,noroot\/zulip,ikasumiwt\/zulip,codeKonami\/zulip,zorojean\/zulip,adnanh\/zulip,jeffcao\/zulip,brainwane\/zulip,sup95\/zulip,ericzhou2008\/zulip,dwrpayne\/zulip,zacps\/zulip,johnnygaddarr\/zulip,shaunstanislaus\/zulip,schatt\/zulip,swinghu\/zulip,reyha\/zulip,tiansiyuan\/zulip,Batterfii\/zulip,so0k\/zulip,brockwhittaker\/zulip,firstblade\/zulip,proliming\/zulip,themass\/zulip,vakila\/zulip,calvinleenyc\/zulip,developerfm\/zulip,codeKonami\/zulip,shubhamdhama\/zulip,mahim97\/zulip,hafeez3000\/zulip,itnihao\/zulip,adnanh\/zulip,voidException\/zulip,AZtheAsian\/zulip,guiquanz\/zulip,umkay\/zulip,susansls\/zulip,fw1121\/zulip,ericzhou2008\/zulip,dwrpayne\/zulip,timabbott\/zulip,dhcrzf\/zulip,he15his\/zulip,fw1121\/zulip,timabbott\/zulip,KJin99\/zulip,adnanh\/zulip,Galexrt\/zulip,verma-varsha\/zulip,wangdeshui\/zulip,kaiyuanheshang\/zulip,wdaher\/zulip,luyifan\/zulip,dxq-git\/zulip,natanovia\/zulip,glovebx\/zulip,armooo\/zulip,PaulPetring\/zulip,easyfmxu\/zulip,MayB\/zulip,joshisa\/zulip,yocome\/zulip,he15his\/zulip,hackerkid\/zulip,natanovia\/zulip,adnanh\/zulip,Diptanshu8\/zulip,xuxiao\/zulip,jackrzhang\/zulip,johnnygaddarr\/zulip,brainwane\/zulip,souravbadami\/zulip,bssrdf\/zulip,Galexrt\/zulip,aakash-cr7\/zulip,dotcool\/zulip,easyfmxu\/zulip,babbage\/zulip,KJin99\/zulip,fw1121\/zulip,Batterfii\/zulip,eastlhu\/zulip,pradiptad\/zulip,KingxBanana\/zulip,xuanhan863\/zulip,umkay\/zulip,AZtheAsian\/zulip,tdr130\/zulip,suxinde2009\/zulip,zorojean\/zulip,zacps\/zulip,zhaoweigg\/zulip,paxapy\/zulip,andersk\/zulip,seapasulli\/zulip,voidException\/zulip,jackrzhang\/zulip,bowlofstew\/zulip,paxapy\/zulip,Suninus\/zulip,showell\/zulip,udxxabp\/zulip,Juanvulcano\/zulip,luyifan\/zulip,voidException\/zulip,hafeez3000\/zulip,arpitpanwar\/zulip,hafeez3000\/zulip,ufosky-server\/zulip,MariaFaBella85\/zulip,babbage\/zulip,codeKonami\/zulip,firstblade\/zulip,ryanbackman\/zulip,eeshangarg\/zulip,dwrpayne\/zulip,christi3k\/zulip,bluesea\/zulip,pradiptad\/zulip,niftynei\/zulip,ikasumiwt\/zulip,wavelets\/zulip,JanzTam\/zulip,zwily\/zulip,he15his\/zulip,praveenaki\/zulip,alliejones\/zulip,peiwei\/zulip,amallia\/zulip,verma-varsha\/zulip,DazWorrall\/zulip,jerryge\/zulip,calvinleenyc\/zulip,ericzhou2008\/zulip,codeKonami\/zulip,luyifan\/zulip,kou\/zulip,arpitpanwar\/zulip,andersk\/zulip,KingxBanana\/zulip,nicholasbs\/zulip,technicalpickles\/zulip,xuxiao\/zulip,nicholasbs\/zulip,jonesgithub\/zulip,zhaoweigg\/zulip,akuseru\/zulip,andersk\/zulip,littledogboy\/zulip,eeshangarg\/zulip,bitemyapp\/zulip,xuxiao\/zulip,esander91\/zulip,ipernet\/zulip,shaunstanislaus\/zulip,Drooids\/zulip,JanzTam\/zulip,Frouk\/zulip,mdavid\/zulip,tiansiyuan\/zulip,jimmy54\/zulip,luyifan\/zulip,jerryge\/zulip,babbage\/zulip,bastianh\/zulip,hj3938\/zulip,proliming\/zulip,glovebx\/zulip,karamcnair\/zulip,PhilSk\/zulip,xuanhan863\/zulip,vaidap\/zulip,Diptanshu8\/zulip,sonali0901\/zulip,synicalsyntax\/zulip,tbutter\/zulip,mohsenSy\/zulip,LeeRisk\/zulip,dnmfarrell\/zulip,itnihao\/zulip,souravbadami\/zulip,hengqujushi\/zulip,ipernet\/zulip,LeeRisk\/zulip,schatt\/zulip,jphilipsen05\/zulip,Batterfii\/zulip,seapasulli\/zulip,showell\/zulip,jerryge\/zulip,ufosky-server\/zulip,bssrdf\/zulip,umkay\/zulip,proliming\/zulip,moria\/zulip,bastianh\/zulip,Jianchun1\/zulip,sharmaeklavya2\/zulip,ApsOps\/zulip,voidException\/zulip,timabbott\/zulip,saitodisse\/zulip,AZtheAsian\/zulip,saitodisse\/zulip,ashwinirudrappa\/zulip,tbutter\/zulip,isht3\/zulip,Diptanshu8\/zulip,Juanvulcano\/zulip,luyifan\/zulip,wavelets\/zulip,zulip\/zulip,kou\/zulip,vakila\/zulip,joyhchen\/zulip,developerfm\/zulip,vabs22\/zulip,j831\/zulip,cosmicAsymmetry\/zulip,itnihao\/zulip,akuseru\/zulip,jrowan\/zulip,johnny9\/zulip,blaze225\/zulip,bowlofstew\/zulip,shrikrishnaholla\/zulip,amanharitsh123\/zulip,amallia\/zulip,avastu\/zulip,lfranchi\/zulip,deer-hope\/zulip,hackerkid\/zulip,zorojean\/zulip,yocome\/zulip,vabs22\/zulip,JPJPJPOPOP\/zulip,andersk\/zulip,themass\/zulip,krtkmj\/zulip,mansilladev\/zulip,jrowan\/zulip,deer-hope\/zulip,susansls\/zulip,zacps\/zulip,arpith\/zulip,mohsenSy\/zulip,vikas-parashar\/zulip,gigawhitlocks\/zulip,AZtheAsian\/zulip,nicholasbs\/zulip,hengqujushi\/zulip,ahmadassaf\/zulip,pradiptad\/zulip,Gabriel0402\/zulip,reyha\/zulip,ryansnowboarder\/zulip,codeKonami\/zulip,ashwinirudrappa\/zulip,zorojean\/zulip,natanovia\/zulip,jeffcao\/zulip,jphilipsen05\/zulip,Gabriel0402\/zulip,yocome\/zulip,karamcnair\/zulip,Vallher\/zulip,Gabriel0402\/zulip,easyfmxu\/zulip,dawran6\/zulip,ikasumiwt\/zulip,shrikrishnaholla\/zulip,ericzhou2008\/zulip,gkotian\/zulip,akuseru\/zulip,voidException\/zulip,suxinde2009\/zulip,Juanvulcano\/zulip,mahim97\/zulip,littledogboy\/zulip,Gabriel0402\/zulip,shaunstanislaus\/zulip,schatt\/zulip,saitodisse\/zulip,xuanhan863\/zulip,zacps\/zulip,PhilSk\/zulip,ahmadassaf\/zulip,developerfm\/zulip,niftynei\/zulip,he15his\/zulip,peguin40\/zulip,samatdav\/zulip,esander91\/zulip,willingc\/zulip,rht\/zulip,natanovia\/zulip,Batterfii\/zulip,mahim97\/zulip,Qgap\/zulip,schatt\/zulip,mdavid\/zulip,shubhamdhama\/zulip,dnmfarrell\/zulip,ericzhou2008\/zulip,zwily\/zulip,swinghu\/zulip,arpith\/zulip,willingc\/zulip,christi3k\/zulip,shaunstanislaus\/zulip,qq1012803704\/zulip,DazWorrall\/zulip,he15his\/zulip,natanovia\/zulip,zwily\/zulip,hj3938\/zulip,yuvipanda\/zulip,Galexrt\/zulip,arpith\/zulip,sharmaeklavya2\/zulip,PaulPetring\/zulip,tdr130\/zulip,moria\/zulip,wangdeshui\/zulip,amallia\/zulip,Drooids\/zulip,dnmfarrell\/zulip,pradiptad\/zulip,MariaFaBella85\/zulip,hayderimran7\/zulip,jrowan\/zulip,blaze225\/zulip,grave-w-grave\/zulip,timabbott\/zulip,jerryge\/zulip,paxapy\/zulip,Frouk\/zulip,joshisa\/zulip,Batterfii\/zulip,johnny9\/zulip,zwily\/zulip,pradiptad\/zulip,dxq-git\/zulip,thomasboyt\/zulip,wweiradio\/zulip,wweiradio\/zulip,qq1012803704\/zulip,amanharitsh123\/zulip,shubhamdhama\/zulip,hafeez3000\/zulip,udxxabp\/zulip,xuanhan863\/zulip,jessedhillon\/zulip,hackerkid\/zulip,mohsenSy\/zulip,huangkebo\/zulip,udxxabp\/zulip,hustlzp\/zulip,kokoar\/zulip,souravbadami\/zulip,moria\/zulip,developerfm\/zulip,MariaFaBella85\/zulip,nicholasbs\/zulip,Galexrt\/zulip,niftynei\/zulip,tdr130\/zulip,bastianh\/zulip,developerfm\/zulip,vakila\/zulip,jainayush975\/zulip,ApsOps\/zulip,SmartPeople\/zulip,brainwane\/zulip,sup95\/zulip,thomasboyt\/zulip,hafeez3000\/zulip,seapasulli\/zulip,brainwane\/zulip,Frouk\/zulip,shubhamdhama\/zulip,dxq-git\/zulip,amyliu345\/zulip,deer-hope\/zulip,reyha\/zulip,susansls\/zulip,johnny9\/zulip,levixie\/zulip,isht3\/zulip,atomic-labs\/zulip,glovebx\/zulip,wweiradio\/zulip,Drooids\/zulip,aliceriot\/zulip,dnmfarrell\/zulip,eastlhu\/zulip,kaiyuanheshang\/zulip,developerfm\/zulip,gigawhitlocks\/zulip,brockwhittaker\/zulip,hackerkid\/zulip,Suninus\/zulip,punchagan\/zulip,levixie\/zulip,armooo\/zulip,he15his\/zulip,hayderimran7\/zulip,hustlzp\/zulip,yocome\/zulip,jphilipsen05\/zulip,JPJPJPOPOP\/zulip,SmartPeople\/zulip,johnnygaddarr\/zulip,ryanbackman\/zulip,lfranchi\/zulip,avastu\/zulip,pradiptad\/zulip,verma-varsha\/zulip,hayderimran7\/zulip,karamcnair\/zulip,vakila\/zulip,sonali0901\/zulip,qq1012803704\/zulip,stamhe\/zulip,bitemyapp\/zulip,zacps\/zulip,gkotian\/zulip,Vallher\/zulip,hafeez3000\/zulip,willingc\/zulip,adnanh\/zulip,atomic-labs\/zulip,EasonYi\/zulip,thomasboyt\/zulip,dwrpayne\/zulip,itnihao\/zulip,kokoar\/zulip,Frouk\/zulip,jimmy54\/zulip,jimmy54\/zulip,amanharitsh123\/zulip,easyfmxu\/zulip,zofuthan\/zulip,zofuthan\/zulip,Cheppers\/zulip,mohsenSy\/zulip,noroot\/zulip,hayderimran7\/zulip,yuvipanda\/zulip,amallia\/zulip,vikas-parashar\/zulip,mansilladev\/zulip,aps-sids\/zulip,KJin99\/zulip,Gabriel0402\/zulip,littledogboy\/zulip,showell\/zulip,zachallaun\/zulip,aakash-cr7\/zulip,peiwei\/zulip,m1ssou\/zulip,TigorC\/zulip,hustlzp\/zulip,LeeRisk\/zulip,shaunstanislaus\/zulip,vaidap\/zulip,aakash-cr7\/zulip,aliceriot\/zulip,cosmicAsymmetry\/zulip,joshisa\/zulip,ipernet\/zulip,LeeRisk\/zulip,hj3938\/zulip,m1ssou\/zulip,sharmaeklavya2\/zulip,LAndreas\/zulip,dattatreya303\/zulip,SmartPeople\/zulip,rht\/zulip,tdr130\/zulip,praveenaki\/zulip,blaze225\/zulip,karamcnair\/zulip,zulip\/zulip,jainayush975\/zulip,Jianchun1\/zulip,Diptanshu8\/zulip,zorojean\/zulip,tommyip\/zulip,bssrdf\/zulip,eastlhu\/zulip,esander91\/zulip,deer-hope\/zulip,zhaoweigg\/zulip,TigorC\/zulip,themass\/zulip,Juanvulcano\/zulip,aliceriot\/zulip,JPJPJPOPOP\/zulip,ahmadassaf\/zulip,eastlhu\/zulip,jimmy54\/zulip,dotcool\/zulip,kaiyuanheshang\/zulip,Jianchun1\/zulip,susansls\/zulip,bluesea\/zulip,swinghu\/zulip,Vallher\/zulip,shaunstanislaus\/zulip,hengqujushi\/zulip,christi3k\/zulip,peiwei\/zulip,littledogboy\/zulip,vikas-parashar\/zulip,Suninus\/zulip,wdaher\/zulip,amanharitsh123\/zulip,fw1121\/zulip,hustlzp\/zulip,adnanh\/zulip,ahmadassaf\/zulip,wavelets\/zulip,tommyip\/zulip,johnny9\/zulip,xuxiao\/zulip,Diptanshu8\/zulip,avastu\/zulip,jessedhillon\/zulip,m1ssou\/zulip,punchagan\/zulip,rht\/zulip,wangdeshui\/zulip,timabbott\/zulip,aliceriot\/zulip,jphilipsen05\/zulip,sup95\/zulip,noroot\/zulip,ryansnowboarder\/zulip,Jianchun1\/zulip,synicalsyntax\/zulip,willingc\/zulip,wdaher\/zulip,nicholasbs\/zulip,guiquanz\/zulip,Vallher\/zulip,kou\/zulip,grave-w-grave\/zulip,avastu\/zulip,souravbadami\/zulip,dxq-git\/zulip,Batterfii\/zulip,noroot\/zulip,m1ssou\/zulip,saitodisse\/zulip,voidException\/zulip,vaidap\/zulip,noroot\/zulip,aakash-cr7\/zulip,Drooids\/zulip,jainayush975\/zulip,bluesea\/zulip,jessedhillon\/zulip,RobotCaleb\/zulip,jackrzhang\/zulip,xuanhan863\/zulip,gkotian\/zulip,DazWorrall\/zulip,armooo\/zulip,easyfmxu\/zulip,peguin40\/zulip,praveenaki\/zulip,LAndreas\/zulip,suxinde2009\/zulip,huangkebo\/zulip,wavelets\/zulip,Jianchun1\/zulip,akuseru\/zulip,ryansnowboarder\/zulip,hackerkid\/zulip,tiansiyuan\/zulip,PaulPetring\/zulip,jonesgithub\/zulip,he15his\/zulip,MariaFaBella85\/zulip,dotcool\/zulip,hackerkid\/zulip,gigawhitlocks\/zulip,bowlofstew\/zulip,KJin99\/zulip,ikasumiwt\/zulip,zulip\/zulip,vaidap\/zulip,grave-w-grave\/zulip,PhilSk\/zulip,amyliu345\/zulip,ikasumiwt\/zulip,bssrdf\/zulip,thomasboyt\/zulip,arpith\/zulip,praveenaki\/zulip,bluesea\/zulip,babbage\/zulip,bluesea\/zulip,guiquanz\/zulip,vakila\/zulip,ApsOps\/zulip,amyliu345\/zulip,thomasboyt\/zulip,PaulPetring\/zulip,ryansnowboarder\/zulip,dhcrzf\/zulip,hafeez3000\/zulip,Suninus\/zulip,eeshangarg\/zulip,MariaFaBella85\/zulip,jeffcao\/zulip,Cheppers\/zulip,alliejones\/zulip,jeffcao\/zulip,avastu\/zulip,suxinde2009\/zulip,so0k\/zulip,aliceriot\/zulip,dhcrzf\/zulip,arpitpanwar\/zulip,udxxabp\/zulip,TigorC\/zulip,niftynei\/zulip,lfranchi\/zulip,arpith\/zulip,tommyip\/zulip,KingxBanana\/zulip,hj3938\/zulip,tommyip\/zulip,cosmicAsymmetry\/zulip,bssrdf\/zulip,bastianh\/zulip,xuxiao\/zulip,sharmaeklavya2\/zulip,PhilSk\/zulip,Cheppers\/zulip,eastlhu\/zulip,jphilipsen05\/zulip,gkotian\/zulip,akuseru\/zulip,LAndreas\/zulip,jimmy54\/zulip,EasonYi\/zulip,rishig\/zulip,umkay\/zulip,deer-hope\/zulip,ApsOps\/zulip,suxinde2009\/zulip,ipernet\/zulip,Gabriel0402\/zulip,bitemyapp\/zulip,aps-sids\/zulip,tdr130\/zulip,easyfmxu\/zulip,eeshangarg\/zulip,cosmicAsymmetry\/zulip,gigawhitlocks\/zulip,dhcrzf\/zulip,kou\/zulip,jimmy54\/zulip,dotcool\/zulip,aliceriot\/zulip,dxq-git\/zulip,wangdeshui\/zulip,peiwei\/zulip,sup95\/zulip,brockwhittaker\/zulip,shubhamdhama\/zulip,isht3\/zulip,glovebx\/zulip,vaidap\/zulip,kou\/zulip,zulip\/zulip,zhaoweigg\/zulip,tbutter\/zulip,tiansiyuan\/zulip,xuanhan863\/zulip,grave-w-grave\/zulip,TigorC\/zulip,zhaoweigg\/zulip,bluesea\/zulip,ryanbackman\/zulip,aps-sids\/zulip,rishig\/zulip,kokoar\/zulip,dawran6\/zulip,EasonYi\/zulip,dattatreya303\/zulip,Qgap\/zulip,j831\/zulip,bitemyapp\/zulip,RobotCaleb\/zulip,Qgap\/zulip,jphilipsen05\/zulip,proliming\/zulip,xuxiao\/zulip,Cheppers\/zulip,ryansnowboarder\/zulip,rishig\/zulip,willingc\/zulip,bastianh\/zulip,reyha\/zulip,MayB\/zulip,ashwinirudrappa\/zulip,vikas-parashar\/zulip,reyha\/zulip,DazWorrall\/zulip,peguin40\/zulip,zorojean\/zulip,jessedhillon\/zulip,ApsOps\/zulip,KJin99\/zulip,developerfm\/zulip,kou\/zulip,KingxBanana\/zulip,gigawhitlocks\/zulip,yuvipanda\/zulip,arpitpanwar\/zulip,firstblade\/zulip,isht3\/zulip,DazWorrall\/zulip,hustlzp\/zulip,proliming\/zulip,proliming\/zulip,Frouk\/zulip,babbage\/zulip,karamcnair\/zulip,fw1121\/zulip,lfranchi\/zulip,mdavid\/zulip,ryanbackman\/zulip,JanzTam\/zulip,mansilladev\/zulip,zulip\/zulip,ashwinirudrappa\/zulip,m1ssou\/zulip,susansls\/zulip,dnmfarrell\/zulip,zhaoweigg\/zulip,andersk\/zulip,gigawhitlocks\/zulip,KJin99\/zulip,peiwei\/zulip,armooo\/zulip,voidException\/zulip,hayderimran7\/zulip,wavelets\/zulip,vabs22\/zulip,peguin40\/zulip,dhcrzf\/zulip,bowlofstew\/zulip,joshisa\/zulip,dawran6\/zulip,EasonYi\/zulip,levixie\/zulip,ericzhou2008\/zulip,shubhamdhama\/zulip,verma-varsha\/zulip,LeeRisk\/zulip,calvinleenyc\/zulip,wweiradio\/zulip,peguin40\/zulip,bastianh\/zulip,themass\/zulip,ipernet\/zulip,glovebx\/zulip,bowlofstew\/zulip,DazWorrall\/zulip,saitodisse\/zulip,zacps\/zulip,ufosky-server\/zulip,krtkmj\/zulip,samatdav\/zulip,rht\/zulip,kaiyuanheshang\/zulip,EasonYi\/zulip,dwrpayne\/zulip,tdr130\/zulip,aps-sids\/zulip,tbutter\/zulip,joyhchen\/zulip,punchagan\/zulip,Suninus\/zulip,tbutter\/zulip,synicalsyntax\/zulip,bssrdf\/zulip,brainwane\/zulip,swinghu\/zulip,wdaher\/zulip,so0k\/zulip,guiquanz\/zulip,mahim97\/zulip,dattatreya303\/zulip,ahmadassaf\/zulip,ashwinirudrappa\/zulip,arpith\/zulip,RobotCaleb\/zulip,zofuthan\/zulip,atomic-labs\/zulip,luyifan\/zulip,dotcool\/zulip,suxinde2009\/zulip,paxapy\/zulip,krtkmj\/zulip,jainayush975\/zulip,RobotCaleb\/zulip,deer-hope\/zulip,Gabriel0402\/zulip,PhilSk\/zulip,vabs22\/zulip,levixie\/zulip,sonali0901\/zulip,dotcool\/zulip,j831\/zulip,schatt\/zulip,dhcrzf\/zulip,zachallaun\/zulip,christi3k\/zulip,yuvipanda\/zulip,shubhamdhama\/zulip,JPJPJPOPOP\/zulip,Drooids\/zulip,schatt\/zulip,saitodisse\/zulip,shrikrishnaholla\/zulip,PaulPetring\/zulip,wdaher\/zulip,EasonYi\/zulip,ipernet\/zulip,technicalpickles\/zulip,shrikrishnaholla\/zulip,yocome\/zulip,firstblade\/zulip,rishig\/zulip,ApsOps\/zulip,joyhchen\/zulip,paxapy\/zulip,armooo\/zulip,ashwinirudrappa\/zulip,tommyip\/zulip,ikasumiwt\/zulip,atomic-labs\/zulip,blaze225\/zulip,esander91\/zulip,Suninus\/zulip,synicalsyntax\/zulip,amanharitsh123\/zulip,samatdav\/zulip,adnanh\/zulip,amallia\/zulip,stamhe\/zulip,qq1012803704\/zulip,hengqujushi\/zulip,qq1012803704\/zulip,mohsenSy\/zulip,joshisa\/zulip,samatdav\/zulip,PaulPetring\/zulip,ApsOps\/zulip,kaiyuanheshang\/zulip,peguin40\/zulip,amyliu345\/zulip,Suninus\/zulip,jrowan\/zulip,hustlzp\/zulip,codeKonami\/zulip,armooo\/zulip,hengqujushi\/zulip,brockwhittaker\/zulip,vabs22\/zulip,joyhchen\/zulip,swinghu\/zulip,bitemyapp\/zulip,technicalpickles\/zulip,glovebx\/zulip,AZtheAsian\/zulip,johnny9\/zulip,joshisa\/zulip,bitemyapp\/zulip,krtkmj\/zulip,JanzTam\/zulip,jackrzhang\/zulip,vikas-parashar\/zulip,calvinleenyc\/zulip,johnnygaddarr\/zulip,vakila\/zulip,hustlzp\/zulip,wweiradio\/zulip,codeKonami\/zulip,zachallaun\/zulip,willingc\/zulip,babbage\/zulip,KingxBanana\/zulip,johnnygaddarr\/zulip,showell\/zulip,schatt\/zulip,dwrpayne\/zulip,MayB\/zulip,tiansiyuan\/zulip,littledogboy\/zulip,mdavid\/zulip,andersk\/zulip,shrikrishnaholla\/zulip,jimmy54\/zulip,seapasulli\/zulip,JanzTam\/zulip,LAndreas\/zulip,levixie\/zulip,yocome\/zulip,wangdeshui\/zulip,shrikrishnaholla\/zulip,huangkebo\/zulip,cosmicAsymmetry\/zulip,grave-w-grave\/zulip,LeeRisk\/zulip,grave-w-grave\/zulip,amyliu345\/zulip,stamhe\/zulip,yuvipanda\/zulip,jeffcao\/zulip,wdaher\/zulip,shrikrishnaholla\/zulip,brockwhittaker\/zulip,jonesgithub\/zulip,zachallaun\/zulip,sup95\/zulip,avastu\/zulip,tbutter\/zulip,vakila\/zulip,jonesgithub\/zulip,AZtheAsian\/zulip,itnihao\/zulip,krtkmj\/zulip,moria\/zulip,Galexrt\/zulip,verma-varsha\/zulip,Diptanshu8\/zulip,luyifan\/zulip,m1ssou\/zulip,paxapy\/zulip,dxq-git\/zulip,yocome\/zulip,jackrzhang\/zulip,kou\/zulip,aps-sids\/zulip,j831\/zulip,saitodisse\/zulip,ahmadassaf\/zulip,aliceriot\/zulip,huangkebo\/zulip,LAndreas\/zulip,udxxabp\/zulip,nicholasbs\/zulip,sharmaeklavya2\/zulip,xuxiao\/zulip,umkay\/zulip,mdavid\/zulip,lfranchi\/zulip,brockwhittaker\/zulip,SmartPeople\/zulip,Jianchun1\/zulip,sharmaeklavya2\/zulip,rishig\/zulip,wweiradio\/zulip,Juanvulcano\/zulip,jainayush975\/zulip,tommyip\/zulip,JPJPJPOPOP\/zulip,firstblade\/zulip,pradiptad\/zulip,alliejones\/zulip,souravbadami\/zulip,praveenaki\/zulip,natanovia\/zulip,fw1121\/zulip,RobotCaleb\/zulip,atomic-labs\/zulip,huangkebo\/zulip,sonali0901\/zulip,rht\/zulip,umkay\/zulip,dhcrzf\/zulip,alliejones\/zulip,wweiradio\/zulip,vaidap\/zulip,ipernet\/zulip,levixie\/zulip,jessedhillon\/zulip,eeshangarg\/zulip,Vallher\/zulip,cosmicAsymmetry\/zulip,tiansiyuan\/zulip,suxinde2009\/zulip,dawran6\/zulip,alliejones\/zulip,esander91\/zulip,sonali0901\/zulip,johnny9\/zulip,bowlofstew\/zulip,dwrpayne\/zulip,ufosky-server\/zulip,akuseru\/zulip,MariaFaBella85\/zulip,Vallher\/zulip,zofuthan\/zulip,showell\/zulip,tiansiyuan\/zulip,aps-sids\/zulip,arpitpanwar\/zulip,qq1012803704\/zulip,ufosky-server\/zulip,verma-varsha\/zulip,so0k\/zulip,KJin99\/zulip,littledogboy\/zulip,ikasumiwt\/zulip,hengqujushi\/zulip,samatdav\/zulip,hj3938\/zulip,m1ssou\/zulip,kaiyuanheshang\/zulip,andersk\/zulip,kokoar\/zulip,vikas-parashar\/zulip,aakash-cr7\/zulip,blaze225\/zulip,Drooids\/zulip,dattatreya303\/zulip,rishig\/zulip,TigorC\/zulip,MayB\/zulip,timabbott\/zulip,jackrzhang\/zulip,karamcnair\/zulip,ryansnowboarder\/zulip,karamcnair\/zulip,wavelets\/zulip,technicalpickles\/zulip,susansls\/zulip,huangkebo\/zulip,zofuthan\/zulip,swinghu\/zulip,littledogboy\/zulip,mohsenSy\/zulip,Qgap\/zulip,deer-hope\/zulip,mahim97\/zulip,tommyip\/zulip,mdavid\/zulip,udxxabp\/zulip,gkotian\/zulip,LeeRisk\/zulip,Cheppers\/zulip,noroot\/zulip,MayB\/zulip,swinghu\/zulip,punchagan\/zulip,armooo\/zulip,punchagan\/zulip,punchagan\/zulip,jrowan\/zulip,zwily\/zulip,Galexrt\/zulip,brainwane\/zulip,Galexrt\/zulip,Qgap\/zulip,Qgap\/zulip,lfranchi\/zulip,showell\/zulip,praveenaki\/zulip,Frouk\/zulip,eastlhu\/zulip,dotcool\/zulip,itnihao\/zulip,arpitpanwar\/zulip,technicalpickles\/zulip,so0k\/zulip,kokoar\/zulip,joyhchen\/zulip,dxq-git\/zulip,natanovia\/zulip,EasonYi\/zulip,blaze225\/zulip,stamhe\/zulip,firstblade\/zulip,ahmadassaf\/zulip,samatdav\/zulip,DazWorrall\/zulip,avastu\/zulip,mansilladev\/zulip,alliejones\/zulip,jessedhillon\/zulip,zofuthan\/zulip,jainayush975\/zulip,RobotCaleb\/zulip,mansilladev\/zulip,ryanbackman\/zulip,brainwane\/zulip,mansilladev\/zulip,TigorC\/zulip,calvinleenyc\/zulip,LAndreas\/zulip,ericzhou2008\/zulip,atomic-labs\/zulip,timabbott\/zulip,shaunstanislaus\/zulip,nicholasbs\/zulip,dnmfarrell\/zulip,hayderimran7\/zulip,isht3\/zulip,guiquanz\/zulip,moria\/zulip,ryansnowboarder\/zulip,aps-sids\/zulip,atomic-labs\/zulip,so0k\/zulip,SmartPeople\/zulip,jeffcao\/zulip,seapasulli\/zulip,Qgap\/zulip,isht3\/zulip,eeshangarg\/zulip,jessedhillon\/zulip,noroot\/zulip,seapasulli\/zulip,MayB\/zulip,souravbadami\/zulip,joyhchen\/zulip,firstblade\/zulip,Drooids\/zulip,huangkebo\/zulip,synicalsyntax\/zulip,amallia\/zulip,sonali0901\/zulip,bowlofstew\/zulip,proliming\/zulip,zulip\/zulip,gkotian\/zulip,xuanhan863\/zulip,showell\/zulip,amallia\/zulip,rht\/zulip,eastlhu\/zulip,stamhe\/zulip,mansilladev\/zulip,mahim97\/zulip,niftynei\/zulip,alliejones\/zulip,wavelets\/zulip,christi3k\/zulip,wdaher\/zulip,zulip\/zulip,thomasboyt\/zulip,willingc\/zulip,umkay\/zulip,esander91\/zulip,zofuthan\/zulip,mdavid\/zulip,stamhe\/zulip,themass\/zulip,PaulPetring\/zulip,rishig\/zulip,Frouk\/zulip,MariaFaBella85\/zulip,peiwei\/zulip,PhilSk\/zulip,zachallaun\/zulip,ufosky-server\/zulip,seapasulli\/zulip,bluesea\/zulip,jonesgithub\/zulip,akuseru\/zulip,krtkmj\/zulip,kaiyuanheshang\/zulip,stamhe\/zulip,themass\/zulip,johnnygaddarr\/zulip,hayderimran7\/zulip,technicalpickles\/zulip,jonesgithub\/zulip,peiwei\/zulip,guiquanz\/zulip,dawran6\/zulip,hj3938\/zulip,kokoar\/zulip,synicalsyntax\/zulip,niftynei\/zulip,guiquanz\/zulip,jerryge\/zulip,johnny9\/zulip"} {"commit":"79c6c71ab6edd8313fd6c9c6441d69ad04d50721","old_file":"update-database\/stackdoc\/namespaces\/microsoftkb.py","new_file":"update-database\/stackdoc\/namespaces\/microsoftkb.py","old_contents":"import re\nimport urllib\n\n\n############### Functions called by stackdoc\n\ndef get_version():\n return 1\n\ndef get_ids(title, body, tags):\n ids = []\n if \"http:\/\/support.microsoft.com\/\":\n urls = re.findall(r'<a href=\"([^\"]+)\"', body)\n for url in urls:\n m = re.match(\"http:\/\/support\\.microsoft\\.com\/(?:default.aspx\/)?kb\/(\\w+)\", url)\n if m:\n ids.append(m.group(1))\n\n return ids\n\ndef get_tags():\n return None # There isn't a reliable set of tags to filter by. Null indicates that we're not filtering\n\n","new_contents":"import re\nimport urllib\n\n\n############### Functions called by stackdoc\n\ndef get_version():\n return 1\n\ndef get_ids(title, body, tags):\n ids = []\n if \"http:\/\/support.microsoft.com\/\":\n urls = re.findall(r'<a href=\"([^\"]+)\"', body)\n for url in urls:\n m = re.match(\"http:\/\/support\\.microsoft\\.com\/(?:default\\.aspx\/)?[kK][bB]\/(\\w+)\", url)\n if m:\n ids.append(m.group(1))\n m2 = re.match(\"http:\/\/support\\.microsoft\\.com\/(?:default\\.aspx)?\\?scid=[kK][bB];[-\\w]+;(\\w+)\", url)\n if m2:\n ids.append(m2.group(1))\n\n return ids\n\ndef get_tags():\n return None # There isn't a reliable set of tags to filter by. Null indicates that we're not filtering\n\n","subject":"Support another form of KB URL.","message":"Support another form of KB URL.\n","lang":"Python","license":"bsd-3-clause","repos":"alnorth\/stackdoc,alnorth\/stackdoc,alnorth\/stackdoc"} {"commit":"49d602528310a9ca71001504632f0d7138fec6ca","old_file":"pylint_runner\/__init__.py","new_file":"pylint_runner\/__init__.py","old_contents":"\"\"\"\nPylint runner\n\"\"\"\n\n__author__ = \"Matthew Peveler\"\n__license__ = \"The MIT License (MIT)\"\n__version__ = \"0.5.0\"\n","new_contents":"\"\"\"\nPylint runner\n\"\"\"\n\n__author__ = \"Matthew Peveler\"\n__license__ = \"The MIT License (MIT)\"\n__version__ = \"0.5.1\"\n","subject":"Bump version to 0.5.1 for official Python 3.6 support","message":"Bump version to 0.5.1 for official Python 3.6 support","lang":"Python","license":"mit","repos":"MasterOdin\/pylint_runner"} {"commit":"640ad3ed45eef21f2b7a71b4fd73a469ebed4b44","old_file":"reobject\/models\/fields.py","new_file":"reobject\/models\/fields.py","old_contents":"import attr\n\n\ndef Field(*args, default=attr.NOTHING, **kwargs):\n if callable(default):\n default = attr.Factory(default)\n\n return attr.ib(*args, default=default, **kwargs)\n\n\ndef ManyToManyField(cls, *args, **kwargs):\n metadata = {\n 'related': {\n 'target': cls,\n 'type': 'ManyToMany',\n }\n }\n\n return attr.ib(*args, **kwargs, metadata=metadata)\n","new_contents":"import attr\n\n\ndef Field(*args, default=attr.NOTHING, **kwargs):\n if callable(default):\n default = attr.Factory(default)\n\n return attr.ib(*args, default=default, **kwargs)\n\n\ndef ManyToManyField(cls, *args, **kwargs):\n metadata = {\n 'related': {\n 'target': cls,\n 'type': 'ManyToMany',\n }\n }\n\n return attr.ib(*args, metadata=metadata, **kwargs)\n","subject":"Fix tests on Python 3.3 and 3.4","message":"Fix tests on Python 3.3 and 3.4\n","lang":"Python","license":"apache-2.0","repos":"onyb\/reobject,onyb\/reobject"} {"commit":"c32bdff4b0ee570ed58cd869830d89e3251cf82a","old_file":"pytils\/test\/__init__.py","new_file":"pytils\/test\/__init__.py","old_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nUnit tests for pytils\n\"\"\"\n__all__ = [\"test_numeral\", \"test_dt\", \"test_translit\", \"test_utils\", \"test_typo\"]\n\nimport unittest\n\ndef get_django_suite():\n try:\n import django\n except ImportError:\n return unittest.TestSuite()\n \n import pytils.test.templatetags\n return pytils.test.templatetags.get_suite()\n\ndef get_suite():\n \"\"\"Return TestSuite for all unit-test of pytils\"\"\"\n suite = unittest.TestSuite()\n for module_name in __all__:\n imported_module = __import__(\"pytils.test.\"+module_name,\n globals(),\n locals(),\n [\"pytils.test\"])\n \n loader = unittest.defaultTestLoader\n suite.addTest(loader.loadTestsFromModule(imported_module))\n suite.addTest(get_django_suite())\n\n return suite\n\ndef run_tests_from_module(module, verbosity=1):\n \"\"\"Run unit-tests for single module\"\"\"\n suite = unittest.TestSuite()\n loader = unittest.defaultTestLoader\n suite.addTest(loader.loadTestsFromModule(module))\n unittest.TextTestRunner(verbosity=verbosity).run(suite)\n\ndef run(verbosity=1):\n \"\"\"Run all unit-test of pytils\"\"\"\n suite = get_suite()\n unittest.TextTestRunner(verbosity=verbosity).run(suite)\n\nif __name__ == '__main__':\n run(2)\n\n","new_contents":"# -*- coding: utf-8 -*-\n\"\"\"\nUnit tests for pytils\n\"\"\"\n__all__ = [\"test_numeral\", \"test_dt\", \"test_translit\", \"test_utils\", \"test_typo\"]\n\nimport unittest\nimport sys\n\ndef get_django_suite():\n try:\n import django\n except ImportError:\n return unittest.TestSuite()\n \n import pytils.test.templatetags\n return pytils.test.templatetags.get_suite()\n\ndef get_suite():\n \"\"\"Return TestSuite for all unit-test of pytils\"\"\"\n suite = unittest.TestSuite()\n for module_name in __all__:\n imported_module = __import__(\"pytils.test.\"+module_name,\n globals(),\n locals(),\n [\"pytils.test\"])\n \n loader = unittest.defaultTestLoader\n suite.addTest(loader.loadTestsFromModule(imported_module))\n suite.addTest(get_django_suite())\n\n return suite\n\ndef run_tests_from_module(module, verbosity=1):\n \"\"\"Run unit-tests for single module\"\"\"\n suite = unittest.TestSuite()\n loader = unittest.defaultTestLoader\n suite.addTest(loader.loadTestsFromModule(module))\n unittest.TextTestRunner(verbosity=verbosity).run(suite)\n\ndef run(verbosity=1):\n \"\"\"Run all unit-test of pytils\"\"\"\n suite = get_suite()\n res = unittest.TextTestRunner(verbosity=verbosity).run(suite)\n if res.errors or res.failures:\n sys.exit(1)\n\nif __name__ == '__main__':\n run(2)\n\n","subject":"Exit with non-0 status if there are failed tests or errors.","message":"Py3: Exit with non-0 status if there are failed tests or errors.\n","lang":"Python","license":"mit","repos":"Forever-Young\/pytils,j2a\/pytils"} {"commit":"7e25472dab7732dc76bfb81d720946c18811962f","old_file":"src\/appengine\/driver.py","new_file":"src\/appengine\/driver.py","old_contents":"\"\"\"List drivers and send them commands.\"\"\"\nimport logging\n\nimport flask\n\nfrom appengine import device, rest\n\n\nclass Query(object):\n def iter(self):\n for name, cls in device.DEVICE_TYPES.iteritems():\n yield Driver(name, cls)\n\n\nclass Driver(object):\n \"\"\"This is a fake for compatibility with the rest module\"\"\"\n\n def __init__(self, name, cls):\n self._name = name\n self._cls = cls\n\n def to_dict(self):\n return {'name': self._name}\n\n # This is a trampoline through to the driver\n # mainly for commands\n def __getattr__(self, name):\n func = getattr(self._cls, name)\n if func is None or not getattr(func, 'is_static', False):\n logging.error('Command %s does not exist or is not a static command',\n name)\n flask.abort(400)\n return func\n\n @staticmethod\n def query():\n return Query()\n\n @staticmethod\n def get_by_id(_id):\n return Driver(_id, device.DEVICE_TYPES[_id])\n\n\n# pylint: disable=invalid-name\nblueprint = flask.Blueprint('driver', __name__)\nrest.register_class(blueprint, Driver, None)\n","new_contents":"\"\"\"List drivers and send them commands.\"\"\"\nimport logging\n\nimport flask\n\nfrom appengine import device, rest\n\n\nclass Query(object):\n def iter(self):\n for name, cls in device.DEVICE_TYPES.iteritems():\n yield Driver(name, cls)\n\n\nclass Driver(object):\n \"\"\"This is a fake for compatibility with the rest module\"\"\"\n\n def __init__(self, name, cls):\n self._name = name\n self._cls = cls\n\n def to_dict(self):\n return {'name': self._name}\n\n # This is a trampoline through to the driver\n # mainly for commands\n def __getattr__(self, name):\n func = getattr(self._cls, name)\n if func is None or not getattr(func, 'is_static', False):\n logging.error('Command %s does not exist or is not a static command',\n name)\n flask.abort(400)\n return func\n\n @staticmethod\n def put():\n pass\n\n @staticmethod\n def query():\n return Query()\n\n @staticmethod\n def get_by_id(_id):\n return Driver(_id, device.DEVICE_TYPES[_id])\n\n\n# pylint: disable=invalid-name\nblueprint = flask.Blueprint('driver', __name__)\nrest.register_class(blueprint, Driver, None)\n","subject":"Fix 'put is not a command' error on static commands","message":"Fix 'put is not a command' error on static commands\n","lang":"Python","license":"mit","repos":"tomwilkie\/awesomation,tomwilkie\/awesomation,tomwilkie\/awesomation,tomwilkie\/awesomation,tomwilkie\/awesomation"} {"commit":"a54933f5fb5e958c890839c58fcba4e658c8e2a0","old_file":"bitbots_head_behavior\/scripts\/testHeadBehaviour.py","new_file":"bitbots_head_behavior\/scripts\/testHeadBehaviour.py","old_contents":"#!\/usr\/bin\/env python3\n# -*- coding:utf-8 -*-\n\nimport rospy\nfrom humanoid_league_msgs.msg import BallInImage, BallRelative, BallsInImage\nfrom trajectory_msgs.msg import JointTrajectory, JointTrajectoryPoint\n\n\ndef run():\n pub_ball = rospy.Publisher(\"ball_in_image\", BallsInImage, queue_size=1)\n pub_hmg = rospy.Publisher(\"head_motor_goals\", JointTrajectory, queue_size=1)\n\n hmg = JointTrajectory()\n goal = JointTrajectoryPoint()\n goal.positions = [0, 0]\n goal.velocities = [0, 0]\n hmg.points = [goal]\n\n\n counter = 320\n direction = 1\n \n rospy.loginfo(\"Create Test\")\n rospy.init_node(\"bitbots_testHeadBehaviour\")\n pub_hmg.publish(hmg)\n\n rate = rospy.Rate(4)\n rospy.logdebug(\"Laeuft...\")\n while not rospy.is_shutdown():\n # Ball in Image\n ball = BallInImage()\n ball.center.x = counter\n if(counter > 340 or counter < 300):\n direction *= -1\n counter += direction\n else:\n counter += direction\n ball.center.y = 200\n ball.diameter = 10\n ball.confidence = 1\n balls = BallsInImage()\n balls.candidates.append(ball)\n\n pub_ball.publish(balls)\n rospy.loginfo(\"Published ball: %s\" % counter)\n rate.sleep()\n\nif __name__ == \"__main__\":\n run()\n\n","new_contents":"#!\/usr\/bin\/env python3\n# -*- coding:utf-8 -*-\n\nimport rospy\nfrom humanoid_league_msgs.msg import BallInImage, BallRelative, BallInImageArray\nfrom trajectory_msgs.msg import JointTrajectory, JointTrajectoryPoint\n\n\ndef run():\n pub_ball = rospy.Publisher(\"ball_in_image\", BallsInImage, queue_size=1)\n pub_hmg = rospy.Publisher(\"head_motor_goals\", JointTrajectory, queue_size=1)\n\n hmg = JointTrajectory()\n goal = JointTrajectoryPoint()\n goal.positions = [0, 0]\n goal.velocities = [0, 0]\n hmg.points = [goal]\n\n\n counter = 320\n direction = 1\n\n rospy.loginfo(\"Create Test\")\n rospy.init_node(\"bitbots_testHeadBehaviour\")\n pub_hmg.publish(hmg)\n\n rate = rospy.Rate(4)\n rospy.logdebug(\"Laeuft...\")\n while not rospy.is_shutdown():\n # Ball in Image\n ball = BallInImage()\n ball.center.x = counter\n if(counter > 340 or counter < 300):\n direction *= -1\n counter += direction\n else:\n counter += direction\n ball.center.y = 200\n ball.diameter = 10\n ball.confidence = 1\n balls = BallInImageArray()\n balls.candidates.append(ball)\n\n pub_ball.publish(balls)\n rospy.loginfo(\"Published ball: %s\" % counter)\n rate.sleep()\n\nif __name__ == \"__main__\":\n run()\n\n","subject":"Test Head Behavior: Apply new HLM","message":"Test Head Behavior: Apply new HLM\n","lang":"Python","license":"bsd-3-clause","repos":"bit-bots\/bitbots_behaviour"} {"commit":"14d0e3b887b469c2b1979352804d8ade3184ef18","old_file":"scripts\/symlinks\/parent\/foogroup.py","new_file":"scripts\/symlinks\/parent\/foogroup.py","old_contents":"#!\/usr\/bin\/env python\nimport json\n\nprint json.dumps({\n \"_meta\": {\n \"hostvars\": {\n 'afoo': {}\n },\n },\n \"foo\": {\n \"hosts\": ['afoo']\n }\n})","new_contents":"#!\/usr\/bin\/env python\nimport json\n\nprint(json.dumps({\n \"_meta\": {\n \"hostvars\": {\n 'afoo': {}\n },\n },\n \"foo\": {\n \"hosts\": ['afoo']\n }\n}))\n","subject":"Fix print statement to be py3 compatible","message":"Fix print statement to be py3 compatible\n","lang":"Python","license":"mit","repos":"AlanCoding\/Ansible-inventory-file-examples,AlanCoding\/Ansible-inventory-file-examples"} {"commit":"24e80d80034084f6d2067df39fdc070e4eb41447","old_file":"diceclient.py","new_file":"diceclient.py","old_contents":"from twisted.internet import reactor, defer\nfrom twisted.internet.protocol import ClientCreator\nfrom twisted.protocols import amp\nfrom ampserver import Sum, Divide\n\n\ndef doMath():\n d1 = ClientCreator(reactor, amp.AMP).connectTCP(\n '127.0.0.1', 1234).addCallback(\n lambda p: p.callRemote(Sum, a=13, b=81)).addCallback(\n lambda result: result['total'])\n def trapZero(result):\n result.trap(ZeroDivisionError)\n print \"Divided by zero: returning INF\"\n return 1e1000\n d2 = ClientCreator(reactor, amp.AMP).connectTCP(\n '127.0.0.1', 1234).addCallback(\n lambda p: p.callRemote(Divide, numerator=1234,\n denominator=0)).addErrback(trapZero)\n def done(result):\n print 'Done with math:', result\n defer.DeferredList([d1, d2]).addCallback(done)\n\nif __name__ == '__main__':\n doMath()\n reactor.run()\n","new_contents":"from twisted.internet import reactor, defer\nfrom twisted.internet.protocol import ClientCreator\nfrom twisted.protocols import amp\nfrom diceserver import Sum, Divide\n\n\ndef doMath():\n d1 = ClientCreator(reactor, amp.AMP).connectTCP(\n '127.0.0.1', 1234).addCallback(\n lambda p: p.callRemote(Sum, a=13, b=81)).addCallback(\n lambda result: result['total'])\n def trapZero(result):\n result.trap(ZeroDivisionError)\n print \"Divided by zero: returning INF\"\n return 1e1000\n d2 = ClientCreator(reactor, amp.AMP).connectTCP(\n '127.0.0.1', 1234).addCallback(\n lambda p: p.callRemote(Divide, numerator=1234,\n denominator=0)).addErrback(trapZero)\n def done(result):\n print 'Done with math:', result\n defer.DeferredList([d1, d2]).addCallback(done)\n\nif __name__ == '__main__':\n doMath()\n reactor.run()\n","subject":"Fix import path to match rename","message":"Fix import path to match rename\n","lang":"Python","license":"mit","repos":"dripton\/ampchat"} {"commit":"14dd9f6cab99be6832ab98291337f4d38faae936","old_file":"fellowms\/forms.py","new_file":"fellowms\/forms.py","old_contents":"from django.forms import ModelForm, widgets\n\nfrom .models import Fellow, Event, Expense, Blog\n\nclass FellowForm(ModelForm):\n class Meta:\n model = Fellow\n exclude = [\n \"home_lon\",\n \"home_lat\",\n \"inauguration_year\",\n \"funding_notes\",\n \"mentor\",\n ]\n\n\nclass EventForm(ModelForm):\n class Meta:\n model = Event\n exclude = [\n \"status\",\n \"budget_approve\",\n \"report_url\",\n ]\n\n # We don't want to expose fellows' data\n # so we will request the email\n # and match on the database.\n labels = {\n 'fellow': 'Fellow',\n 'url': \"Event's homepage url\",\n 'name': \"Event's name\",\n }\n\n\nclass ExpenseForm(ModelForm):\n class Meta:\n model = Expense\n exclude = [\n 'id',\n 'status',\n ]\n\n\nclass BlogForm(ModelForm):\n class Meta:\n model = Blog\n exclude = [\n 'status',\n ]\n","new_contents":"from django.forms import ModelForm, widgets\n\nfrom .models import Fellow, Event, Expense, Blog\n\nclass FellowForm(ModelForm):\n class Meta:\n model = Fellow\n exclude = [\n \"user\",\n \"home_lon\",\n \"home_lat\",\n \"inauguration_year\",\n \"funding_notes\",\n \"mentor\",\n ]\n\n\nclass EventForm(ModelForm):\n class Meta:\n model = Event\n exclude = [\n \"status\",\n \"budget_approve\",\n \"report_url\",\n ]\n\n # We don't want to expose fellows' data\n # so we will request the email\n # and match on the database.\n labels = {\n 'fellow': 'Fellow',\n 'url': \"Event's homepage url\",\n 'name': \"Event's name\",\n }\n\n\nclass ExpenseForm(ModelForm):\n class Meta:\n model = Expense\n exclude = [\n 'id',\n 'status',\n ]\n\n\nclass BlogForm(ModelForm):\n class Meta:\n model = Blog\n exclude = [\n 'status',\n ]\n","subject":"Exclude user field from form","message":"Exclude user field from form\n","lang":"Python","license":"bsd-3-clause","repos":"softwaresaved\/fat,softwaresaved\/fat,softwaresaved\/fat,softwaresaved\/fat"} {"commit":"785208c904caacd69cb98f9ea44ee9f720752baf","old_file":"src\/tmlib\/imextract\/argparser.py","new_file":"src\/tmlib\/imextract\/argparser.py","old_contents":"'''\nArguments of the command line program.\n'''\n\nfrom . import __version__\nfrom .cli import Imextract\nfrom .args import ImextractInitArgs\n\n\nparser, subparsers = Imextract.get_parser_and_subparsers(\n required_subparsers=['init', 'run', 'submit', 'cleanup'])\n\nparser.description = '''\n Extract images from heterogeneous microscopic image file formats\n and store each 2D plane in a separate PNG file.\n'''\nparser.version = __version__\n\ninit_parser = subparsers.choices['init']\ninit_extra_group = init_parser.add_argument_group(\n 'additional program-specific arguments')\nImextractInitArgs().add_to_argparser(init_extra_group)\n\nfor name in subparsers.choices:\n subparsers.choices[name].set_defaults(handler=Imextract.call)\n","new_contents":"'''\nArguments of the command line program.\n'''\n\nfrom . import __version__\nfrom .cli import Imextract\nfrom .args import ImextractInitArgs\n\n\nparser, subparsers = Imextract.get_parser_and_subparsers(\n required_subparsers=['init', 'run', 'submit', 'cleanup', 'log'])\n\nparser.description = '''\n Extract images from heterogeneous microscopic image file formats\n and store each 2D plane in a separate PNG file.\n'''\nparser.version = __version__\n\ninit_parser = subparsers.choices['init']\ninit_extra_group = init_parser.add_argument_group(\n 'additional program-specific arguments')\nImextractInitArgs().add_to_argparser(init_extra_group)\n\nfor name in subparsers.choices:\n subparsers.choices[name].set_defaults(handler=Imextract.call)\n","subject":"Fix bug in imextract argument parser module","message":"Fix bug in imextract argument parser module\n","lang":"Python","license":"agpl-3.0","repos":"TissueMAPS\/TmLibrary,TissueMAPS\/TmLibrary,TissueMAPS\/TmLibrary,TissueMAPS\/TmLibrary,TissueMAPS\/TmLibrary"} {"commit":"629bfe7ba928bc9650217b90190409708740ee82","old_file":"lib\/cretonne\/meta\/isa\/intel\/defs.py","new_file":"lib\/cretonne\/meta\/isa\/intel\/defs.py","old_contents":"\"\"\"\nIntel definitions.\n\nCommonly used definitions.\n\"\"\"\nfrom __future__ import absolute_import\nfrom cdsl.isa import TargetISA, CPUMode\nimport base.instructions\nfrom . import instructions as x86\n\nISA = TargetISA('intel', [base.instructions.GROUP, x86.GROUP])\n\n# CPU modes for 32-bit and 64-bit operation.\nI32 = CPUMode('I32', ISA)\nI64 = CPUMode('I64', ISA)\n","new_contents":"\"\"\"\nIntel definitions.\n\nCommonly used definitions.\n\"\"\"\nfrom __future__ import absolute_import\nfrom cdsl.isa import TargetISA, CPUMode\nimport base.instructions\nfrom . import instructions as x86\n\nISA = TargetISA('intel', [base.instructions.GROUP, x86.GROUP])\n\n# CPU modes for 32-bit and 64-bit operation.\nI64 = CPUMode('I64', ISA)\nI32 = CPUMode('I32', ISA)\n","subject":"Define I64 before I32 for better encoding table compression.","message":"Define I64 before I32 for better encoding table compression.\n\nThe encoding list compression algorithm is not the sharpest knife in the\ndrawer. It can reuse subsets of I64 encoding lists for I32 instructions,\nbut only when the I64 lists are defined first.\n\nWith this change and the previous change to the encoding list format, we\nget the following table sizes for the Intel ISA:\n\nENCLISTS: 1478 B -> 662 B\nLEVEL2: 1072 B (unchanged)\nLEVEL1: 32 B -> 48 B\nTotal: 2582 B -> 1782 B (-31%)\n","lang":"Python","license":"apache-2.0","repos":"sunfishcode\/cretonne,stoklund\/cretonne,sunfishcode\/cretonne,stoklund\/cretonne,stoklund\/cretonne,sunfishcode\/cretonne"} {"commit":"d028f66964249bab928a29d92ab4cff075352546","old_file":"integration\/main.py","new_file":"integration\/main.py","old_contents":"from spec import Spec, skip\n\n\nclass Tessera(Spec):\n def is_importable(self):\n import tessera\n assert tessera.app\n assert tessera.db\n","new_contents":"from contextlib import contextmanager\nimport os\nfrom shutil import rmtree\nfrom tempfile import mkdtemp\n\nfrom spec import Spec, skip\n\n\n@contextmanager\ndef _tmp():\n try:\n tempdir = mkdtemp()\n yield tempdir\n finally:\n rmtree(tempdir)\n\n@contextmanager\ndef _db():\n with _tmp() as tempdir:\n from tessera import app, db\n # Temp db location\n path = os.path.join(tempdir, 'tessera.db')\n dbfile = 'sqlite:\/\/\/{0}'.format(path)\n # Inform app of that location & setup\n app.config.from_object(_config(SQLALCHEMY_DATABASE_URI=dbfile))\n db.create_all()\n # Let test have its way with that temp db\n yield db\n\n\nclass Config(object):\n pass\n\ndef _config(**options):\n config = Config()\n for key, value in options.iteritems():\n setattr(config, key, value)\n\n\nclass Tessera(Spec):\n def is_importable(self):\n import tessera\n assert tessera.app\n assert tessera.db\n\n def creates_a_nonempty_database_schema(self):\n with _db() as db:\n meta = db.MetaData()\n meta.reflect(db.engine)\n assert len(meta.tables) > 0\n","subject":"Add temp DB test harness + basic test","message":"Add temp DB test harness + basic test\n","lang":"Python","license":"apache-2.0","repos":"tessera-metrics\/tessera,jmptrader\/tessera,aalpern\/tessera,Slach\/tessera,filippog\/tessera,aalpern\/tessera,aalpern\/tessera,section-io\/tessera,urbanairship\/tessera,aalpern\/tessera,urbanairship\/tessera,Slach\/tessera,jmptrader\/tessera,urbanairship\/tessera,Slach\/tessera,urbanairship\/tessera,urbanairship\/tessera,tessera-metrics\/tessera,section-io\/tessera,aalpern\/tessera,tessera-metrics\/tessera,filippog\/tessera,tessera-metrics\/tessera,section-io\/tessera,jmptrader\/tessera,jmptrader\/tessera,filippog\/tessera,section-io\/tessera,Slach\/tessera,jmptrader\/tessera,tessera-metrics\/tessera"} {"commit":"1100830d3b48262dd9b94d96eb50d75c8ff69fe4","old_file":"Cogs\/Emoji.py","new_file":"Cogs\/Emoji.py","old_contents":"import discord\nfrom discord.ext import commands\nfrom Cogs import GetImage\n\ndef setup(bot):\n bot.add_cog(Emoji(bot))\n\nclass Emoji(commands.Cog):\n\n def __init__(self, bot):\n self.bot = bot\n\n @commands.command()\n async def emoji(self, ctx, emoji = None):\n '''Outputs your CUSTOM emoji... but bigger! (Does not work with standard discord emojis)'''\n emojiparts = emoji.replace(\"<\",\"\").replace(\">\",\"\").split(\":\") if emoji else []\n if not len(emojiparts) == 3:\n await ctx.send(\"Usage: `{}emoji [emoji]` - must be a CUSTOM emoji, and not just the name.\".format(ctx.prefix))\n return\n emoji_obj = discord.PartialEmoji(animated=len(emojiparts[0]) > 0, name=emojiparts[1], id=emojiparts[2])\n if not emoji_obj.url:\n await ctx.send(\"Could not find url for emoji :(\")\n return\n f = await GetImage.download(emoji_obj.url)\n if not f:\n await ctx.send(\"I couldn't get that emoji :(\")\n return\n await ctx.send(file=discord.File(f))","new_contents":"import discord\nfrom discord.ext import commands\nfrom Cogs import GetImage\n\ndef setup(bot):\n bot.add_cog(Emoji(bot))\n\nclass Emoji(commands.Cog):\n\n def __init__(self, bot):\n self.bot = bot\n\n @commands.command()\n async def emoji(self, ctx, emoji = None):\n '''Outputs the passed emoji... but bigger!'''\n if emoji is None:\n await ctx.send(\"Usage: `{}emoji [emoji]`\".format(ctx.prefix))\n return\n if len(emoji) < 3:\n # Try to get just the unicode\n h = \"-\".join([hex(ord(x)).lower()[2:] for x in emoji])\n url = \"https:\/\/raw.githubusercontent.com\/twitter\/twemoji\/gh-pages\/2\/72x72\/{}.png\".format(h)\n f = await GetImage.download(url)\n if not f:\n await ctx.send(\"I couldn't get that emoji :(\")\n else:\n await ctx.send(file=discord.File(f))\n return\n emojiparts = emoji.replace(\"<\",\"\").replace(\">\",\"\").split(\":\") if emoji else []\n if not len(emojiparts) == 3:\n await ctx.send(\"Usage: `{}emoji [emoji]`\".format(ctx.prefix))\n return\n emoji_obj = discord.PartialEmoji(animated=len(emojiparts[0]) > 0, name=emojiparts[1], id=emojiparts[2])\n if not emoji_obj.url:\n await ctx.send(\"Could not find a url for that emoji :(\")\n return\n f = await GetImage.download(emoji_obj.url)\n if not f:\n await ctx.send(\"I couldn't get that emoji :(\")\n return\n await ctx.send(file=discord.File(f))\n","subject":"Add support for built-in emojis","message":"Add support for built-in emojis","lang":"Python","license":"mit","repos":"corpnewt\/CorpBot.py,corpnewt\/CorpBot.py"} {"commit":"6464028097b13b5d03969c20bae56f9f70acbbd1","old_file":"saleor\/cart\/middleware.py","new_file":"saleor\/cart\/middleware.py","old_contents":"from __future__ import unicode_literals\n\nfrom . import SessionCart, CART_SESSION_KEY\n\n\nclass CartMiddleware(object):\n '''\n Saves the cart instance into the django session.\n '''\n\n def process_request(self, request):\n try:\n cart_data = request.session[CART_SESSION_KEY]\n cart = SessionCart.from_storage(cart_data)\n except KeyError:\n cart = SessionCart()\n setattr(request, 'cart', cart)\n\n def process_response(self, request, response):\n if hasattr(request, 'cart'):\n request.session[CART_SESSION_KEY] = request.cart.for_storage()\n return response\n","new_contents":"from __future__ import unicode_literals\n\nfrom . import SessionCart, CART_SESSION_KEY\n\n\nclass CartMiddleware(object):\n '''\n Saves the cart instance into the django session.\n '''\n\n def process_request(self, request):\n try:\n cart_data = request.session[CART_SESSION_KEY]\n cart = SessionCart.from_storage(cart_data)\n except KeyError:\n cart = SessionCart()\n setattr(request, 'cart', cart)\n\n def process_response(self, request, response):\n if hasattr(request, 'cart') and request.cart.modified:\n request.session[CART_SESSION_KEY] = request.cart.for_storage()\n return response\n","subject":"Store cart in session only when it was modified","message":"Store cart in session only when it was modified\n","lang":"Python","license":"bsd-3-clause","repos":"HyperManTT\/ECommerceSaleor,taedori81\/saleor,car3oon\/saleor,UITools\/saleor,rodrigozn\/CW-Shop,mociepka\/saleor,spartonia\/saleor,UITools\/saleor,arth-co\/saleor,UITools\/saleor,paweltin\/saleor,hongquan\/saleor,Drekscott\/Motlaesaleor,UITools\/saleor,avorio\/saleor,josesanch\/saleor,tfroehlich82\/saleor,maferelo\/saleor,spartonia\/saleor,dashmug\/saleor,taedori81\/saleor,HyperManTT\/ECommerceSaleor,Drekscott\/Motlaesaleor,UITools\/saleor,jreigel\/saleor,josesanch\/saleor,dashmug\/saleor,maferelo\/saleor,paweltin\/saleor,arth-co\/saleor,Drekscott\/Motlaesaleor,KenMutemi\/saleor,car3oon\/saleor,mociepka\/saleor,taedori81\/saleor,dashmug\/saleor,laosunhust\/saleor,taedori81\/saleor,jreigel\/saleor,josesanch\/saleor,itbabu\/saleor,hongquan\/saleor,rchav\/vinerack,laosunhust\/saleor,paweltin\/saleor,HyperManTT\/ECommerceSaleor,KenMutemi\/saleor,avorio\/saleor,rchav\/vinerack,arth-co\/saleor,hongquan\/saleor,arth-co\/saleor,mociepka\/saleor,jreigel\/saleor,laosunhust\/saleor,spartonia\/saleor,Drekscott\/Motlaesaleor,itbabu\/saleor,tfroehlich82\/saleor,rodrigozn\/CW-Shop,KenMutemi\/saleor,paweltin\/saleor,maferelo\/saleor,rchav\/vinerack,avorio\/saleor,spartonia\/saleor,tfroehlich82\/saleor,itbabu\/saleor,rodrigozn\/CW-Shop,laosunhust\/saleor,avorio\/saleor,car3oon\/saleor"} {"commit":"3dfa781ce8e073f40eda3d80794ad1caff5d5920","old_file":"samples\/migrateAccount.py","new_file":"samples\/migrateAccount.py","old_contents":"#### Migrate person to a new account within the same Org\r\n\r\n# Requires admin role\r\n# Useful when migrating to Enterprise Logins.\r\n# Reassigns all items\/groups to new owner and\r\n# adds userTo to all groups which userFrom is a member.'''\r\n\r\nfrom agoTools.admin import Admin\r\nmyAgol = Admin('<username>') # Replace <username> your ADMIN account\r\n\r\n# for migrating a single account...\r\nmyAgol.migrateAccount(myAgol, '<userFrom>', '<userTo>') # Replace with usernames between which you are moving items\r\n\r\n# for migrating a batch of accounts...\r\nmyAgol.migrateAccounts(myAgol, <path to user mapping CSV>) # Replace with path to CSV file with col1=userFrom, col2=userTo","new_contents":"#### Migrate a member to a new account within the same Organization\r\n\r\n# Requires admin role\r\n# Useful when migrating to Enterprise Logins\r\n# Reassigns all items\/groups to new owner\r\n# Adds userTo to all groups which userFrom is a member\r\n\r\nfrom agoTools.admin import Admin\r\nmyAgol = Admin('<username>') # Replace <username> your ADMIN account\r\n\r\n# un-comment one of the lines below, depending on which workflow you wish to use\r\n\r\n### for migrating a single account...\r\n# myAgol.migrateAccount(myAgol, '<userFrom>', '<userTo>') # Replace with usernames between which you are moving items\r\n\r\n### for migrating multiple accounts...\r\n# myAgol.migrateAccounts(myAgol, <path to user mapping CSV>) # Replace with path to CSV file with col1=userFrom, col2=userTo","subject":"Enhance comments in Migrate Account sample","message":"Enhance comments in Migrate Account sample\n","lang":"Python","license":"apache-2.0","repos":"oevans\/ago-tools"} {"commit":"a90c2eecf95323a6f968e1313c3d7852e4eb25b2","old_file":"speeches\/management\/commands\/populatespeakers.py","new_file":"speeches\/management\/commands\/populatespeakers.py","old_contents":"from django.core.management.base import NoArgsCommand\nfrom django.conf import settings\nfrom popit import PopIt\nfrom speeches.models import Speaker\n\nclass Command(NoArgsCommand):\n help = 'Populates the database with people from Popit'\n\n def handle_noargs(self, **options):\n api = PopIt(instance = settings.POPIT_INSTANCE,\n hostname = settings.POPIT_HOSTNAME,\n api_version = settings.POPIT_API_VERSION)\n results = api.person.get()\n for person in results['results']:\n speaker, created = Speaker.objects.get_or_create(popit_id=person['_id'])\n # we ignore created for now, just always set the name\n speaker.name = person['name']\n speaker.save();\n","new_contents":"import logging\nfrom django.core.management.base import NoArgsCommand\nfrom django.conf import settings\nfrom popit import PopIt\nfrom speeches.models import Speaker\n\nlogger = logging.getLogger(__name__)\n\nclass Command(NoArgsCommand):\n\n help = 'Populates the database with people from Popit'\n\n def handle_noargs(self, **options):\n api = PopIt(instance = settings.POPIT_INSTANCE,\n hostname = settings.POPIT_HOSTNAME,\n api_version = settings.POPIT_API_VERSION)\n results = api.person.get()\n for person in results['results']:\n\n logger.warn('Processing: {0}'.format(person['meta']['api_url']))\n \n speaker, created = Speaker.objects.get_or_create(popit_url=person['meta']['api_url'])\n\n logger.warn('Person was created? {0}'.format(created))\n logger.warn('Persons id in the spoke db is: {0}'.format(speaker.id))\n\n # we ignore created for now, just always set the name\n speaker.name = person['name']\n speaker.save();\n","subject":"Update speaker population command to set popit_url instead of popit_id","message":"Update speaker population command to set popit_url instead of popit_id\n","lang":"Python","license":"agpl-3.0","repos":"opencorato\/sayit,opencorato\/sayit,opencorato\/sayit,opencorato\/sayit"} {"commit":"a5ef9a5d141ba5fd0d1d6c983cd8ac82079a1782","old_file":"run_tests.py","new_file":"run_tests.py","old_contents":"#!\/usr\/bin\/env python3\n\nimport os\nimport tempfile\nfrom distutils.sysconfig import get_python_lib\n\nfrom coalib.tests.TestHelper import TestHelper\n\n\nif __name__ == '__main__':\n parser = TestHelper.create_argparser(description=\"Runs coalas tests.\")\n parser.add_argument(\"-b\",\n \"--ignore-bear-tests\",\n help=\"ignore bear tests\",\n action=\"store_true\")\n parser.add_argument(\"-m\",\n \"--ignore-main-tests\",\n help=\"ignore main program tests\",\n action=\"store_true\")\n\n testhelper = TestHelper(parser)\n\n if not testhelper.args.ignore_main_tests:\n testhelper.add_test_files(os.path.abspath(os.path.join(\"coalib\",\n \"tests\")))\n if not testhelper.args.ignore_bear_tests:\n testhelper.add_test_files(os.path.abspath(os.path.join(\"bears\",\n \"tests\")))\n\n ignore_list = [\n os.path.join(tempfile.gettempdir(), \"**\"),\n os.path.join(get_python_lib(), \"**\"),\n os.path.join(\"coalib\", \"tests\", \"**\"),\n os.path.join(\"bears\", \"tests\", \"**\")\n ]\n\n exit(testhelper.execute_python3_files(ignore_list))\n","new_contents":"#!\/usr\/bin\/env python3\n\nimport os\nimport tempfile\nfrom distutils.sysconfig import get_python_lib\n\nfrom coalib.tests.TestHelper import TestHelper\n\n\nif __name__ == '__main__':\n parser = TestHelper.create_argparser(description=\"Runs coalas tests.\")\n parser.add_argument(\"-b\",\n \"--ignore-bear-tests\",\n help=\"ignore bear tests\",\n action=\"store_true\")\n parser.add_argument(\"-m\",\n \"--ignore-main-tests\",\n help=\"ignore main program tests\",\n action=\"store_true\")\n\n testhelper = TestHelper(parser)\n\n if not testhelper.args.ignore_main_tests:\n testhelper.add_test_files(os.path.abspath(os.path.join(\"coalib\",\n \"tests\")))\n if not testhelper.args.ignore_bear_tests:\n testhelper.add_test_files(os.path.abspath(os.path.join(\"bears\",\n \"tests\")))\n\n ignore_list = [\n os.path.join(tempfile.gettempdir(), \"**\"),\n os.path.join(os.path.dirname(get_python_lib()), \"**\"),\n os.path.join(\"coalib\", \"tests\", \"**\"),\n os.path.join(\"bears\", \"tests\", \"**\")\n ]\n\n exit(testhelper.execute_python3_files(ignore_list))\n","subject":"Update run_test.py to fix coverage","message":"tests: Update run_test.py to fix coverage\n","lang":"Python","license":"agpl-3.0","repos":"Asalle\/coala,ManjiriBirajdar\/coala,jayvdb\/coala,Asnelchristian\/coala,RJ722\/coala,abhiroyg\/coala,FeodorFitsner\/coala,meetmangukiya\/coala,sils1297\/coala,Tanmay28\/coala,yashLadha\/coala,Asalle\/coala,scottbelden\/coala,stevemontana1980\/coala,sophiavanvalkenburg\/coala,Tanmay28\/coala,JohnS-01\/coala,Nosferatul\/coala,yashLadha\/coala,SambitAcharya\/coala,arjunsinghy96\/coala,d6e\/coala,sagark123\/coala,Tanmay28\/coala,mr-karan\/coala,SanketDG\/coala,CruiseDevice\/coala,sagark123\/coala,refeed\/coala,NalinG\/coala,SambitAcharya\/coala,AbdealiJK\/coala,sudheesh001\/coala,kartikeys98\/coala,coala-analyzer\/coala,karansingh1559\/coala,NalinG\/coala,CruiseDevice\/coala,Shade5\/coala,RJ722\/coala,lonewolf07\/coala,rresol\/coala,NiklasMM\/coala,impmihai\/coala,SanketDG\/coala,dagdaggo\/coala,Tanmay28\/coala,incorrectusername\/coala,coala-analyzer\/coala,CruiseDevice\/coala,Shade5\/coala,vinc456\/coala,yashtrivedi96\/coala,NiklasMM\/coala,coala-analyzer\/coala,nemaniarjun\/coala,JohnS-01\/coala,abhiroyg\/coala,coala\/coala,Tanmay28\/coala,swatilodha\/coala,Shade5\/coala,SambitAcharya\/coala,dagdaggo\/coala,incorrectusername\/coala,mr-karan\/coala,nemaniarjun\/coala,stevemontana1980\/coala,Asnelchristian\/coala,MattAllmendinger\/coala,yashtrivedi96\/coala,coala\/coala,MattAllmendinger\/coala,shreyans800755\/coala,dagdaggo\/coala,damngamerz\/coala,meetmangukiya\/coala,SambitAcharya\/coala,arush0311\/coala,Balaji2198\/coala,ManjiriBirajdar\/coala,rimacone\/testing2,andreimacavei\/coala,sils1297\/coala,rresol\/coala,AbdealiJK\/coala,AdeshAtole\/coala,netman92\/coala,Nosferatul\/coala,Uran198\/coala,arjunsinghy96\/coala,yland\/coala,karansingh1559\/coala,SambitAcharya\/coala,arush0311\/coala,vinc456\/coala,NalinG\/coala,shreyans800755\/coala,NalinG\/coala,tushar-rishav\/coala,incorrectusername\/coala,arafsheikh\/coala,andreimacavei\/coala,Tanmay28\/coala,d6e\/coala,tltuan\/coala,sophiavanvalkenburg\/coala,svsn2117\/coala,arush0311\/coala,lonewolf07\/coala,meetmangukiya\/coala,aptrishu\/coala,MattAllmendinger\/coala,swatilodha\/coala,SanketDG\/coala,RJ722\/coala,damngamerz\/coala,aptrishu\/coala,kartikeys98\/coala,scriptnull\/coala,rresol\/coala,sils1297\/coala,ayushin78\/coala,tltuan\/coala,ayushin78\/coala,swatilodha\/coala,saurabhiiit\/coala,impmihai\/coala,svsn2117\/coala,FeodorFitsner\/coala,shreyans800755\/coala,NiklasMM\/coala,d6e\/coala,arjunsinghy96\/coala,coala\/coala,Asalle\/coala,scriptnull\/coala,abhiroyg\/coala,FeodorFitsner\/coala,AdeshAtole\/coala,Uran198\/coala,tushar-rishav\/coala,tushar-rishav\/coala,yland\/coala,karansingh1559\/coala,vinc456\/coala,aptrishu\/coala,MariosPanag\/coala,impmihai\/coala,Balaji2198\/coala,netman92\/coala,mr-karan\/coala,djkonro\/coala,netman92\/coala,nemaniarjun\/coala,damngamerz\/coala,scriptnull\/coala,NalinG\/coala,kartikeys98\/coala,MariosPanag\/coala,Asnelchristian\/coala,scriptnull\/coala,refeed\/coala,Balaji2198\/coala,djkonro\/coala,saurabhiiit\/coala,arafsheikh\/coala,scriptnull\/coala,scottbelden\/coala,sudheesh001\/coala,Tanmay28\/coala,Uran198\/coala,sophiavanvalkenburg\/coala,MariosPanag\/coala,refeed\/coala,jayvdb\/coala,jayvdb\/coala,svsn2117\/coala,rimacone\/testing2,stevemontana1980\/coala,sudheesh001\/coala,yland\/coala,scriptnull\/coala,ManjiriBirajdar\/coala,AbdealiJK\/coala,Tanmay28\/coala,andreimacavei\/coala,NalinG\/coala,saurabhiiit\/coala,AdeshAtole\/coala,ayushin78\/coala,lonewolf07\/coala,scriptnull\/coala,tltuan\/coala,rimacone\/testing2,yashLadha\/coala,SambitAcharya\/coala,scottbelden\/coala,sagark123\/coala,JohnS-01\/coala,NalinG\/coala,arafsheikh\/coala,SambitAcharya\/coala,Nosferatul\/coala,djkonro\/coala,yashtrivedi96\/coala"} {"commit":"6de9457215e5a41a40acaf428132f46ab94fed2c","old_file":"miniraf\/combine.py","new_file":"miniraf\/combine.py","old_contents":"import astropy.io.fits as fits\nimport numpy as np\nimport sys\n\nfrom .util import stack_fits_data\n\nMETHOD_MAP = {\"median\": lambda x: np.median(x, axis=0, overwrite_input=True),\n \"average\": lambda x: np.average(x, axis=0),\n \"sum\": lambda x: np.sum(x, axis=0)}\n\ndef create_parser(subparsers):\n parser_combine = subparsers.add_parser(\"combine\", help=\"combine help\")\n parser_combine.add_argument(\"-m\", \"--method\", choices=[\"median\", \"average\", \"sum\"], required=True)\n parser_combine.add_argument(\"-o\", \"--output\", metavar=\"OUTFILE\", default=sys.stdout.buffer)\n parser_combine.add_argument(\"file\", nargs=\"+\")\n parser_combine.set_defaults(func=main)\n\ndef main(args):\n image_stack = stack_fits_data(args.file)\n result = METHOD_MAP[args.method](image_stack)\n hdu = fits.PrimaryHDU(result)\n hdu.writeto(args.output)\n","new_contents":"import astropy.io.fits as fits\nimport numpy as np\nimport sys\n\nfrom .util import stack_fits_data\n\nMETHOD_MAP = {\"median\": lambda x: np.median(x, axis=0, overwrite_input=True),\n \"average\": lambda x: np.mean(x, axis=0),\n \"sum\": lambda x: np.sum(x, axis=0)}\n\ndef create_parser(subparsers):\n parser_combine = subparsers.add_parser(\"combine\", help=\"combine help\")\n parser_combine.add_argument(\"-m\", \"--method\", choices=[\"median\", \"average\", \"sum\"], required=True)\n parser_combine.add_argument(\"-o\", \"--output\", metavar=\"OUTFILE\", default=sys.stdout.buffer)\n parser_combine.add_argument(\"file\", nargs=\"+\")\n parser_combine.set_defaults(func=main)\n\ndef main(args):\n image_stack = stack_fits_data(args.file)\n result = METHOD_MAP[args.method](image_stack)\n hdu = fits.PrimaryHDU(result)\n hdu.writeto(args.output)\n","subject":"Use np.mean instead for unweighted mean","message":"Use np.mean instead for unweighted mean\n\nSigned-off-by: Lizhou Sha <d6acb26e253550574bc1141efa0eb5e6de15daeb@mit.edu>\n","lang":"Python","license":"mit","repos":"vulpicastor\/miniraf"} {"commit":"81f2a561ac27d13fb43edae1fb94b237951ff9cc","old_file":"tests\/rietveld\/test_braggtree.py","new_file":"tests\/rietveld\/test_braggtree.py","old_contents":"from __future__ import absolute_import, print_function\nimport unittest\nfrom qtpy.QtWidgets import QApplication\nfrom addie.rietveld.braggtree import BraggTree, BankRegexException\n\n\nclass BraggTreeTests(unittest.TestCase):\n def setUp(self):\n self.main_window = QApplication([])\n\n def tearDown(self):\n self.main_window.quit()\n\n def test_get_bank_id(self):\n \"\"\"Test we can extract a bank id from bank workspace name\"\"\"\n braggtree = BraggTree(None)\n target = 12345\n bank_wksp_name = \"Bank {} - 90.0\".format(target)\n bank_id = braggtree._get_bank_id(bank_wksp_name)\n self.assertEqual(int(bank_id), target)\n\n def test_get_bank_id_exception(self):\n \"\"\"Test for raised exception from a bad workspace name\"\"\"\n braggtree = BraggTree(None)\n bad_ws = \"Bank jkl 1 -- 90.0\"\n self.assertRaises(BankRegexException, braggtree._get_bank_id, bad_ws)\n\n def test_do_plot_ws_exception(self):\n \"\"\"Test for raised exception from MainWindow==None\"\"\"\n braggtree = BraggTree(None)\n self.assertRaises(NotImplementedError, braggtree.do_plot_ws)\n\n\nif __name__ == '__main__':\n unittest.main() # pragma: no cover\n","new_contents":"from __future__ import absolute_import, print_function\nimport pytest\nfrom addie.main import MainWindow\nfrom addie.rietveld.braggtree import BraggTree, BankRegexException\n\n@pytest.fixture\ndef braggtree():\n return BraggTree(None)\n\n\ndef test_get_bank_id(qtbot, braggtree):\n \"\"\"Test we can extract a bank id from bank workspace name\"\"\"\n target = 12345\n bank_wksp_name = \"Bank {} - 90.0\".format(target)\n bank_id = braggtree._get_bank_id(bank_wksp_name)\n assert int(bank_id) == target\n\ndef test_get_bank_id_exception(qtbot, braggtree):\n \"\"\"Test for raised exception from a bad workspace name\"\"\"\n bad_ws = \"Bank jkl 1 -- 90.0\"\n with pytest.raises(BankRegexException) as e:\n braggtree._get_bank_id(bad_ws)\n\ndef test_do_plot_ws_exception(qtbot, braggtree):\n \"\"\"Test for raised exception from MainWindow==None\"\"\"\n with pytest.raises(NotImplementedError) as e:\n braggtree.do_plot_ws()\n","subject":"Refactor BraggTree test to use pytest-qt","message":"Refactor BraggTree test to use pytest-qt\n","lang":"Python","license":"mit","repos":"neutrons\/FastGR,neutrons\/FastGR,neutrons\/FastGR"} {"commit":"f0c732644a2bd110a9842423d1cccae8c3539a0e","old_file":"lbrynet\/__init__.py","new_file":"lbrynet\/__init__.py","old_contents":"import logging\n\n__version__ = \"0.20.0rc10\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","new_contents":"import logging\n\n__version__ = \"0.20.0rc11\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","subject":"Bump version 0.20.0rc10 --> 0.20.0rc11","message":"Bump version 0.20.0rc10 --> 0.20.0rc11\n\nSigned-off-by: Jack Robison <40884020c67726395ea162083a125620dc32cdab@lbry.io>\n","lang":"Python","license":"mit","repos":"lbryio\/lbry,lbryio\/lbry,lbryio\/lbry"} {"commit":"941c0d8d8f8d4a98db71ba3e79100de6d6b86294","old_file":"lbrynet\/__init__.py","new_file":"lbrynet\/__init__.py","old_contents":"import logging\n\n__version__ = \"0.14.2rc6\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","new_contents":"import logging\n\n__version__ = \"0.14.2rc7\"\nversion = tuple(__version__.split('.'))\n\nlogging.getLogger(__name__).addHandler(logging.NullHandler())\n","subject":"Bump version 0.14.2rc6 --> 0.14.2rc7","message":"Bump version 0.14.2rc6 --> 0.14.2rc7\n\nSigned-off-by: Jack Robison <40884020c67726395ea162083a125620dc32cdab@lbry.io>\n","lang":"Python","license":"mit","repos":"zestyr\/lbry,lbryio\/lbry,zestyr\/lbry,lbryio\/lbry,zestyr\/lbry,lbryio\/lbry"} {"commit":"5daa628d59576f00d0c5d49358a800dd728c6fdf","old_file":"necropsy\/models.py","new_file":"necropsy\/models.py","old_contents":"# -*- coding: utf-8 -*-\nfrom django.db import models\n\n# Create your models here.\nclass Necropsy (models.Model):\n\tclinical_information = models.TextField(null=True, blank=True)\n\tmacroscopic = models.TextField(null=True, blank=True)\n\tmicroscopic = models.TextField(null=True, blank=True)\n\tconclusion = models.TextField(null=True, blank=True)\n\tnotes = models.TextField(null=True, blank=True)\n\tfooter = models.TextField(null=True, blank=True)\n\n","new_contents":"# -*- coding: utf-8 -*-\n\nfrom django.db import models\nfrom modeling.exam import Exam\nfrom modeling.report import ReportStatus\n\n\nclass NecropsyStatus(models.Model):\n description = models.CharField(max_length=50)\n\nclass Necropsy(models.Model):\n clinical_information = models.TextField(null=True, blank=True)\n main_disease = models.TextField(null=True, blank=True)\n consequential_final_disease = models.TextField(null=True, blank=True)\n contributors_disease = models.TextField(null=True, blank=True)\n consequential_disease = models.TextField(null=True, blank=True)\n other_disases = models.TextField(null=True, blank=True)\n note = models.TextField(null=True, blank=True)\n footer = models.TextField(null=True, blank=True)\n status = models.ForeignKey(NecropsyStatus)\n exam = models.ForeignKey(Exam)\n\nclass NecropsyReport(models.Model):\n clinical_information = models.TextField(null=True, blank=True)\n main_disease = models.TextField(null=True, blank=True)\n consequential_final_disease = models.TextField(null=True, blank=True)\n contributors_disease = models.TextField(null=True, blank=True)\n consequential_disease = models.TextField(null=True, blank=True)\n other_disases = models.TextField(null=True, blank=True)\n note = models.TextField(null=True, blank=True)\n footer = models.TextField(null=True, blank=True)\n status = models.ForeignKey(ReportStatus)\n necropsy = models.ForeignKey(Necropsy)\n","subject":"Add NecropsyReport in Model Necropsy","message":"Add NecropsyReport in Model Necropsy\n","lang":"Python","license":"mit","repos":"msfernandes\/anato-hub,msfernandes\/anato-hub,msfernandes\/anato-hub,msfernandes\/anato-hub"} {"commit":"2ba4e0758c04bebcd1dcde78e99605d0b9460abf","old_file":"foldatlas\/monitor.py","new_file":"foldatlas\/monitor.py","old_contents":"import os\n\n# must call \"sudo apt-get install sendmail\" first...\n\n\n# if sts != 0:\n# print(\"Sendmail exit status \"+str(sts))\n\n\n\ndef send_error(recipient, error_details):\n\tSENDMAIL = \"\/usr\/sbin\/sendmail\" # sendmail location\n\n\tp = os.popen(\"%s -t\" % SENDMAIL, \"w\")\n\tp.write(\"To: \"+recipient+\"\\n\")\n\tp.write(\"Subject: FoldAtlas error\\n\")\n\tp.write(\"\\n\") # blank line separating headers from body\n\tp.write(\"Some text\\n\")\n\tp.write(\"some more text\\n\")\n\tsts = p.close()\n","new_contents":"import traceback\nimport os\nimport urllib.request # the lib that handles the url stuff\n\ntest_url = \"http:\/\/www.foldatlas.com\/transcript\/AT2G45180.1\"\nrecipient = \"matthew.gs.norris@gmail.com\"\nsearch_str = \"AT2G45180.1\"\n\ndef run_test():\n\ttry:\t\n\t\tdata = urllib.request.urlopen(test_url) # it's a file like object and works just like a file\n\t\ttext = str(data.read())\n\n\t\tif search_str in text:\n\t\t\tprint(\"It worked!\")\n\t\telse:\n\t\t\tsend_error(text)\n\texcept:\n\t\tsend_error(traceback.format_exc())\n\ndef send_error(error_details):\n\tprint(\"FAILED\")\n\tSENDMAIL = \"\/usr\/sbin\/sendmail\" # sendmail location\n\n\tp = os.popen(\"%s -t\" % SENDMAIL, \"w\")\n\tp.write(\"To: \"+recipient+\"\\n\")\n\tp.write(\"Subject: FoldAtlas error\\n\")\n\tp.write(\"\\n\") # blank line separating headers from body\n\tp.write(error_details)\n\tsts = p.close()\n\nrun_test()","subject":"Monitor now checks and emails","message":"Monitor now checks and emails\n","lang":"Python","license":"mit","repos":"mnori\/foldatlas,mnori\/foldatlas,mnori\/foldatlas,mnori\/foldatlas"} {"commit":"6b0774eab70c42fbdd28869b6bcdab9b81183b8e","old_file":"run_tests.py","new_file":"run_tests.py","old_contents":"#!\/usr\/bin\/env python\n\n# tests require pytest-cov and pytest-xdist\nimport os\nimport signal\nimport sys\nfrom bluesky.testing.noseclasses import KnownFailure\nimport pytest\n\ntry:\n from pcaspy import Driver, SimpleServer\n from multiprocessing import Process\n\n def to_subproc():\n\n prefix = 'BSTEST:'\n pvdb = {\n 'VAL': {\n 'prec': 3,\n },\n }\n\n class myDriver(Driver):\n def __init__(self):\n super(myDriver, self).__init__()\n\n if __name__ == '__main__':\n server = SimpleServer()\n server.createPV(prefix, pvdb)\n driver = myDriver()\n\n # process CA transactions\n while True:\n try:\n server.process(0.1)\n except KeyboardInterrupt:\n break\n\n p = Process(target=to_subproc)\nexcept ImportError:\n p = None\n\n\ndef run():\n if p is not None:\n p.start()\n try:\n args = ['--cov bluesky']\n args.extend(sys.argv)\n pytest.main(args)\n finally:\n if p is not None:\n os.kill(p.pid, signal.SIGINT)\n p.join()\n\nif __name__ == '__main__':\n run()\n","new_contents":"#!\/usr\/bin\/env python\n\n# tests require pytest-cov and pytest-xdist\nimport os\nimport signal\nimport sys\nimport pytest\n\ntry:\n from pcaspy import Driver, SimpleServer\n from multiprocessing import Process\n\n def to_subproc():\n\n prefix = 'BSTEST:'\n pvdb = {\n 'VAL': {\n 'prec': 3,\n },\n }\n\n class myDriver(Driver):\n def __init__(self):\n super(myDriver, self).__init__()\n\n if __name__ == '__main__':\n server = SimpleServer()\n server.createPV(prefix, pvdb)\n driver = myDriver()\n\n # process CA transactions\n while True:\n try:\n server.process(0.1)\n except KeyboardInterrupt:\n break\n\n p = Process(target=to_subproc)\nexcept ImportError:\n p = None\n\n\ndef run():\n if p is not None:\n p.start()\n try:\n # adding rxs to show extra info on skips and xfails\n args = ['--cov bluesky -srx -vv']\n args.extend(sys.argv)\n pytest.main(args)\n finally:\n if p is not None:\n os.kill(p.pid, signal.SIGINT)\n p.join()\n\nif __name__ == '__main__':\n run()\n","subject":"Remove deleted subpackage. Add better args to pytest","message":"TST: Remove deleted subpackage. Add better args to pytest\n","lang":"Python","license":"bsd-3-clause","repos":"ericdill\/bluesky,ericdill\/bluesky"} {"commit":"c3a184a188d18f87bad2d7f34a2dfd3a7cca4827","old_file":"signac\/common\/errors.py","new_file":"signac\/common\/errors.py","old_contents":"# Copyright (c) 2016 The Regents of the University of Michigan\n# All rights reserved.\n# This software is licensed under the BSD 3-Clause License.\n\n\nclass Error(Exception):\n pass\n\n\nclass ConfigError(Error, RuntimeError):\n pass\n\n\nclass AuthenticationError(Error, RuntimeError):\n\n def __str__(self):\n if len(self.args) > 0:\n return \"Failed to authenticate with host '{}'.\".format(\n self.args[0])\n else:\n return \"Failed to authenticate with host.\"\n\n\nclass ExportError(Error, RuntimeError):\n pass\n\n\nclass FileNotFoundError(Error, FileNotFoundError):\n pass\n\n\nclass FetchError(FileNotFoundError):\n pass\n","new_contents":"# Copyright (c) 2016 The Regents of the University of Michigan\n# All rights reserved.\n# This software is licensed under the BSD 3-Clause License.\nfrom . import six\n\nclass Error(Exception):\n pass\n\n\nclass ConfigError(Error, RuntimeError):\n pass\n\n\nclass AuthenticationError(Error, RuntimeError):\n\n def __str__(self):\n if len(self.args) > 0:\n return \"Failed to authenticate with host '{}'.\".format(\n self.args[0])\n else:\n return \"Failed to authenticate with host.\"\n\n\nclass ExportError(Error, RuntimeError):\n pass\n\n\nif six.PY2:\n class FileNotFoundError(Error, IOError):\n pass\nelse:\n class FileNotFoundError(Error, FileNotFoundError):\n pass\n\n\nclass FetchError(FileNotFoundError):\n pass\n","subject":"Fix py27 issue in error module.","message":"Fix py27 issue in error module.\n\nInherit signac internal FileNotFoundError class from IOError\ninstead of FileNotFoundError in python 2.7.\n","lang":"Python","license":"bsd-3-clause","repos":"csadorf\/signac,csadorf\/signac"} {"commit":"54e78b61db2660a57762b0f0115d532b308386e4","old_file":"opal\/tests\/test_core_commandline.py","new_file":"opal\/tests\/test_core_commandline.py","old_contents":"\"\"\"\nUnittests for opal.core.commandline\n\"\"\"\nfrom opal.core.test import OpalTestCase\n\nfrom opal.core import commandline\n","new_contents":"\"\"\"\nUnittests for opal.core.commandline\n\"\"\"\nfrom mock import patch, MagicMock\n\nfrom opal.core.test import OpalTestCase\n\nfrom opal.core import commandline\n\n\nclass StartprojectTestCase(OpalTestCase):\n\n def test_startproject(self):\n mock_args = MagicMock(name='Mock Args')\n mock_args.name = 'projectname'\n with patch.object(commandline.scaffold_utils, 'start_project') as sp:\n commandline.startproject(mock_args)\n sp.assert_called_with('projectname', commandline.USERLAND_HERE)\n\n\nclass StartpluginTestCase(OpalTestCase):\n\n def test_startplugin(self):\n mock_args = MagicMock(name='Mock Args')\n mock_args.name = 'pluginname'\n with patch.object(commandline.scaffold_utils, 'start_plugin') as sp:\n commandline.startplugin(mock_args)\n sp.assert_called_with('pluginname', commandline.USERLAND_HERE)\n","subject":"Add simple basic unittests for some of our commandline argparse target functions","message":"Add simple basic unittests for some of our commandline argparse target functions\n","lang":"Python","license":"agpl-3.0","repos":"khchine5\/opal,khchine5\/opal,khchine5\/opal"} {"commit":"c00a55b8337dbc354921c195dfa4becc7ee1346a","old_file":"ipython\/profile_default\/startup\/00-imports.py","new_file":"ipython\/profile_default\/startup\/00-imports.py","old_contents":"\"\"\"Imports for IPython\"\"\"\n\n# pylint: disable=W0611\n\n# import this\n\nimport os\nimport re\nimport sys\nimport inspect\npyprint = print\n\nmores = []\n\ntry:\n from rich.console import Console\n console = Console(color_system=\"standard\")\n print = console.print\n mores += [\"rich\"]\nexcept ImportError:\n pass\n\ntry:\n from importlib import reload\nexcept ImportError:\n def reload(x):\n raise NotImplementedError(\"importlib.reload is not available\")\n\ntry:\n import requests\n mores += [\"requests\"]\nexcept ModuleNotFoundError:\n pass\n\ntry:\n import pysyte\n from pysyte.types import paths\n from pysyte.types.paths import path\n from pysyte import cli\nexcept ImportError as e:\n print(e)\n sys.stderr.write(\"pip install pysyte # please\")\n\ntry:\n from pathlib import Path\n mores += [\"Path\"]\nexcept ImportError:\n pass\n\nmore = \", \".join([\" \"] + mores) if mores else \"\"\nexecutable = sys.executable.replace(os.environ['HOME'], '~')\nversion = sys.version.split()[0]\nstdout = lambda x: sys.stdout.write(f\"{x}\\n\")\n\nstdout(f\"import os, re, sys, inspect, pysyte, paths, path, cli{more}\")\n\nstdout(\"\")\nstdout(f\"{executable} {version}\")\n\n","new_contents":"\"\"\"Imports for IPython\"\"\"\n\n# pylint: disable=W0611\n\n# import this\n\nimport os\nimport re\nimport sys\nimport inspect\npyprint = print\n\nmores = []\n\ntry:\n from rich.console import Console\n console = Console(color_system=\"standard\")\n print = console.print\n mores += [\"rich\"]\n from rich import pretty\n pretty.install()\nexcept ImportError:\n pass\n\ntry:\n from importlib import reload\nexcept ImportError:\n def reload(x):\n raise NotImplementedError(\"importlib.reload is not available\")\n\ntry:\n import requests\n mores += [\"requests\"]\nexcept ModuleNotFoundError:\n pass\n\ntry:\n import pysyte\n from pysyte.types import paths\n from pysyte.types.paths import path\n from pysyte import cli\nexcept ImportError as e:\n print(e)\n sys.stderr.write(\"pip install pysyte # please\")\n\ntry:\n from pathlib import Path\n mores += [\"Path\"]\nexcept ImportError:\n pass\n\nmore = \", \".join([\" \"] + mores) if mores else \"\"\nexecutable = sys.executable.replace(os.environ['HOME'], '~')\nversion = sys.version.split()[0]\nstdout = lambda x: sys.stdout.write(f\"{x}\\n\")\n\nstdout(f\"import os, re, sys, inspect, pysyte, paths, path, cli{more}\")\n\nstdout(\"\")\nstdout(f\"{executable} {version}\")\n\n","subject":"Use rich for printing in ipython","message":"Use rich for printing in ipython\n","lang":"Python","license":"mit","repos":"jalanb\/jab,jalanb\/dotjab,jalanb\/dotjab,jalanb\/jab"} {"commit":"80ca0bebce22f64d0d01377493126ed95d8a64cb","old_file":"falcom\/luhn.py","new_file":"falcom\/luhn.py","old_contents":"# Copyright (c) 2017 The Regents of the University of Michigan.\n# All Rights Reserved. Licensed according to the terms of the Revised\n# BSD License. See LICENSE.txt for details.\n\ndef get_check_digit_from_checkable_int (number):\n return (9 * ((number \/\/ 10) + rotate_digit(number % 10))) % 10\n\ndef rotate_digit (digit):\n if digit > 4:\n return (digit * 2) - 9\n\n else:\n return digit * 2\n\ndef get_check_digit (number = None):\n if number:\n return get_check_digit_from_checkable_int(int(number))\n\n else:\n return None\n","new_contents":"# Copyright (c) 2017 The Regents of the University of Michigan.\n# All Rights Reserved. Licensed according to the terms of the Revised\n# BSD License. See LICENSE.txt for details.\n\ndef rotate_digit (digit):\n if digit > 4:\n return (digit * 2) - 9\n\n else:\n return digit * 2\n\ndef get_check_digit_from_checkable_int (number):\n return (9 * ((number \/\/ 10) + rotate_digit(number % 10))) % 10\n\ndef get_check_digit (number = None):\n if number:\n return get_check_digit_from_checkable_int(int(number))\n\n else:\n return None\n","subject":"Reorder methods to make sense","message":"Reorder methods to make sense\n","lang":"Python","license":"bsd-3-clause","repos":"mlibrary\/image-conversion-and-validation,mlibrary\/image-conversion-and-validation"} {"commit":"d5ee1185f0249d2e29f78866eb29552921b69ec9","old_file":"config.py","new_file":"config.py","old_contents":"import os\nimport jinja2\n\nbasedir = os.path.abspath(os.path.dirname(__file__))\n\n\nclass Config:\n @staticmethod\n def init_app(app):\n repo_root = os.path.abspath(os.path.dirname(__file__))\n template_folders = [\n os.path.join(repo_root,\n 'bower_components\/govuk_template\/views\/layouts'),\n os.path.join(repo_root, 'app\/templates')\n ]\n jinja_loader = jinja2.FileSystemLoader(template_folders)\n app.jinja_loader = jinja_loader\n\n\nclass Test(Config):\n DEBUG = True\n\n\nclass Development(Config):\n DEBUG = True,\n BASE_TEMPLATE_DATA = {\n 'asset_path': '\/static\/',\n 'header_class': 'with-proposition'\n }\n\n\nclass Live(Config):\n DEBUG = False\n\n\nconfig = {\n 'live': Live,\n 'development': Development,\n 'test': Test,\n 'default': Development\n}\n","new_contents":"import os\nimport jinja2\n\nbasedir = os.path.abspath(os.path.dirname(__file__))\n\n\nclass Config(object):\n DEBUG = False\n STATIC_URL_PATH = '\/supplier\/static'\n ASSET_PATH = STATIC_URL_PATH + '\/'\n BASE_TEMPLATE_DATA = {\n 'asset_path': ASSET_PATH,\n 'header_class': 'with-proposition'\n }\n\n @staticmethod\n def init_app(app):\n repo_root = os.path.abspath(os.path.dirname(__file__))\n template_folders = [\n os.path.join(repo_root,\n 'bower_components\/govuk_template\/views\/layouts'),\n os.path.join(repo_root, 'app\/templates')\n ]\n jinja_loader = jinja2.FileSystemLoader(template_folders)\n app.jinja_loader = jinja_loader\n\n\nclass Test(Config):\n DEBUG = True\n\n\nclass Development(Config):\n DEBUG = True,\n\n\nclass Live(Config):\n DEBUG = False\n\n\nconfig = {\n 'live': Live,\n 'development': Development,\n 'test': Test,\n 'default': Development\n}\n","subject":"Add supplier\/ prefix to static file paths","message":"Add supplier\/ prefix to static file paths\n","lang":"Python","license":"mit","repos":"mtekel\/digitalmarketplace-supplier-frontend,alphagov\/digitalmarketplace-supplier-frontend,alphagov\/digitalmarketplace-supplier-frontend,mtekel\/digitalmarketplace-supplier-frontend,alphagov\/digitalmarketplace-supplier-frontend,alphagov\/digitalmarketplace-supplier-frontend,mtekel\/digitalmarketplace-supplier-frontend,mtekel\/digitalmarketplace-supplier-frontend"} {"commit":"286dced2c23b90dba53848423d6f29873779d177","old_file":"config.py","new_file":"config.py","old_contents":"import os\n\n\nclass Config:\n SECRET_KEY = os.environ.get('SECRET_KEY')\n SQLALCHEMY_COMMIT_ON_TEARDOWN = True\n\n @staticmethod\n def init_app(app):\n pass\n\n\nclass DevelopmentConfig(Config):\n DEBUG = True\n SQLALCHEMY_DATABASE_URI = os.environ.get('DEV_DATABASE_URL')\n\n\nclass TestingConfig(Config):\n TESTING = True\n SQLALCHEMY_DATABASE_URI = os.environ.get('TEST_DATABASE_URL')\n\n\nclass ProductionConfig(Config):\n SQLALCHEMY_DATABASE_URI = os.environ.get('DATABASE_URL')\n\n @classmethod\n def init_app(cls, app):\n pass\n\n\nclass HerokuConfig(ProductionConfig):\n @classmethod\n def init_app(cls, app):\n ProductionConfig.init_app(app)\n\n # # log to stderr\n # import logging\n # from logging import StreamHandler\n # file_handler = StreamHandler()\n # file_handler.setLevel(logging.WARNING)\n # app.logger.addHandler(file_handler)\n\nconfig = {\n 'development': DevelopmentConfig,\n 'testing': TestingConfig,\n 'production': ProductionConfig,\n 'heroku': HerokuConfig,\n\n 'default': DevelopmentConfig\n}","new_contents":"import os\n\n\nclass Config:\n SECRET_KEY = os.environ.get('SECRET_KEY')\n SQLALCHEMY_COMMIT_ON_TEARDOWN = True\n\n @staticmethod\n def init_app(app):\n pass\n\n\nclass DevelopmentConfig(Config):\n DEBUG = True\n SQLALCHEMY_DATABASE_URI = os.environ.get('DEV_DATABASE_URL')\n\n\nclass TestingConfig(Config):\n TESTING = True\n\n if os.environ.get('TEST_DATABASE_URL'):\n SQLALCHEMY_DATABASE_URI = os.environ.get('TEST_DATABASE_URL')\n else:\n basedir = os.path.abspath(os.path.dirname(__file__))\n SQLALCHEMY_DATABASE_URI = 'sqlite:\/\/\/' + os.path.join(basedir, 'data-test.sqlite')\n\n\nclass ProductionConfig(Config):\n SQLALCHEMY_DATABASE_URI = os.environ.get('DATABASE_URL')\n\n @classmethod\n def init_app(cls, app):\n pass\n\n\nclass HerokuConfig(ProductionConfig):\n @classmethod\n def init_app(cls, app):\n ProductionConfig.init_app(app)\n\n # # log to stderr\n # import logging\n # from logging import StreamHandler\n # file_handler = StreamHandler()\n # file_handler.setLevel(logging.WARNING)\n # app.logger.addHandler(file_handler)\n\nconfig = {\n 'development': DevelopmentConfig,\n 'testing': TestingConfig,\n 'production': ProductionConfig,\n 'heroku': HerokuConfig,\n\n 'default': DevelopmentConfig\n}","subject":"Use sqlite as DB for test if none set in environment","message":"Use sqlite as DB for test if none set in environment\n","lang":"Python","license":"mit","repos":"boltzj\/movies-in-sf"} {"commit":"a6e46fc5429840fd3ff47c03d8b0d9f3b28c7811","old_file":"src\/sentry\/api\/endpoints\/group_events_latest.py","new_file":"src\/sentry\/api\/endpoints\/group_events_latest.py","old_contents":"from __future__ import absolute_import\n\nfrom rest_framework.response import Response\n\nfrom sentry.api import client\nfrom sentry.api.base import DocSection\nfrom sentry.api.bases.group import GroupEndpoint\n\n\nclass GroupEventsLatestEndpoint(GroupEndpoint):\n doc_section = DocSection.EVENTS\n\n def get(self, request, group):\n \"\"\"\n Retrieve the latest sample for an aggregate\n\n Return details on the latest sample for this aggregate.\n\n {method} {path}\n\n \"\"\"\n event = group.get_latest_event()\n\n try:\n return client.get('\/events\/{}\/'.format(event.id), request.user, request.auth)\n except client.ApiError as e:\n return Response(e.body, status=e.status_code)\n","new_contents":"from __future__ import absolute_import\n\nfrom rest_framework.response import Response\n\nfrom sentry.api import client\nfrom sentry.api.base import DocSection\nfrom sentry.api.bases.group import GroupEndpoint\n\n\nclass GroupEventsLatestEndpoint(GroupEndpoint):\n doc_section = DocSection.EVENTS\n\n def get(self, request, group):\n \"\"\"\n Retrieve the latest sample for an aggregate\n\n Return details on the latest sample for this aggregate.\n\n {method} {path}\n\n \"\"\"\n event = group.get_latest_event()\n if not event:\n return Response({'detail': 'No events found for group'}, status=404)\n\n try:\n return client.get('\/events\/{}\/'.format(event.id), request.user, request.auth)\n except client.ApiError as e:\n return Response(e.body, status=e.status_code)\n","subject":"Handle no latest event (fixes GH-1727)","message":"Handle no latest event (fixes GH-1727)\n","lang":"Python","license":"bsd-3-clause","repos":"imankulov\/sentry,hongliang5623\/sentry,fotinakis\/sentry,BuildingLink\/sentry,gencer\/sentry,mitsuhiko\/sentry,BuildingLink\/sentry,beeftornado\/sentry,mvaled\/sentry,daevaorn\/sentry,wong2\/sentry,ifduyue\/sentry,jean\/sentry,ifduyue\/sentry,Kryz\/sentry,jean\/sentry,kevinlondon\/sentry,fotinakis\/sentry,pauloschilling\/sentry,korealerts1\/sentry,1tush\/sentry,alexm92\/sentry,pauloschilling\/sentry,ngonzalvez\/sentry,JamesMura\/sentry,zenefits\/sentry,fuziontech\/sentry,1tush\/sentry,gencer\/sentry,gencer\/sentry,JamesMura\/sentry,jean\/sentry,BuildingLink\/sentry,mvaled\/sentry,mvaled\/sentry,korealerts1\/sentry,gencer\/sentry,felixbuenemann\/sentry,daevaorn\/sentry,wong2\/sentry,ifduyue\/sentry,wong2\/sentry,nicholasserra\/sentry,imankulov\/sentry,fuziontech\/sentry,ngonzalvez\/sentry,beeftornado\/sentry,felixbuenemann\/sentry,Kryz\/sentry,zenefits\/sentry,looker\/sentry,fotinakis\/sentry,kevinlondon\/sentry,felixbuenemann\/sentry,fuziontech\/sentry,pauloschilling\/sentry,Natim\/sentry,JackDanger\/sentry,mitsuhiko\/sentry,mvaled\/sentry,looker\/sentry,Natim\/sentry,songyi199111\/sentry,korealerts1\/sentry,BayanGroup\/sentry,zenefits\/sentry,BuildingLink\/sentry,daevaorn\/sentry,hongliang5623\/sentry,JamesMura\/sentry,ifduyue\/sentry,imankulov\/sentry,looker\/sentry,beeftornado\/sentry,nicholasserra\/sentry,alexm92\/sentry,songyi199111\/sentry,JackDanger\/sentry,hongliang5623\/sentry,JamesMura\/sentry,looker\/sentry,nicholasserra\/sentry,JamesMura\/sentry,BayanGroup\/sentry,zenefits\/sentry,ngonzalvez\/sentry,zenefits\/sentry,daevaorn\/sentry,mvaled\/sentry,jean\/sentry,kevinlondon\/sentry,Kryz\/sentry,gencer\/sentry,jean\/sentry,mvaled\/sentry,songyi199111\/sentry,fotinakis\/sentry,BuildingLink\/sentry,looker\/sentry,ifduyue\/sentry,JackDanger\/sentry,BayanGroup\/sentry,1tush\/sentry,Natim\/sentry,alexm92\/sentry"} {"commit":"666fc19e2949a30cbe40bf6020c141e84dfcae1e","old_file":"app\/soc\/models\/project_survey.py","new_file":"app\/soc\/models\/project_survey.py","old_contents":"#!\/usr\/bin\/python2.5\n#\n# Copyright 2009 the Melange authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"This module contains the ProjectSurvey model.\n\"\"\"\n\n__authors__ = [\n '\"Daniel Diniz\" <ajaksu@gmail.com>',\n '\"Lennard de Rijk\" <ljvderijk@gmail.com>',\n]\n\n\nfrom soc.models.survey import Survey\n\n\nclass ProjectSurvey(Survey):\n \"\"\"Survey for Students that have a StudentProject.\n \"\"\"\n\n def __init__(self, *args, **kwargs):\n super(ProjectSurvey, self).__init__(*args, **kwargs)\n # TODO: prefix has to be set to gsoc_program once data has been transferred\n self.prefix = 'program'\n self.taking_access = 'student'\n","new_contents":"#!\/usr\/bin\/python2.5\n#\n# Copyright 2009 the Melange authors.\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\n\"\"\"This module contains the ProjectSurvey model.\n\"\"\"\n\n__authors__ = [\n '\"Daniel Diniz\" <ajaksu@gmail.com>',\n '\"Lennard de Rijk\" <ljvderijk@gmail.com>',\n]\n\n\nfrom soc.models.survey import Survey\n\n\nclass ProjectSurvey(Survey):\n \"\"\"Survey for Students that have a StudentProject.\n \"\"\"\n\n def __init__(self, *args, **kwargs):\n super(ProjectSurvey, self).__init__(*args, **kwargs)\n self.prefix = 'gsoc_program'\n self.taking_access = 'student'\n","subject":"Set the default prefix for ProjectSurveys to gsoc_program.","message":"Set the default prefix for ProjectSurveys to gsoc_program.\n","lang":"Python","license":"apache-2.0","repos":"rhyolight\/nupic.son,rhyolight\/nupic.son,rhyolight\/nupic.son"} {"commit":"1b9aa9909b284489c9f8a5d38b1c5520d5916dc7","old_file":"feature_extraction\/measurements\/__init__.py","new_file":"feature_extraction\/measurements\/__init__.py","old_contents":"from collections import defaultdict\nfrom feature_extraction.util import DefaultAttributeDict\n\nclass Measurement(object):\n\t\"\"\"\n\tA generic feature measurement.\n\n\tAttributes\n\t----------\n\tdefault_options\n\t\tCan be set by subclasses to set default option values\n\t\"\"\"\n\n\tdefault_options = {}\n\n\tdef __init__(self, options=None):\n\t\t\"\"\"\n\t\tWhen initializing this measurement, options can be passed.\n\t\tThese are exposed to internal algorithms as `self.options`.\n\n\t\tParameters\n\t\t----------\n\t\toptions : dict\n\t\t\tA dict of options for this measurement.\n\t\t\"\"\"\n\t\tself.options = DefaultAttributeDict()\n\t\tself.options.update(self.default_options or {})\n\t\tself.options.update(options or {})\n\nfrom .pixelaverage import PixelAverage\nfrom .texture_haralick import HaralickTexture\n","new_contents":"from collections import defaultdict\nfrom feature_extraction.util import AttributeDict\n\nclass Measurement(object):\n\t\"\"\"\n\tA generic feature measurement.\n\n\tAttributes\n\t----------\n\tdefault_options\n\t\tCan be set by subclasses to set default option values\n\t\"\"\"\n\n\tdefault_options = {}\n\n\tdef __init__(self, options=None):\n\t\t\"\"\"\n\t\tWhen initializing this measurement, options can be passed.\n\t\tThese are exposed to internal algorithms as `self.options`.\n\n\t\tParameters\n\t\t----------\n\t\toptions : dict\n\t\t\tA dict of options for this measurement.\n\t\t\"\"\"\n\t\tself.options = AttributeDict()\n\t\tself.options.update(self.default_options or {})\n\t\tself.options.update(options or {})\n\nfrom .pixelaverage import PixelAverage\nfrom .texture_haralick import HaralickTexture\n","subject":"Switch back to AttributeDict for measurement options","message":"Switch back to AttributeDict for measurement options\n","lang":"Python","license":"apache-2.0","repos":"widoptimization-willett\/feature-extraction"} {"commit":"f0bca27d58fb4bc74b6627275486dbfd159954d6","old_file":"tests\/test_datafeed_fms_teams.py","new_file":"tests\/test_datafeed_fms_teams.py","old_contents":"import unittest2\nimport datetime\n\nfrom google.appengine.ext import testbed\n\nfrom datafeeds.datafeed_fms import DatafeedFms\n\n\nclass TestDatafeedFmsTeams(unittest2.TestCase):\n\n def setUp(self):\n self.testbed = testbed.Testbed()\n self.testbed.activate()\n self.testbed.init_urlfetch_stub()\n self.testbed.init_memcache_stub()\n\n self.datafeed = DatafeedFms()\n\n def tearDown(self):\n self.testbed.deactivate()\n\n def test_getFmsTeamList(self):\n teams = self.datafeed.getFmsTeamList()\n self.find177(teams)\n\n def find177(self, teams):\n found_177 = False\n for team in teams:\n if team.team_number == 177:\n found_177 = True\n self.assertEqual(team.name, \"UTC \/ Ensign Bickford Aerospace & Defense & South Windsor High School\")\n #self.assertEqual(team.address, u\"South Windsor, CT, USA\")\n self.assertEqual(team.nickname, \"Bobcat Robotics\")\n\n self.assertTrue(found_177)\n self.assertTrue(len(teams) > 0)\n","new_contents":"import unittest2\nimport datetime\n\nfrom google.appengine.ext import testbed\n\nfrom datafeeds.datafeed_fms import DatafeedFms\n\n\nclass TestDatafeedFmsTeams(unittest2.TestCase):\n\n def setUp(self):\n self.testbed = testbed.Testbed()\n self.testbed.activate()\n self.testbed.init_urlfetch_stub()\n self.testbed.init_memcache_stub()\n\n self.datafeed = DatafeedFms()\n\n def tearDown(self):\n self.testbed.deactivate()\n\n def test_getFmsTeamList(self):\n teams = self.datafeed.getFmsTeamList()\n self.find177(teams)\n\n def find177(self, teams):\n found_177 = False\n for team in teams:\n if team.team_number == 177:\n found_177 = True\n self.assertEqual(team.name, \"ClearEdge Power \/ United Technologies \/ Gain Talent \/ EBA&D & South Windsor High School\")\n #self.assertEqual(team.address, u\"South Windsor, CT, USA\")\n self.assertEqual(team.nickname, \"Bobcat Robotics\")\n\n self.assertTrue(found_177)\n self.assertTrue(len(teams) > 0)\n","subject":"Update test case for 2014","message":"Update test case for 2014\n","lang":"Python","license":"mit","repos":"tsteward\/the-blue-alliance,verycumbersome\/the-blue-alliance,1fish2\/the-blue-alliance,fangeugene\/the-blue-alliance,synth3tk\/the-blue-alliance,synth3tk\/the-blue-alliance,bvisness\/the-blue-alliance,nwalters512\/the-blue-alliance,verycumbersome\/the-blue-alliance,synth3tk\/the-blue-alliance,1fish2\/the-blue-alliance,verycumbersome\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,josephbisch\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,josephbisch\/the-blue-alliance,nwalters512\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,josephbisch\/the-blue-alliance,1fish2\/the-blue-alliance,bdaroz\/the-blue-alliance,tsteward\/the-blue-alliance,bvisness\/the-blue-alliance,synth3tk\/the-blue-alliance,bvisness\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,synth3tk\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,tsteward\/the-blue-alliance,josephbisch\/the-blue-alliance,bdaroz\/the-blue-alliance,bvisness\/the-blue-alliance,bvisness\/the-blue-alliance,fangeugene\/the-blue-alliance,tsteward\/the-blue-alliance,bdaroz\/the-blue-alliance,verycumbersome\/the-blue-alliance,nwalters512\/the-blue-alliance,bvisness\/the-blue-alliance,bdaroz\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,fangeugene\/the-blue-alliance,1fish2\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,verycumbersome\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,nwalters512\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,the-blue-alliance\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,fangeugene\/the-blue-alliance,fangeugene\/the-blue-alliance,josephbisch\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,nwalters512\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,synth3tk\/the-blue-alliance,verycumbersome\/the-blue-alliance,bdaroz\/the-blue-alliance,bdaroz\/the-blue-alliance,fangeugene\/the-blue-alliance,1fish2\/the-blue-alliance,tsteward\/the-blue-alliance,1fish2\/the-blue-alliance,josephbisch\/the-blue-alliance,jaredhasenklein\/the-blue-alliance,tsteward\/the-blue-alliance,phil-lopreiato\/the-blue-alliance,nwalters512\/the-blue-alliance"} {"commit":"c43e120319248a804328893aad34fc774c4928d3","old_file":"stdup\/kde.py","new_file":"stdup\/kde.py","old_contents":"# -*- coding: utf-8 -*-\n\n# Copyright 2013 Jacek Mitręga\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import unicode_literals\n\nimport envoy\nimport logging\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass KdeWindowManager(object):\n\n def __init__(self, workspace=6):\n self.prev_workspace = 1\n self.workspace = 6\n\n def show(self):\n envoy.run('qdbus org.kde.kwin \/KWin org.kde.KWin.setCurrentDesktop 6',\n timeout=2)\n # envoy.run('killall firefox', timeout=2)\n # envoy.connect('firefox http:\/\/standup-backend.herokuapp.com\/?room=12')\n\n def hide(self):\n envoy.run(('qdbus org.kde.kwin \/KWin '\n 'org.kde.KWin.setCurrentDesktop {}')\n .format(self.prev_workspace),\n timeout=2)\n # envoy.run('killall firefox', timeout=2)\n","new_contents":"# -*- coding: utf-8 -*-\n\n# Copyright 2013 Jacek Mitręga\n\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n\nfrom __future__ import unicode_literals\n\nimport envoy\nimport logging\n\n\nlogger = logging.getLogger(__name__)\n\n\nclass KdeWindowManager(object):\n\n def __init__(self, workspace=6):\n self.prev_workspace = 1\n self.workspace = 6\n\n def show(self):\n logger.info('kde show')\n envoy.run('qdbus org.kde.kwin \/KWin org.kde.KWin.setCurrentDesktop 6',\n timeout=2)\n # envoy.run('killall firefox', timeout=2)\n # envoy.connect('firefox http:\/\/standup-backend.herokuapp.com\/?room=12')\n\n def hide(self):\n logger.info('kde hide')\n envoy.run(('qdbus org.kde.kwin \/KWin '\n 'org.kde.KWin.setCurrentDesktop {}')\n .format(self.prev_workspace),\n timeout=2)\n # envoy.run('killall firefox', timeout=2)\n","subject":"Add KDE show & hide logging","message":"Add KDE show & hide logging\n","lang":"Python","license":"apache-2.0","repos":"waawal\/standup-desktop,waawal\/standup-desktop"} {"commit":"e4e3f0d7270e93e6123dbf05e1f51993e38d970c","old_file":"tests\/cpydiff\/types_exception_subclassinit.py","new_file":"tests\/cpydiff\/types_exception_subclassinit.py","old_contents":"\"\"\"\ncategories: Types,Exception\ndescription: Exception.__init__ raises TypeError if overridden and called by subclass\ncause: Unknown\nworkaround: Unknown\n\"\"\"\nclass A(Exception):\n def __init__(self):\n Exception.__init__(self)\n\na = A()\n","new_contents":"\"\"\"\ncategories: Types,Exception\ndescription: Exception.__init__ method does not exist.\ncause: Subclassing native classes is not fully supported in MicroPython.\nworkaround: Call using ``super()`` instead::\n\nclass A(Exception):\n def __init__(self):\n super().__init__()\n\"\"\"\nclass A(Exception):\n def __init__(self):\n Exception.__init__(self)\n\na = A()\n","subject":"Update subclassing Exception case and give work-around.","message":"tests\/cpydiff: Update subclassing Exception case and give work-around.\n","lang":"Python","license":"mit","repos":"adafruit\/micropython,pfalcon\/micropython,adafruit\/circuitpython,blazewicz\/micropython,adafruit\/circuitpython,blazewicz\/micropython,ryannathans\/micropython,ryannathans\/micropython,henriknelson\/micropython,MrSurly\/micropython,trezor\/micropython,adafruit\/micropython,adafruit\/micropython,dmazzella\/micropython,pramasoul\/micropython,tobbad\/micropython,adafruit\/circuitpython,selste\/micropython,henriknelson\/micropython,pozetroninc\/micropython,trezor\/micropython,kerneltask\/micropython,infinnovation\/micropython,ryannathans\/micropython,infinnovation\/micropython,pramasoul\/micropython,bvernoux\/micropython,dmazzella\/micropython,blazewicz\/micropython,tobbad\/micropython,pfalcon\/micropython,pozetroninc\/micropython,adafruit\/circuitpython,infinnovation\/micropython,henriknelson\/micropython,pramasoul\/micropython,pfalcon\/micropython,kerneltask\/micropython,tobbad\/micropython,adafruit\/circuitpython,MrSurly\/micropython,tobbad\/micropython,bvernoux\/micropython,adafruit\/micropython,trezor\/micropython,henriknelson\/micropython,tralamazza\/micropython,pfalcon\/micropython,pramasoul\/micropython,kerneltask\/micropython,ryannathans\/micropython,pozetroninc\/micropython,MrSurly\/micropython,infinnovation\/micropython,trezor\/micropython,dmazzella\/micropython,pozetroninc\/micropython,MrSurly\/micropython,selste\/micropython,swegener\/micropython,swegener\/micropython,tobbad\/micropython,swegener\/micropython,bvernoux\/micropython,blazewicz\/micropython,adafruit\/circuitpython,swegener\/micropython,pozetroninc\/micropython,swegener\/micropython,selste\/micropython,kerneltask\/micropython,blazewicz\/micropython,dmazzella\/micropython,tralamazza\/micropython,selste\/micropython,bvernoux\/micropython,tralamazza\/micropython,adafruit\/micropython,ryannathans\/micropython,MrSurly\/micropython,bvernoux\/micropython,selste\/micropython,infinnovation\/micropython,trezor\/micropython,pfalcon\/micropython,tralamazza\/micropython,kerneltask\/micropython,pramasoul\/micropython,henriknelson\/micropython"} {"commit":"1f7979edaa918a52702bea5de6f2bdd7a8e60796","old_file":"encryption.py","new_file":"encryption.py","old_contents":"import base64\nfrom Crypto.Cipher import AES\nfrom Crypto import Random\n\ndef encrypt(raw, key):\n raw = pad(raw)\n iv = Random.new().read(AES.block_size)\n cipher = AES.new(key, AES.MODE_CBC, iv)\n return base64.b64encode(iv + cipher.encrypt(raw))\n\ndef decrypt(enc, key):\n enc = base64.b64decode(enc)\n iv = enc[:AES.block_size]\n cipher = AES.new(key, AES.MODE_CBC, iv)\n result = unpad(cipher.decrypt(enc[AES.block_size:])).decode('utf-8')\n return result\n\ndef pad(s):\n bs = 32\n return s + (bs - len(s) % bs) * chr(bs - len(s) % bs)\n\ndef unpad(s):\n return s[:-ord(s[len(s)-1:])]\n\n","new_contents":"import base64\nfrom Crypto.Cipher import AES\nfrom Crypto import Random\n\ndef encrypt(raw, key):\n raw = pad(raw)\n iv = Random.new().read(AES.block_size)\n cipher = AES.new(key, AES.MODE_CBC, iv)\n return base64.b64encode(iv + cipher.encrypt(raw)).decode('utf-8')\n\ndef decrypt(enc, key):\n enc = base64.b64decode(enc)\n iv = enc[:AES.block_size]\n cipher = AES.new(key, AES.MODE_CBC, iv)\n result = unpad(cipher.decrypt(enc[AES.block_size:])).decode('utf-8')\n return result\n\ndef pad(s):\n bs = 32\n return s + (bs - len(s) % bs) * chr(bs - len(s) % bs)\n\ndef unpad(s):\n return s[:-ord(s[len(s)-1:])]\n\n","subject":"Add decode(utf-8) to return on encrypt","message":"Add decode(utf-8) to return on encrypt","lang":"Python","license":"mit","repos":"regexpressyourself\/passman"} {"commit":"68374c16d66cdeea9dbce620dc29d375e3009070","old_file":"bcbio\/bam\/fasta.py","new_file":"bcbio\/bam\/fasta.py","old_contents":"from Bio import SeqIO\n\ndef sequence_length(fasta):\n \"\"\"\n return a dict of the lengths of sequences in a fasta file\n \"\"\"\n file_handle = open(fasta)\n in_handle = SeqIO.parse(file_handle, \"fasta\")\n records = {record.id: len(record) for record in in_handle}\n file_handle.close()\n return records\n","new_contents":"from Bio import SeqIO\n\ndef sequence_length(fasta):\n \"\"\"\n return a dict of the lengths of sequences in a fasta file\n \"\"\"\n sequences = SeqIO.parse(fasta, \"fasta\")\n records = {record.id: len(record) for record in sequences}\n return records\n\ndef sequence_names(fasta):\n \"\"\"\n return a list of the sequence IDs in a FASTA file\n \"\"\"\n sequences = SeqIO.parse(fasta, \"fasta\")\n records = [record.id for record in sequences]\n return records\n","subject":"Add function to get list of sequence names from a FASTA file.","message":"Add function to get list of sequence names from a FASTA file.\n\nRefactor to be simpler.\n","lang":"Python","license":"mit","repos":"vladsaveliev\/bcbio-nextgen,biocyberman\/bcbio-nextgen,chapmanb\/bcbio-nextgen,chapmanb\/bcbio-nextgen,biocyberman\/bcbio-nextgen,chapmanb\/bcbio-nextgen,vladsaveliev\/bcbio-nextgen,lbeltrame\/bcbio-nextgen,vladsaveliev\/bcbio-nextgen,brainstorm\/bcbio-nextgen,lbeltrame\/bcbio-nextgen,biocyberman\/bcbio-nextgen,a113n\/bcbio-nextgen,brainstorm\/bcbio-nextgen,brainstorm\/bcbio-nextgen,a113n\/bcbio-nextgen,lbeltrame\/bcbio-nextgen,a113n\/bcbio-nextgen"} {"commit":"e6d7ec55bf00960d42b3288ae5e0e501030d5fa9","old_file":"test\/library\/gyptest-shared-obj-install-path.py","new_file":"test\/library\/gyptest-shared-obj-install-path.py","old_contents":"#!\/usr\/bin\/env python\n\n# Copyright (c) 2010 Google Inc. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\n\"\"\"\nVerifies that .so files that are order only dependencies are specified by\ntheir install location rather than by their alias.\n\"\"\"\n\nimport os\nimport TestGyp\n\ntest = TestGyp.TestGyp(formats=['make'])\n\ntest.run_gyp('shared_dependency.gyp',\n chdir='src')\ntest.relocate('src', 'relocate\/src')\n\ntest.build('shared_dependency.gyp', test.ALL, chdir='relocate\/src')\n\nwith open('relocate\/src\/Makefile') as makefile:\n make_contents = makefile.read()\n\n# If we remove the code to generate lib1, Make should still be able\n# to build lib2 since lib1.so already exists.\nmake_contents = make_contents.replace('include lib1.target.mk', '')\nwith open('relocate\/src\/Makefile', 'w') as makefile:\n makefile.write(make_contents)\n\ntest.build('shared_dependency.gyp', test.ALL, chdir='relocate\/src')\n\ntest.pass_test()\n","new_contents":"#!\/usr\/bin\/env python\n\n# Copyright (c) 2010 Google Inc. All rights reserved.\n# Use of this source code is governed by a BSD-style license that can be\n# found in the LICENSE file.\n\n\"\"\"\nVerifies that .so files that are order only dependencies are specified by\ntheir install location rather than by their alias.\n\"\"\"\n\n# Python 2.5 needs this for the with statement.\nfrom __future__ import with_statement\n\nimport os\nimport TestGyp\n\ntest = TestGyp.TestGyp(formats=['make'])\n\ntest.run_gyp('shared_dependency.gyp',\n chdir='src')\ntest.relocate('src', 'relocate\/src')\n\ntest.build('shared_dependency.gyp', test.ALL, chdir='relocate\/src')\n\nwith open('relocate\/src\/Makefile') as makefile:\n make_contents = makefile.read()\n\n# If we remove the code to generate lib1, Make should still be able\n# to build lib2 since lib1.so already exists.\nmake_contents = make_contents.replace('include lib1.target.mk', '')\nwith open('relocate\/src\/Makefile', 'w') as makefile:\n makefile.write(make_contents)\n\ntest.build('shared_dependency.gyp', test.ALL, chdir='relocate\/src')\n\ntest.pass_test()\n","subject":"Add with_statement import for python2.5.","message":"Add with_statement import for python2.5.\n\nSee http:\/\/www.python.org\/dev\/peps\/pep-0343\/ which describes\nthe with statement.\nReview URL: http:\/\/codereview.chromium.org\/5690003","lang":"Python","license":"bsd-3-clause","repos":"witwall\/gyp,witwall\/gyp,witwall\/gyp,witwall\/gyp,witwall\/gyp"} {"commit":"4e1d0ce04c762d60eedf5bd2ecdd689fb706cbc2","old_file":"anserv\/cronjobs\/__init__.py","new_file":"anserv\/cronjobs\/__init__.py","old_contents":"registered = {}\nregistered_lock = {}\nparameters = {}\n\ndef register(f=None, lock=True, params={}):\n \"\"\"Decorator to add the function to the cronjob library.\n\n @cronjobs.register\n def my_task():\n print('I can be run once\/machine at a time.')\n\n @cronjobs.register(lock=False)\n def my_task():\n print('I am concurrent friendly!')\n\n \"\"\"\n\n def decorator(f, lock=lock):\n registered[f.__name__] = f\n parameters[f.__name__] = params\n if lock:\n registered_lock[f.__name__] = f\n return f\n\n if callable(f):\n return decorator(f, lock)\n return decorator\n","new_contents":"registered = {}\nregistered_lock = {}\nparameters = {}\nfrom decorator import decorator\n\ndef register(f=None, lock=True, params={}):\n \"\"\"Decorator to add the function to the cronjob library.\n\n @cronjobs.register\n def my_task():\n print('I can be run once\/machine at a time.')\n\n @cronjobs.register(lock=False)\n def my_task():\n print('I am concurrent friendly!')\n\n \"\"\"\n\n def cron_decorator(f, lock=lock):\n registered[f.__name__] = f\n parameters[f.__name__] = params\n if lock:\n registered_lock[f.__name__] = f\n return f\n\n if callable(f):\n return decorator(cron_decorator(f, lock),f)\n return cron_decorator\n","subject":"Change decorators in cron to preserve signature","message":"Change decorators in cron to preserve signature\n","lang":"Python","license":"agpl-3.0","repos":"edx\/edxanalytics,edx\/edxanalytics,edx\/edxanalytics,edx\/insights,edx\/insights,edx\/edxanalytics"} {"commit":"c3a15b4753ecfe7436b08456da90efb7be545a50","old_file":"test\/test_exceptions.py","new_file":"test\/test_exceptions.py","old_contents":"from neomodel import StructuredNode, StringProperty, DoesNotExist, CypherException\nimport pickle\n\n\nclass Person(StructuredNode):\n name = StringProperty(unique_index=True)\n\n\ndef test_cypher_exception_can_be_displayed():\n print CypherException(\"SOME QUERY\", (), \"ERROR\", None, None)\n\ndef test_object_does_not_exist():\n try:\n Person.nodes.get(name=\"johnny\")\n except Person.DoesNotExist as e:\n pickle_instance = pickle.dumps(e)\n assert pickle_instance\n assert pickle.loads(pickle_instance)\n assert isinstance(pickle.loads(pickle_instance), DoesNotExist)\n\n\ndef test_raise_does_not_exist():\n try:\n raise DoesNotExist(\"My Test Message\")\n except DoesNotExist as e:\n pickle_instance = pickle.dumps(e)\n assert pickle_instance\n assert pickle.loads(pickle_instance)\n assert isinstance(pickle.loads(pickle_instance), DoesNotExist)\n","new_contents":"from neomodel import StructuredNode, StringProperty, DoesNotExist, CypherException\nimport pickle\n\n\nclass Person(StructuredNode):\n name = StringProperty(unique_index=True)\n\n\ndef test_cypher_exception_can_be_displayed():\n print(CypherException(\"SOME QUERY\", (), \"ERROR\", None, None))\n\ndef test_object_does_not_exist():\n try:\n Person.nodes.get(name=\"johnny\")\n except Person.DoesNotExist as e:\n pickle_instance = pickle.dumps(e)\n assert pickle_instance\n assert pickle.loads(pickle_instance)\n assert isinstance(pickle.loads(pickle_instance), DoesNotExist)\n\n\ndef test_raise_does_not_exist():\n try:\n raise DoesNotExist(\"My Test Message\")\n except DoesNotExist as e:\n pickle_instance = pickle.dumps(e)\n assert pickle_instance\n assert pickle.loads(pickle_instance)\n assert isinstance(pickle.loads(pickle_instance), DoesNotExist)\n","subject":"Fix syntax Error for compability with python 3.X","message":"Fix syntax Error for compability with python 3.X\n","lang":"Python","license":"mit","repos":"robinedwards\/neomodel,wcooley\/neomodel,robinedwards\/neomodel,pombredanne\/neomodel,fpieper\/neomodel"} {"commit":"14a4b836853909763b8961dfcdc58477607180fd","old_file":"protocols\/views.py","new_file":"protocols\/views.py","old_contents":"from django.shortcuts import render\nfrom django.conf.urls.defaults import *\nfrom django.contrib.auth.decorators import user_passes_test\n\n\nfrom .forms import ProtocolForm, TopicFormSet\n\n\ndef can_add_protocols(user):\n return user.is_authenticated() and user.has_perm('protocols.add_protocol')\n\n\n@user_passes_test(can_add_protocols)\ndef add(request):\n data = request.POST if request.POST else None\n protocol_form = ProtocolForm(data)\n topic_form = TopicFormSet(data)\n\n if protocol_form.is_valid() and topic_form.is_valid():\n protocol_form.save()\n topic_form.save()\n\n return render(request, 'protocols\/add.html', locals())\n","new_contents":"from django.shortcuts import render\nfrom django.conf.urls import *\nfrom django.contrib.auth.decorators import user_passes_test\n\n\nfrom .forms import ProtocolForm, TopicFormSet\n\n\ndef can_add_protocols(user):\n return user.is_authenticated() and user.has_perm('protocols.add_protocol')\n\n\n@user_passes_test(can_add_protocols)\ndef add(request):\n data = request.POST if request.POST else None\n protocol_form = ProtocolForm(data)\n topic_form = TopicFormSet(data)\n\n if protocol_form.is_valid() and topic_form.is_valid():\n protocol_form.save()\n topic_form.save()\n\n return render(request, 'protocols\/add.html', locals())\n","subject":"Change django.conf.urls.defaults (it is depricated)","message":"Change django.conf.urls.defaults (it is depricated)\n","lang":"Python","license":"mit","repos":"Hackfmi\/Diaphanum,Hackfmi\/Diaphanum"} {"commit":"6fa5c20f4d3b6ea9716adbf4c5fd50739f2f987e","old_file":"protractor\/test.py","new_file":"protractor\/test.py","old_contents":"# -*- coding: utf-8 -*-\n\nimport os\nimport subprocess\n\n\nclass ProtractorTestCaseMixin(object):\n protractor_conf = 'protractor.conf.js'\n suite = None\n specs = None\n\n @classmethod\n def setUpClass(cls):\n super(ProtractorTestCaseMixin, cls).setUpClass()\n with open(os.devnull, 'wb') as f:\n subprocess.call(['webdriver-manager', 'update'], stdout=f, stderr=f)\n cls.webdriver = subprocess.Popen(\n ['webdriver-manager', 'start'], stdout=f, stderr=f)\n\n @classmethod\n def tearDownClass(cls):\n cls.webdriver.kill()\n super(ProtractorTestCaseMixin, cls).tearDownClass()\n\n def test_run(self):\n protractor_command = 'protractor {}'.format(self.protractor_conf)\n if self.specs:\n protractor_command += ' --specs {}'.format(','.join(self.specs))\n if self.suite:\n protractor_command += ' --suite {}'.format(self.suite)\n protractor_command += ' --params.live_server_url={}'.format(self.live_server_url)\n return_code = subprocess.call(protractor_command.split())\n self.assertEqual(return_code, 0)\n","new_contents":"# -*- coding: utf-8 -*-\n\nimport os\nimport subprocess\n\n\nclass ProtractorTestCaseMixin(object):\n protractor_conf = 'protractor.conf.js'\n suite = None\n specs = None\n\n @classmethod\n def setUpClass(cls):\n super(ProtractorTestCaseMixin, cls).setUpClass()\n with open(os.devnull, 'wb') as f:\n subprocess.call(['webdriver-manager', 'update'], stdout=f, stderr=f)\n cls.webdriver = subprocess.Popen(\n ['webdriver-manager', 'start'], stdout=f, stderr=f)\n\n @classmethod\n def tearDownClass(cls):\n cls.webdriver.kill()\n super(ProtractorTestCaseMixin, cls).tearDownClass()\n\n def get_protractor_params(self):\n \"\"\"A hook for adding params that protractor will receive.\"\"\"\n return {\n 'live_server_url': self.live_server_url\n }\n\n def test_run(self):\n protractor_command = 'protractor {}'.format(self.protractor_conf)\n if self.specs:\n protractor_command += ' --specs {}'.format(','.join(self.specs))\n if self.suite:\n protractor_command += ' --suite {}'.format(self.suite)\n for key, value in self.get_protractor_params().iteritems():\n protractor_command += ' --params.{key}={value}'.format(\n key=key, value=value\n )\n return_code = subprocess.call(protractor_command.split())\n self.assertEqual(return_code, 0)\n","subject":"Add hook for protactor params","message":"Add hook for protactor params\n","lang":"Python","license":"mit","repos":"jpulec\/django-protractor,penguin359\/django-protractor"} {"commit":"ac850c8f9284fbe6fd8e6318431d5e4856f26c7c","old_file":"openquake\/calculators\/tests\/classical_risk_test.py","new_file":"openquake\/calculators\/tests\/classical_risk_test.py","old_contents":"import unittest\nfrom nose.plugins.attrib import attr\n\nfrom openquake.qa_tests_data.classical_risk import (\n case_1, case_2, case_3, case_4)\nfrom openquake.calculators.tests import CalculatorTestCase\n\n\nclass ClassicalRiskTestCase(CalculatorTestCase):\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_1(self):\n raise unittest.SkipTest\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_2(self):\n raise unittest.SkipTest\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_3(self):\n out = self.run_calc(case_3.__file__, 'job_haz.ini,job_risk.ini',\n exports='csv')\n [fname] = out['avg_losses-rlzs', 'csv']\n self.assertEqualFiles('expected\/rlz-000-avg_loss.csv', fname)\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_4(self):\n out = self.run_calc(case_4.__file__, 'job_haz.ini,job_risk.ini',\n exports='csv')\n fnames = out['avg_losses-rlzs', 'csv']\n self.assertEqualFiles('expected\/rlz-000-avg_loss.csv', fnames[0])\n self.assertEqualFiles('expected\/rlz-001-avg_loss.csv', fnames[1])\n","new_contents":"import unittest\nfrom nose.plugins.attrib import attr\n\nfrom openquake.qa_tests_data.classical_risk import (\n case_1, case_2, case_3, case_4)\nfrom openquake.calculators.tests import CalculatorTestCase\n\n\nclass ClassicalRiskTestCase(CalculatorTestCase):\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_1(self):\n out = self.run_calc(case_1.__file__, 'job_risk.ini', exports='xml')\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_2(self):\n raise unittest.SkipTest\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_3(self):\n out = self.run_calc(case_3.__file__, 'job_haz.ini,job_risk.ini',\n exports='csv')\n [fname] = out['avg_losses-rlzs', 'csv']\n self.assertEqualFiles('expected\/rlz-000-avg_loss.csv', fname)\n\n @attr('qa', 'risk', 'classical_risk')\n def test_case_4(self):\n out = self.run_calc(case_4.__file__, 'job_haz.ini,job_risk.ini',\n exports='csv')\n fnames = out['avg_losses-rlzs', 'csv']\n self.assertEqualFiles('expected\/rlz-000-avg_loss.csv', fnames[0])\n self.assertEqualFiles('expected\/rlz-001-avg_loss.csv', fnames[1])\n","subject":"Work on classical_risk test_case_1 and test_case_2","message":"Work on classical_risk test_case_1 and test_case_2\n","lang":"Python","license":"agpl-3.0","repos":"gem\/oq-engine,gem\/oq-engine,gem\/oq-engine,gem\/oq-engine,gem\/oq-engine"} {"commit":"9a5229fe7ae4a240d91bfae59b61c5e8dda1aa13","old_file":"bucketeer\/test\/test_commit.py","new_file":"bucketeer\/test\/test_commit.py","old_contents":"import unittest, boto, os\nfrom bucketeer import commit\n\nclass BuckeeterTest(unittest.TestCase):\n\n # Constants - TODO move to config file\n global existing_bucket, test_dir, test_file\n existing_bucket = 'bucket.exists'\n test_dir = 'bucketeer_test_dir'\n test_file = 'bucketeer_test_file'\n\n def setUp(self):\n connection = boto.connect_s3()\n\n # Create a bucket to test on existing bucket\n bucket = connection.create_bucket(existing_bucket)\n\n # Create directory to house test files\n os.makedirs(test_dir)\n\n # Create test file\n open(test_dir + '\/' + test_file, 'w').close()\n\n return\n\n def tearDown(self):\n connection = boto.connect_s3()\n\n # Remove all files uploaded to s3\n bucket = connection.get_bucket(existing_bucket)\n for s3_file in bucket.list():\n bucket.delete_key(s3_file.key)\n\n # Remove bucket created to test on existing bucket\n bucket = connection.delete_bucket(existing_bucket)\n\n # Remove test file\n os.remove(test_dir + '\/' + test_file)\n\n # Remove directory created to house test files\n os.rmdir(test_dir)\n\n return\n\n def testMain(self):\n self.assertTrue(commit)\n\n\nif __name__ == '__main__':\n unittest.main()\n","new_contents":"import unittest, boto, os\nfrom bucketeer import commit\n\nclass BuckeeterTest(unittest.TestCase):\n\n # Constants - TODO move to config file\n global existing_bucket, test_dir, test_file\n existing_bucket = 'bucket.exists'\n test_dir = 'bucketeer_test_dir'\n test_file = 'bucketeer_test_file'\n\n def setUp(self):\n connection = boto.connect_s3()\n\n # Create a bucket to test on existing bucket\n bucket = connection.create_bucket(existing_bucket)\n\n # Create directory to house test files\n os.makedirs(test_dir)\n\n # Create test file\n open(test_dir + '\/' + test_file, 'w').close()\n\n return\n\n def tearDown(self):\n connection = boto.connect_s3()\n\n # Remove all files uploaded to s3\n bucket = connection.get_bucket(existing_bucket)\n for s3_file in bucket.list():\n bucket.delete_key(s3_file.key)\n\n # Remove bucket created to test on existing bucket\n bucket = connection.delete_bucket(existing_bucket)\n\n # Remove test file\n os.remove(test_dir + '\/' + test_file)\n\n # Remove directory created to house test files\n os.rmdir(test_dir)\n\n return\n\n def testMain(self):\n self.assertTrue(commit)\n\n def testNewFileUploadExistingBucket(self):\n result = commit.commit_to_s3(existing_bucket, test_dir)\n self.assertTrue(result)\n\nif __name__ == '__main__':\n unittest.main()\n","subject":"Add test for new file upload to existing bucket","message":"Add test for new file upload to existing bucket\n","lang":"Python","license":"mit","repos":"mgarbacz\/bucketeer"} {"commit":"5da51e1820c03a76dfdb9926023848b7399691da","old_file":"inthe_am\/taskmanager\/models\/usermetadata.py","new_file":"inthe_am\/taskmanager\/models\/usermetadata.py","old_contents":"from django.conf import settings\nfrom django.contrib.auth.models import User\nfrom django.db import models\n\n\nclass UserMetadata(models.Model):\n user = models.ForeignKey(\n User, related_name=\"metadata\", unique=True, on_delete=models.CASCADE\n )\n tos_version = models.IntegerField(default=0)\n tos_accepted = models.DateTimeField(default=None, null=True,)\n privacy_policy_version = models.IntegerField(default=0)\n privacy_policy_accepted = models.DateTimeField(default=None, null=True,)\n colorscheme = models.CharField(default=\"dark-yellow-green.theme\", max_length=255,)\n\n @property\n def tos_up_to_date(self):\n return self.tos_version == settings.TOS_VERSION\n\n @property\n def privacy_policy_up_to_date(self):\n return self.privacy_policy_version == settings.PRIVACY_POLICY_VERSION\n\n @classmethod\n def get_for_user(cls, user):\n meta, created = UserMetadata.objects.get_or_create(user=user)\n return meta\n\n def save(self, *args, **kwargs):\n super().save(*args, **kwargs)\n from . import TaskStore\n\n if self.tos_up_to_date and self.privacy_policy_up_to_date:\n store = TaskStore.get_for_user(self.user)\n store.taskd_account.resume()\n\n def __str__(self):\n return self.user.username\n\n class Meta:\n app_label = \"taskmanager\"\n","new_contents":"from django.conf import settings\nfrom django.contrib.auth.models import User\nfrom django.db import models\n\n\nclass UserMetadata(models.Model):\n user = models.OneToOneField(\n User, related_name=\"metadata\", on_delete=models.CASCADE\n )\n tos_version = models.IntegerField(default=0)\n tos_accepted = models.DateTimeField(default=None, null=True,)\n privacy_policy_version = models.IntegerField(default=0)\n privacy_policy_accepted = models.DateTimeField(default=None, null=True,)\n colorscheme = models.CharField(default=\"dark-yellow-green.theme\", max_length=255,)\n\n @property\n def tos_up_to_date(self):\n return self.tos_version == settings.TOS_VERSION\n\n @property\n def privacy_policy_up_to_date(self):\n return self.privacy_policy_version == settings.PRIVACY_POLICY_VERSION\n\n @classmethod\n def get_for_user(cls, user):\n meta, created = UserMetadata.objects.get_or_create(user=user)\n return meta\n\n def save(self, *args, **kwargs):\n super().save(*args, **kwargs)\n from . import TaskStore\n\n if self.tos_up_to_date and self.privacy_policy_up_to_date:\n store = TaskStore.get_for_user(self.user)\n store.taskd_account.resume()\n\n def __str__(self):\n return self.user.username\n\n class Meta:\n app_label = \"taskmanager\"\n","subject":"Change mapping to avoid warning","message":"Change mapping to avoid warning\n","lang":"Python","license":"agpl-3.0","repos":"coddingtonbear\/inthe.am,coddingtonbear\/inthe.am,coddingtonbear\/inthe.am,coddingtonbear\/inthe.am,coddingtonbear\/inthe.am"} {"commit":"9e41b1b8d19b27cd6bd1bb81fb34c9a3adf30ad5","old_file":"entrypoint.py","new_file":"entrypoint.py","old_contents":"#!\/usr\/bin\/python3\n#\n# Define containerized environment for running Diosix on Qemu\n#\n# On Google Cloud Run: Creates HTTP server on port 8080\n# or whatever was specified using the PORT system variable.\n# Outputs via the HTTP port. This requires K_SERVICE to be set.\n#\n# On all other environments: Log to stdout\n#\n# syntax: entrypoint.py <command>\n#\n# Author: Chris Williams <diodesign@tuta.io>\n#\n\nimport os\nimport sys\n\nglobal command_result\n\nfrom flask import Flask\n\nif __name__ == \"__main__\":\n if (os.environ.get('K_SERVICE')) != '':\n print('Running HTTP service for Google Cloud')\n # app = Flask(__name__)\n # @app.route('\/')\n # def ContainerService():\n # return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\\n'\n # app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))\n else:\n print('Running locally')\n # stream = os.popen('. $HOME\/.cargo\/env && cd \/build\/diosix && {}'.format(' '.join(sys.argv[1:])))\n # output = stream.read()\n # output\n","new_contents":"#!\/usr\/bin\/python3\n#\n# Define containerized environment for running Diosix on Qemu\n#\n# On Google Cloud Run: Creates HTTP server on port 8080\n# or whatever was specified using the PORT system variable.\n# Outputs via the HTTP port. This requires K_SERVICE to be set.\n#\n# On all other environments: Log to stdout\n#\n# syntax: entrypoint.py <command>\n#\n# Author: Chris Williams <diodesign@tuta.io>\n#\n\nimport os\nimport sys\n\nglobal command_result\n\nfrom flask import Flask\n\nif __name__ == \"__main__\":\n print('{} {} {}', os.environ.get('K_SERVICE'), os.environ.get('K_REVISION'), os.environ.get('K_CONFIGURATION'))\n\n if (os.environ.get('K_SERVICE')) != '':\n print('Running HTTP service for Google Cloud')\n # app = Flask(__name__)\n # @app.route('\/')\n # def ContainerService():\n # return 'Container built. Use docker images and docker run in the Google Cloud shell to run this container.\\n'\n # app.run(debug=True,host='0.0.0.0',port=int(os.environ.get('PORT', 8080)))\n else:\n print('Running locally')\n # stream = os.popen('. $HOME\/.cargo\/env && cd \/build\/diosix && {}'.format(' '.join(sys.argv[1:])))\n # output = stream.read()\n # output\n","subject":"Debug Google Cloud Run support","message":"Debug Google Cloud Run support\n","lang":"Python","license":"mit","repos":"diodesign\/diosix"} {"commit":"95fffa0fbe744b9087547a14a97fb7dd0e68ba76","old_file":"chainer\/functions\/__init__.py","new_file":"chainer\/functions\/__init__.py","old_contents":"# Non-parameterized functions\nfrom accuracy import accuracy\nfrom basic_math import exp, log\nfrom concat import concat\nfrom copy import copy\nfrom dropout import dropout\nfrom identity import identity\nfrom leaky_relu import leaky_relu\nfrom lstm import lstm\nfrom mean_squared_error import mean_squared_error\nfrom pooling_2d import average_pooling_2d, max_pooling_2d\nfrom relu import relu\nfrom sigmoid import sigmoid\nfrom softmax import softmax\nfrom softmax_cross_entropy import softmax_cross_entropy\nfrom sum import sum\nfrom tanh import tanh\n\n# Parameterized layers\nfrom batch_normalization import BatchNormalization\nfrom convolution_2d import Convolution2D\nfrom embed_id import EmbedID\nfrom inception import Inception\nfrom linear import Linear\nfrom parameter import Parameter\nfrom prelu import PReLU\n","new_contents":"\"\"\"Collection of :class:`~chainer.Function` implementations.\"\"\"\n\n# Parameterized function classes\nfrom batch_normalization import BatchNormalization\nfrom convolution_2d import Convolution2D\nfrom embed_id import EmbedID\nfrom inception import Inception\nfrom linear import Linear\nfrom parameter import Parameter\nfrom prelu import PReLU\n\n# Array manipulation functions\nfrom concat import concat\nfrom copy import copy\nfrom dropout import dropout\nfrom identity import identity\n\n# Activation functions\nfrom basic_math import exp, log\nfrom leaky_relu import leaky_relu\nfrom lstm import lstm\nfrom relu import relu\nfrom sigmoid import sigmoid\nfrom softmax import softmax\nfrom tanh import tanh\n\n# Pooling functions\nfrom pooling_2d import average_pooling_2d, max_pooling_2d\n\n# Loss, evaluation and aggregation\nfrom accuracy import accuracy\nfrom mean_squared_error import mean_squared_error\nfrom softmax_cross_entropy import softmax_cross_entropy\nfrom sum import sum\n","subject":"Sort function imports to fit with documentation order","message":"Sort function imports to fit with documentation order\n","lang":"Python","license":"mit","repos":"kikusu\/chainer,niboshi\/chainer,ttakamura\/chainer,wkentaro\/chainer,okuta\/chainer,cupy\/cupy,ktnyt\/chainer,nushio3\/chainer,cupy\/cupy,ytoyama\/yans_chainer_hackathon,chainer\/chainer,okuta\/chainer,jnishi\/chainer,keisuke-umezawa\/chainer,kashif\/chainer,cupy\/cupy,cupy\/cupy,muupan\/chainer,chainer\/chainer,masia02\/chainer,jnishi\/chainer,niboshi\/chainer,t-abe\/chainer,ktnyt\/chainer,keisuke-umezawa\/chainer,ttakamura\/chainer,niboshi\/chainer,tkerola\/chainer,tereka114\/chainer,chainer\/chainer,jnishi\/chainer,hvy\/chainer,minhpqn\/chainer,pfnet\/chainer,hvy\/chainer,AlpacaDB\/chainer,aonotas\/chainer,wkentaro\/chainer,sou81821\/chainer,Kaisuke5\/chainer,wavelets\/chainer,muupan\/chainer,ysekky\/chainer,benob\/chainer,okuta\/chainer,ikasumi\/chainer,sinhrks\/chainer,wkentaro\/chainer,kuwa32\/chainer,ktnyt\/chainer,woodshop\/complex-chainer,laysakura\/chainer,jnishi\/chainer,yanweifu\/chainer,t-abe\/chainer,tscohen\/chainer,keisuke-umezawa\/chainer,benob\/chainer,keisuke-umezawa\/chainer,anaruse\/chainer,umitanuki\/chainer,kikusu\/chainer,wkentaro\/chainer,hvy\/chainer,tigerneil\/chainer,nushio3\/chainer,truongdq\/chainer,chainer\/chainer,niboshi\/chainer,kiyukuta\/chainer,woodshop\/chainer,AlpacaDB\/chainer,ronekko\/chainer,rezoo\/chainer,sinhrks\/chainer,cemoody\/chainer,elviswf\/chainer,hvy\/chainer,okuta\/chainer,1986ks\/chainer,ktnyt\/chainer,hidenori-t\/chainer,jfsantos\/chainer,bayerj\/chainer,truongdq\/chainer,delta2323\/chainer"} {"commit":"896f402c79dd3bbe7d5cbc6e59787398a96b3747","old_file":"runtests.py","new_file":"runtests.py","old_contents":"import argparse\nimport os\nimport sys\n\n# Force this to happen before loading django\ntry:\n os.environ[\"DJANGO_SETTINGS_MODULE\"] = \"testtinymce.settings\"\n test_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n sys.path.insert(0, test_dir)\nexcept ImportError:\n pass\nelse:\n import django\n from django.conf import settings\n from django.test.utils import get_runner\n\n\ndef runtests(verbosity=1, failfast=False):\n django.setup()\n TestRunner = get_runner(settings)\n test_runner = TestRunner(interactive=True, verbosity=verbosity, failfast=failfast)\n failures = test_runner.run_tests([\"tinymce\"])\n sys.exit(bool(failures))\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description=\"Run the django-tinymce test suite.\")\n parser.add_argument(\n \"-v\",\n \"--verbosity\",\n default=1,\n type=int,\n choices=[0, 1, 2, 3],\n help=\"Verbosity level; 0=minimal output, 1=normal output, 2=all output\",\n )\n parser.add_argument(\n \"--failfast\",\n action=\"store_true\",\n help=\"Stop running the test suite after first failed test.\",\n )\n options = parser.parse_args()\n runtests(verbosity=options.verbosity, failfast=options.failfast)\n","new_contents":"import argparse\nimport os\nimport sys\n\n# Force this to happen before loading django\ntry:\n os.environ[\"DJANGO_SETTINGS_MODULE\"] = \"testtinymce.settings\"\n test_dir = os.path.dirname(os.path.dirname(os.path.abspath(__file__)))\n sys.path.insert(0, test_dir)\nexcept ImportError:\n pass\nelse:\n import django\n from django.conf import settings\n from django.test.utils import get_runner\n\n\ndef runtests(modules=[\"tinymce\"], verbosity=1, failfast=False):\n django.setup()\n TestRunner = get_runner(settings)\n test_runner = TestRunner(interactive=True, verbosity=verbosity, failfast=failfast)\n failures = test_runner.run_tests(modules)\n sys.exit(bool(failures))\n\n\nif __name__ == \"__main__\":\n parser = argparse.ArgumentParser(description=\"Run the django-tinymce test suite.\")\n parser.add_argument(\n \"modules\",\n nargs=\"*\",\n metavar=\"module\",\n help='Optional path(s) to test modules; e.g. \"tinymce\" or '\n '\"tinymce.tests.test_widgets\".',\n )\n parser.add_argument(\n \"-v\",\n \"--verbosity\",\n default=1,\n type=int,\n choices=[0, 1, 2, 3],\n help=\"Verbosity level; 0=minimal output, 1=normal output, 2=all output\",\n )\n parser.add_argument(\n \"--failfast\",\n action=\"store_true\",\n help=\"Stop running the test suite after first failed test.\",\n )\n options = parser.parse_args()\n runtests(modules=options.modules, verbosity=options.verbosity, failfast=options.failfast)\n","subject":"Add ability to run subset of tests","message":"Add ability to run subset of tests\n","lang":"Python","license":"mit","repos":"aljosa\/django-tinymce,aljosa\/django-tinymce,aljosa\/django-tinymce,aljosa\/django-tinymce"} {"commit":"5e2697b55f1720c4c144840e680004fb28a3cfcc","old_file":"runtests.py","new_file":"runtests.py","old_contents":"#!\/usr\/bin\/env python\nimport sys\n\nfrom django.conf import settings\n\n\nif not settings.configured:\n settings.configure(\n DATABASES={\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': ':memory:',\n }\n },\n INSTALLED_APPS=(\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'td_biblio',\n ),\n ROOT_URLCONF='td_biblio.urls',\n SITE_ID=1,\n SECRET_KEY='this-is-just-for-tests-so-not-that-secret',\n )\n\n\nfrom django.test.utils import get_runner\n\n\ndef runtests():\n TestRunner = get_runner(settings)\n test_runner = TestRunner(verbosity=1, interactive=True, failfast=False)\n failures = test_runner.run_tests(['td_biblio', ])\n sys.exit(failures)\n\n\nif __name__ == '__main__':\n runtests()\n","new_contents":"#!\/usr\/bin\/env python\nimport sys\n\nfrom django.conf import settings\n\n\nif not settings.configured:\n settings.configure(\n DATABASES={\n 'default': {\n 'ENGINE': 'django.db.backends.sqlite3',\n 'NAME': ':memory:',\n }\n },\n INSTALLED_APPS=(\n 'django.contrib.auth',\n 'django.contrib.contenttypes',\n 'td_biblio',\n ),\n ROOT_URLCONF='td_biblio.urls',\n SITE_ID=1,\n SECRET_KEY='this-is-just-for-tests-so-not-that-secret',\n )\n\n\nfrom django.test.utils import get_runner\n\n\ndef runtests():\n TestRunner = get_runner(settings)\n test_runner = TestRunner(verbosity=1, interactive=True, failfast=False)\n failures = test_runner.run_tests([\n 'td_biblio.tests.test_commands',\n 'td_biblio.tests.test_factories',\n 'td_biblio.tests.test_models',\n 'td_biblio.tests.test_views',\n ])\n sys.exit(failures)\n\n\nif __name__ == '__main__':\n runtests()\n","subject":"Add more flexibity to run tests independantly","message":"Add more flexibity to run tests independantly\n","lang":"Python","license":"mit","repos":"TailorDev\/django-tailordev-biblio,TailorDev\/django-tailordev-biblio,TailorDev\/django-tailordev-biblio,TailorDev\/django-tailordev-biblio"} {"commit":"946220075802cc59f3b34d3557c0b749c526c4b1","old_file":"runtests.py","new_file":"runtests.py","old_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport sys\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"farnsworth.settings\")\nthis_dir = os.path.abspath(os.path.dirname(__file__))\nif this_dir not in sys.path:\n\tsys.path.insert(0, this_dir)\n\nfrom django.test.utils import get_runner\nfrom django.conf import settings\n\ndef runtests():\n\tTestRunner = get_runner(settings)\n\ttest_runner = TestRunner(verbosity=1, interactive=True, failfast=False)\n\tfailures = test_runner.run_tests([\"base\", \"threads\", \"events\", \"managers\"])\n\tsys.exit(bool(failures))\n\nif __name__ == \"__main__\":\n\truntests()\n","new_contents":"#!\/usr\/bin\/env python\n\nimport os\nimport sys\n\nos.environ.setdefault(\"DJANGO_SETTINGS_MODULE\", \"farnsworth.settings\")\nthis_dir = os.path.abspath(os.path.dirname(__file__))\nif this_dir not in sys.path:\n\tsys.path.insert(0, this_dir)\n\nfrom django.test.utils import get_runner\nfrom django.conf import settings\n\ndef runtests():\n\tTestRunner = get_runner(settings)\n\ttest_runner = TestRunner(verbosity=1, interactive=True, failfast=False)\n\tfailures = test_runner.run_tests([\n \"base\",\n \"threads\",\n \"events\",\n \"managers\",\n \"workshift\",\n ])\n\tsys.exit(bool(failures))\n\nif __name__ == \"__main__\":\n\truntests()\n","subject":"Add workshift to the list of tests","message":"Add workshift to the list of tests\n","lang":"Python","license":"bsd-2-clause","repos":"knagra\/farnsworth,knagra\/farnsworth,knagra\/farnsworth,knagra\/farnsworth"} {"commit":"45e758b56370f5bb34ff28c4660837fd9037b945","old_file":"dom\/automation\/detect_malloc_errors.py","new_file":"dom\/automation\/detect_malloc_errors.py","old_contents":"#!\/usr\/bin\/env python\n\n# Look for \"szone_error\" (Tiger), \"malloc_error_break\" (Leopard), \"MallocHelp\" (?)\n# which are signs of malloc being unhappy (double free, out-of-memory, etc).\n\ndef amiss(logPrefix):\n foundSomething = False\n\n currentFile = file(logPrefix + \"-err\", \"r\")\n \n pline = \"\"\n ppline = \"\"\n\n for line in currentFile:\n line = line.strip(\"\\x07\").rstrip(\"\\n\")\n \n if (-1 != line.find(\"szone_error\")\n or -1 != line.find(\"malloc_error_break\")\n or -1 != line.find(\"MallocHelp\")):\n if (-1 != pline.find(\"can't allocate region\")):\n print \"\"\n print ppline\n print pline\n print line\n foundSomething = True\n break # Don't flood the log with repeated malloc failures\n\n ppline = pline\n pline = line\n\n currentFile.close()\n \n return foundSomething\n","new_contents":"#!\/usr\/bin\/env python\n\n# Look for \"szone_error\" (Tiger), \"malloc_error_break\" (Leopard), \"MallocHelp\" (?)\n# which are signs of malloc being unhappy (double free, out-of-memory, etc).\n\ndef amiss(logPrefix):\n foundSomething = False\n\n currentFile = file(logPrefix + \"-err\", \"r\")\n \n pline = \"\"\n ppline = \"\"\n\n for line in currentFile:\n line = line.strip(\"\\x07\").rstrip(\"\\n\")\n \n if (-1 != line.find(\"szone_error\")\n or -1 != line.find(\"malloc_error_break\")\n or -1 != line.find(\"MallocHelp\")):\n if (-1 == pline.find(\"can't allocate region\")):\n print \"\"\n print ppline\n print pline\n print line\n foundSomething = True\n break # Don't flood the log with repeated malloc failures\n\n ppline = pline\n pline = line\n\n currentFile.close()\n \n return foundSomething\n","subject":"Fix reversed condition for ignoring \"can't allocate region\" errors","message":"Fix reversed condition for ignoring \"can't allocate region\" errors\n","lang":"Python","license":"mpl-2.0","repos":"MozillaSecurity\/funfuzz,nth10sd\/funfuzz,nth10sd\/funfuzz,MozillaSecurity\/funfuzz,MozillaSecurity\/funfuzz,nth10sd\/funfuzz"} {"commit":"94788bd7a7ba0a7799689c4613a2afbcc377649b","old_file":"games\/migrations\/0016_auto_20161209_1256.py","new_file":"games\/migrations\/0016_auto_20161209_1256.py","old_contents":"# -*- coding: utf-8 -*-\n# Generated by Django 1.10.2 on 2016-12-09 11:56\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations\nfrom django.core.management import call_command\n\n\ndef create_revisions(apps, schema_editor):\n call_command('createinitialrevisions')\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('games', '0015_installer_draft'),\n ]\n\n operations = [\n migrations.RunPython(create_revisions)\n ]\n","new_contents":"# -*- coding: utf-8 -*-\n# Generated by Django 1.10.2 on 2016-12-09 11:56\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations\nfrom django.core.management import call_command\n\n\ndef create_revisions(apps, schema_editor):\n call_command('createinitialrevisions')\n\n\nclass Migration(migrations.Migration):\n\n dependencies = [\n ('games', '0015_installer_draft'),\n ('reversion', '0001_squashed_0004_auto_20160611_1202'),\n ]\n\n operations = [\n migrations.RunPython(create_revisions)\n ]\n","subject":"Add dependency to reversion data migration","message":"Add dependency to reversion data migration\n","lang":"Python","license":"agpl-3.0","repos":"Turupawn\/website,Turupawn\/website,lutris\/website,lutris\/website,lutris\/website,Turupawn\/website,Turupawn\/website,lutris\/website"} {"commit":"4c3a2a61c6a8cb5e0ece14bced4ec8b33df45400","old_file":"tests\/simple\/_util.py","new_file":"tests\/simple\/_util.py","old_contents":"#######################################################\n# Copyright (c) 2015, ArrayFire\n# All rights reserved.\n#\n# This file is distributed under 3-clause BSD license.\n# The complete license agreement can be obtained at:\n# http:\/\/arrayfire.com\/licenses\/BSD-3-Clause\n########################################################\n\nimport arrayfire as af\n\ndef display_func(verbose):\n if (verbose):\n return af.display\n else:\n def eval_func(foo):\n res = foo\n return eval_func\n\ndef print_func(verbose):\n def print_func_impl(*args):\n if (verbose):\n print(args)\n else:\n res = [args]\n return print_func_impl\n\nclass _simple_test_dict(dict):\n\n def __init__(self):\n self.print_str = \"Simple %16s: %s\"\n super(_simple_test_dict, self).__init__()\n\n def run(self, name_list=None, verbose=False):\n test_list = name_list if name_list is not None else self.keys()\n for key in test_list:\n\n try:\n test = self[key]\n except:\n print(self.print_str % (key, \"NOTFOUND\"))\n continue\n\n try:\n test(verbose)\n print(self.print_str % (key, \"PASSED\"))\n except:\n print(self.print_str % (key, \"FAILED\"))\n\ntests = _simple_test_dict()\n","new_contents":"#######################################################\n# Copyright (c) 2015, ArrayFire\n# All rights reserved.\n#\n# This file is distributed under 3-clause BSD license.\n# The complete license agreement can be obtained at:\n# http:\/\/arrayfire.com\/licenses\/BSD-3-Clause\n########################################################\n\nimport traceback\nimport logging\nimport arrayfire as af\n\ndef display_func(verbose):\n if (verbose):\n return af.display\n else:\n def eval_func(foo):\n res = foo\n return eval_func\n\ndef print_func(verbose):\n def print_func_impl(*args):\n if (verbose):\n print(args)\n else:\n res = [args]\n return print_func_impl\n\nclass _simple_test_dict(dict):\n\n def __init__(self):\n self.print_str = \"Simple %16s: %s\"\n super(_simple_test_dict, self).__init__()\n\n def run(self, name_list=None, verbose=False):\n test_list = name_list if name_list is not None else self.keys()\n for key in test_list:\n try:\n test = self[key]\n except:\n print(self.print_str % (key, \"NOTFOUND\"))\n continue\n\n try:\n test(verbose)\n print(self.print_str % (key, \"PASSED\"))\n except Exception as e:\n print(self.print_str % (key, \"FAILED\"))\n if (verbose):\n logging.error(traceback.format_exc())\n\n\ntests = _simple_test_dict()\n","subject":"Add proper logging to tests when in verbose mode","message":"Add proper logging to tests when in verbose mode\n","lang":"Python","license":"bsd-3-clause","repos":"arrayfire\/arrayfire_python,pavanky\/arrayfire-python,arrayfire\/arrayfire-python"} {"commit":"361af42be2c3044a15480572befb1405a603b4ab","old_file":"VALDprepare.py","new_file":"VALDprepare.py","old_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf8 -*-\n\n# My imports\nimport argparse\nimport gzip\n\n\ndef _parser():\n parser = argparse.ArgumentParser(description='Prepare the data downloaded '\n 'from VALD.')\n parser.add_argument('input', help='input compressed file')\n parser.add_argument('-o', '--output',\n help='Optional output',\n default=False)\n return parser.parse_args()\n\n\ndef main(input, output=False):\n if not isinstance(input, str):\n raise TypeError('Input must be a str. A %s was parsed' % type(input))\n if not isinstance(output, str) and output:\n raise TypeError('Output must be a str. A %s was parsed' % type(output))\n\n # TODO: Check if the input exists\n\n fname = input.rpartition('.')[0]\n if not output:\n output = '%s.dat' % fname\n oref = '%s.ref' % fname\n\n fout = ''\n fref = ''\n with gzip.open(input, 'r') as lines:\n for i, line in enumerate(lines):\n if i < 2:\n fout += '# %s' % line.replace(\"'\", '')\n else:\n fout += line.replace(\"'\", '')\n if 'References' in line:\n break\n\n with open(output, 'w') as fo:\n fo.write(fout)\n\n\nif __name__ == '__main__':\n args = _parser()\n input, output = args.input, args.output\n main(input, output)\n","new_contents":"#!\/usr\/bin\/env python\n# -*- coding: utf8 -*-\n\n# My imports\nimport argparse\nimport gzip\nimport os\n\n\ndef _parser():\n parser = argparse.ArgumentParser(description='Prepare the data downloaded '\n 'from VALD.')\n parser.add_argument('input', help='input compressed file', type=str)\n parser.add_argument('-o', '--output',\n help='Optional output',\n default=False, type=str)\n return parser.parse_args()\n\n\ndef main(input, output=False):\n\n if not os.path.isfile(input):\n raise IOError('File: %s does not exists' % input)\n\n fname = input.rpartition('.')[0]\n if not output:\n output = '%s.dat' % fname\n oref = '%s.ref' % fname\n\n fout = ''\n fref = ''\n with gzip.open(input, 'r') as lines:\n for i, line in enumerate(lines):\n if i < 2:\n fout += '# %s' % line.replace(\"'\", '')\n else:\n fout += line.replace(\"'\", '')\n if 'References' in line:\n break\n\n with open(output, 'w') as fo:\n fo.write(fout)\n\n\nif __name__ == '__main__':\n args = _parser()\n input, output = args.input, args.output\n main(input, output)\n","subject":"Check if the file exists before doing anything else.","message":"Check if the file exists before doing anything else.\n","lang":"Python","license":"mit","repos":"DanielAndreasen\/astro_scripts"} {"commit":"62f681803401d05fd0a5e554d4d6c7210dcc7c17","old_file":"cbv\/management\/commands\/load_all_django_versions.py","new_file":"cbv\/management\/commands\/load_all_django_versions.py","old_contents":"import os\nimport re\n\nfrom django.conf import settings\nfrom django.core.management import call_command, BaseCommand\n\n\nclass Command(BaseCommand):\n \"\"\"Load the Django project fixtures and all version fixtures\"\"\"\n\n def handle(self, **options):\n fixtures_dir = os.path.join(settings.DIRNAME, 'cbv', 'fixtures')\n self.stdout.write('Loading project.json')\n call_command('loaddata', 'cbv\/fixtures\/project.json')\n version_fixtures = [re.match(r'((?:\\d+\\.){2,3}json)', filename) for filename in os.listdir(fixtures_dir)]\n for match in version_fixtures:\n try:\n fixture = match.group()\n except AttributeError:\n continue\n self.stdout.write('Loading {}'.format(fixture))\n call_command('loaddata', 'cbv\/fixtures\/{}'.format(fixture))\n","new_contents":"import glob\nimport os\n\nfrom django.core.management import call_command, BaseCommand\n\n\nclass Command(BaseCommand):\n \"\"\"Load the Django project fixtures and all version fixtures\"\"\"\n\n def handle(self, **options):\n self.stdout.write('Loading project.json')\n call_command('loaddata', 'cbv\/fixtures\/project.json')\n version_fixtures = glob.glob(os.path.join('cbv', 'fixtures', '*.*.*json'))\n for fixture in version_fixtures:\n self.stdout.write('Loading {}'.format(fixture))\n call_command('loaddata', fixture)\n","subject":"Use glob for finding version fixtures","message":"Use glob for finding version fixtures\n\nThanks @ghickman!\n","lang":"Python","license":"bsd-2-clause","repos":"refreshoxford\/django-cbv-inspector,refreshoxford\/django-cbv-inspector,refreshoxford\/django-cbv-inspector,refreshoxford\/django-cbv-inspector"} {"commit":"b6416ba4c32aaeddb567be4486854d6415c3048e","old_file":"tornwamp\/customize.py","new_file":"tornwamp\/customize.py","old_contents":"\"\"\"\nTornWAMP user-configurable structures.\n\"\"\"\nfrom tornwamp.processors import GoodbyeProcessor, HelloProcessor, pubsub, rpc\nfrom tornwamp.messages import Code\n\nprocessors = {\n Code.HELLO: HelloProcessor,\n Code.GOODBYE: GoodbyeProcessor,\n Code.SUBSCRIBE: pubsub.SubscribeProcessor,\n Code.CALL: rpc.CallProcessor\n}\n# 2: 'welcome',\n# 3: 'abort',\n# 4: 'challenge',\n# 5: 'authenticate',\n# 7: 'heartbeat',\n# 8: 'error',\n# 16: 'publish',\n# 17: 'published',\n# 32: 'subscribe',\n# 33: 'subscribed',\n# 34: 'unsubscribe',\n# 35: 'unsubscribed',\n# 36: 'event',\n# 49: 'cancel',\n# 50: 'result',\n# 64: 'register',\n# 65: 'registered',\n# 66: 'unregister',\n# 67: 'unregistered',\n# 68: 'invocation',\n# 69: 'interrupt',\n# 70: 'yield'\n","new_contents":"\"\"\"\nTornWAMP user-configurable structures.\n\"\"\"\nfrom tornwamp.processors import GoodbyeProcessor, HelloProcessor, pubsub, rpc\nfrom tornwamp.messages import Code\n\nprocessors = {\n Code.HELLO: HelloProcessor,\n Code.GOODBYE: GoodbyeProcessor,\n Code.SUBSCRIBE: pubsub.SubscribeProcessor,\n Code.CALL: rpc.CallProcessor,\n Code.PUBLISH: pubsub.PublishProcessor\n}\n# 2: 'welcome',\n# 3: 'abort',\n# 4: 'challenge',\n# 5: 'authenticate',\n# 7: 'heartbeat',\n# 8: 'error',\n# 16: 'publish',\n# 17: 'published',\n# 32: 'subscribe',\n# 33: 'subscribed',\n# 34: 'unsubscribe',\n# 35: 'unsubscribed',\n# 36: 'event',\n# 49: 'cancel',\n# 50: 'result',\n# 64: 'register',\n# 65: 'registered',\n# 66: 'unregister',\n# 67: 'unregistered',\n# 68: 'invocation',\n# 69: 'interrupt',\n# 70: 'yield'\n","subject":"Add PublishProcessor to processors' list","message":"Add PublishProcessor to processors' list\n","lang":"Python","license":"apache-2.0","repos":"ef-ctx\/tornwamp"} {"commit":"369964986df0ca558c2e340bc8d15272296af67e","old_file":"tools\/debug_launcher.py","new_file":"tools\/debug_launcher.py","old_contents":"from __future__ import print_function\nimport sys\nimport os\nimport time\nimport socket\nimport argparse\nimport subprocess\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--launch-adapter')\nparser.add_argument('--lldb')\nparser.add_argument('--wait-port')\n\nargs = parser.parse_args()\n\nif args.launch_adapter:\n lldb = args.lldb or 'lldb'\n cmd = [lldb, '-b',\n '-O', 'command script import %s' % args.launch_adapter,\n '-O', 'script import ptvsd; ptvsd.enable_attach(address=(\"0.0.0.0\", 3000)); ptvsd.wait_for_attach(); adapter.run_tcp_session(4711)',\n ]\n print('Launching', cmd)\n subprocess.Popen(cmd, preexec_fn=lambda: os.setsid())\n\nif args.wait_port:\n port = int(args.wait_port)\n print('Waiting for port %d' % port)\n\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n while True:\n result = sock.connect_ex(('127.0.0.1', port))\n if result == 0:\n break\n time.sleep(0.5)\n\n print('Port opened')\n sock.shutdown(socket.SHUT_WR)\n sock.close()\n","new_contents":"from __future__ import print_function\nimport sys\nimport os\nimport time\nimport socket\nimport argparse\nimport subprocess\n\nparser = argparse.ArgumentParser()\nparser.add_argument('--launch-adapter')\nparser.add_argument('--lldb')\nparser.add_argument('--wait-port')\n\nargs = parser.parse_args()\n\nif args.launch_adapter:\n lldb = args.lldb or 'lldb'\n cmd = [lldb, '-b',\n '-O', 'command script import %s' % args.launch_adapter,\n '-O', 'script sys.argv=[\"lldb\"]; import ptvsd; ptvsd.enable_attach(address=(\"0.0.0.0\", 3000)); ptvsd.wait_for_attach()',\n '-O', 'script adapter.run_tcp_session(4711)',\n ]\n print('Launching', cmd)\n if sys.platform != 'win32':\n subprocess.Popen(cmd, preexec_fn=lambda: os.setsid())\n else:\n subprocess.Popen(cmd, creationflags=subprocess.CREATE_NEW_CONSOLE)\n\nif args.wait_port:\n port = int(args.wait_port)\n print('Waiting for port %d' % port)\n\n sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)\n while True:\n result = sock.connect_ex(('127.0.0.1', port))\n if result == 0:\n break\n time.sleep(0.5)\n\n print('Port opened')\n sock.shutdown(socket.SHUT_WR)\n sock.close()\n","subject":"Fix python debugging on Windows.","message":"Fix python debugging on Windows.\n","lang":"Python","license":"mit","repos":"vadimcn\/vscode-lldb,vadimcn\/vscode-lldb,vadimcn\/vscode-lldb,vadimcn\/vscode-lldb,vadimcn\/vscode-lldb,vadimcn\/vscode-lldb,vadimcn\/vscode-lldb"} {"commit":"d4aa2b1a0a72696ce34f5aa2f5e588fc3a72e622","old_file":"cfgrib\/__main__.py","new_file":"cfgrib\/__main__.py","old_contents":"\nimport argparse\nimport sys\n\nfrom . import eccodes\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('--selfcheck', default=False, action='store_true')\n args = parser.parse_args()\n if args.selfcheck:\n eccodes.codes_get_api_version()\n print(\"Your system is ready.\")\n else:\n raise RuntimeError(\"Command not recognised. See usage with --help.\")\n\n\nif __name__ == '__main__':\n main()\n","new_contents":"#\n# Copyright 2017-2018 European Centre for Medium-Range Weather Forecasts (ECMWF).\n#\n# Licensed under the Apache License, Version 2.0 (the \"License\");\n# you may not use this file except in compliance with the License.\n# You may obtain a copy of the License at\n#\n# http:\/\/www.apache.org\/licenses\/LICENSE-2.0\n#\n# Unless required by applicable law or agreed to in writing, software\n# distributed under the License is distributed on an \"AS IS\" BASIS,\n# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.\n# See the License for the specific language governing permissions and\n# limitations under the License.\n#\n# Authors:\n# Alessandro Amici - B-Open - https:\/\/bopen.eu\n#\n\nimport argparse\n\nfrom . import eccodes\n\n\ndef main():\n parser = argparse.ArgumentParser()\n parser.add_argument('--selfcheck', default=False, action='store_true')\n args = parser.parse_args()\n if args.selfcheck:\n eccodes.codes_get_api_version()\n print(\"Your system is ready.\")\n else:\n raise RuntimeError(\"Command not recognised. See usage with --help.\")\n\n\nif __name__ == '__main__':\n main()\n","subject":"Add copyright noticeand Authors comment.","message":"Add copyright noticeand Authors comment.\n","lang":"Python","license":"apache-2.0","repos":"ecmwf\/cfgrib"} {"commit":"9fec06c6acf57b4d49b9c49b7e1d3b5c90e2c9c4","old_file":"blog\/admin.py","new_file":"blog\/admin.py","old_contents":"from django.contrib import admin\n\nfrom .models import Post\n\n\n@admin.register(Post)\nclass PostAdmin(admin.ModelAdmin):\n # list view\n date_hierarchy = 'pub_date'\n list_display = ('title', 'pub_date')\n list_filter = ('pub_date',)\n search_fields = ('title', 'text')\n # form view\n fieldsets = (\n (None, {\n 'fields': (\n 'title', 'slug', 'author', 'text',\n )}),\n ('Related', {\n 'fields': (\n 'tags', 'startups')}),\n )\n filter_horizontal = ('startups',)\n filter_vertical = ('tags',)\n prepopulated_fields = {\"slug\": (\"title\",)}\n","new_contents":"from django.contrib import admin\n\nfrom .models import Post\n\n\n@admin.register(Post)\nclass PostAdmin(admin.ModelAdmin):\n # list view\n date_hierarchy = 'pub_date'\n list_display = ('title', 'pub_date')\n list_filter = ('pub_date',)\n search_fields = ('title', 'text')\n # form view\n fieldsets = (\n (None, {\n 'fields': (\n 'title', 'slug', 'author', 'text',\n )}),\n ('Related', {\n 'fields': (\n 'tags', 'startups')}),\n )\n filter_horizontal = ('tags', 'startups',)\n prepopulated_fields = {\"slug\": (\"title\",)}\n","subject":"Use horizontal filter for M2M in PostAdmin.","message":"Ch23: Use horizontal filter for M2M in PostAdmin.\n","lang":"Python","license":"bsd-2-clause","repos":"jambonrose\/DjangoUnleashed-1.8,jambonrose\/DjangoUnleashed-1.8"} {"commit":"b0273cc12abaf9a3f9f2e6c534d82bd7581c240e","old_file":"ctypeslib\/test\/test_dynmodule.py","new_file":"ctypeslib\/test\/test_dynmodule.py","old_contents":"# Basic test of dynamic code generation\nimport unittest\n\nimport stdio\nfrom ctypes import POINTER, c_int\n\nclass DynModTest(unittest.TestCase):\n def test_fopen(self):\n self.failUnlessEqual(stdio.fopen.restype, POINTER(stdio.FILE))\n self.failUnlessEqual(stdio.fopen.argtypes, [stdio.STRING, stdio.STRING])\n\n def test_constants(self):\n self.failUnlessEqual(stdio.O_RDONLY, 0)\n self.failUnlessEqual(stdio.O_WRONLY, 1)\n self.failUnlessEqual(stdio.O_RDWR, 2)\n\nif __name__ == \"__main__\":\n unittest.main()\n","new_contents":"# Basic test of dynamic code generation\nimport unittest\nimport os, glob\n\nimport stdio\nfrom ctypes import POINTER, c_int\n\nclass DynModTest(unittest.TestCase):\n def tearDown(self):\n for fnm in glob.glob(stdio._gen_basename + \".*\"):\n try:\n os.remove(fnm)\n except IOError:\n pass\n\n def test_fopen(self):\n self.failUnlessEqual(stdio.fopen.restype, POINTER(stdio.FILE))\n self.failUnlessEqual(stdio.fopen.argtypes, [stdio.STRING, stdio.STRING])\n\n def test_constants(self):\n self.failUnlessEqual(stdio.O_RDONLY, 0)\n self.failUnlessEqual(stdio.O_WRONLY, 1)\n self.failUnlessEqual(stdio.O_RDWR, 2)\n\nif __name__ == \"__main__\":\n unittest.main()\n","subject":"Clean up generated files in the tearDown method.","message":"Clean up generated files in the tearDown method.\n\n\ngit-svn-id: ac2c3632cb6543e7ab5fafd132c7fe15057a1882@52711 6015fed2-1504-0410-9fe1-9d1591cc4771\n","lang":"Python","license":"mit","repos":"trolldbois\/ctypeslib,trolldbois\/ctypeslib,luzfcb\/ctypeslib,trolldbois\/ctypeslib,luzfcb\/ctypeslib,luzfcb\/ctypeslib"} {"commit":"0945e04edcb4739069f4263bbd022bff4320606e","old_file":"examples\/LKE_example.py","new_file":"examples\/LKE_example.py","old_contents":"# for local run, before pygraphc packaging\nimport sys\nsys.path.insert(0, '..\/pygraphc\/misc')\nfrom LKE import *\nsys.path.insert(0, '..\/pygraphc\/clustering')\nfrom ClusterUtility import *\nfrom ClusterEvaluation import *\n\nip_address = '161.166.232.17'\nstandard_path = '\/home\/hudan\/Git\/labeled-authlog\/dataset\/' + ip_address\nstandard_file = standard_path + 'auth.log.anon.labeled'\nanalyzed_file = 'auth.log.anon'\nprediction_file = 'lke-result-' + ip_address + '.txt'\nOutputPath = '.\/results'\npara = Para(path=standard_path, logname=analyzed_file, save_path=OutputPath)\n\nmyparser = LKE(para)\ntime = myparser.main_process()\nclusters = myparser.get_clusters()\noriginal_logs = myparser.get_logs()\n\nClusterUtility.set_cluster_label_id(None, clusters, original_logs, prediction_file)\nhomogeneity_completeness_vmeasure = ClusterEvaluation.get_homogeneity_completeness_vmeasure(standard_file,\n prediction_file)\n\nprint homogeneity_completeness_vmeasure\nprint ('The running time of LKE is', time)\n","new_contents":"# for local run, before pygraphc packaging\nimport sys\nsys.path.insert(0, '..\/pygraphc\/misc')\nfrom LKE import *\nsys.path.insert(0, '..\/pygraphc\/evaluation')\nfrom ExternalEvaluation import *\n\nip_address = '161.166.232.17'\nstandard_path = '\/home\/hudan\/Git\/labeled-authlog\/dataset\/' + ip_address\nstandard_file = standard_path + 'auth.log.anon.labeled'\nanalyzed_file = 'auth.log.anon'\nprediction_file = 'lke-result-' + ip_address + '.txt'\nOutputPath = '.\/results'\npara = Para(path=standard_path, logname=analyzed_file, save_path=OutputPath)\n\nmyparser = LKE(para)\ntime = myparser.main_process()\nclusters = myparser.get_clusters()\noriginal_logs = myparser.logs\n\nExternalEvaluation.set_cluster_label_id(None, clusters, original_logs, prediction_file)\nhomogeneity_completeness_vmeasure = ExternalEvaluation.get_homogeneity_completeness_vmeasure(standard_file,\n prediction_file)\n\nprint homogeneity_completeness_vmeasure\nprint ('The running time of LKE is', time)\n","subject":"Change module path for cluster evaluation and edit how to get original logs","message":"Change module path for cluster evaluation and edit how to get original logs\n","lang":"Python","license":"mit","repos":"studiawan\/pygraphc"} {"commit":"05ac14957f8ef81a50de4f8662bb08a5d0a1bdeb","old_file":"backdrop\/collector\/__init__.py","new_file":"backdrop\/collector\/__init__.py","old_contents":"__VERSION__ = \"0.0.3\"\n__AUTHOR__ = \"GDS Developers\"\n__AUTHOR_EMAIL__ = \"\"\n","new_contents":"__VERSION__ = \"0.0.4\"\n__AUTHOR__ = \"GDS Developers\"\n__AUTHOR_EMAIL__ = \"\"\n","subject":"Bump version so as not to pull down bad logging","message":"Bump version so as not to pull down bad logging\n","lang":"Python","license":"mit","repos":"alphagov\/performanceplatform-collector,alphagov\/performanceplatform-collector,alphagov\/performanceplatform-collector,gds-attic\/backdrop-collector,gds-attic\/backdrop-collector"} {"commit":"7c49517c3c24d239c2bd44d82916b4f3d90ca1e2","old_file":"utilities\/__init__.py","new_file":"utilities\/__init__.py","old_contents":"#! \/usr\/bin\/env python\n\nfrom subprocess import Popen, PIPE\n\n\ndef launch(cmd):\n \"\"\"\n Fork the specified command, returning a tuple of (stdout, stderr)\n \"\"\"\n return Popen(cmd, shell=True, stdout=PIPE, stderr=PIPE).communicate()\n\n\ndef get_stdout(cmd):\n \"\"\"\n Fork the specified command, returning stdout\n \"\"\"\n return launch(cmd)[0]\n\n\ndef get_stderr(cmd):\n \"\"\"\n Fork the specified command, returning stderr\n \"\"\"\n return launch(cmd)[1]\n","new_contents":"#! \/usr\/bin\/env python\n\nfrom subprocess import Popen, PIPE\n\n\ndef popen(cmd):\n \"\"\"\n Fork the specified command, returning a tuple of (stdout, stderr)\n \"\"\"\n return Popen(cmd, shell=True, stdout=PIPE, stderr=PIPE).communicate()\n\n\ndef get_stdout(cmd):\n \"\"\"\n Fork the specified command, returning stdout\n \"\"\"\n return popen(cmd)[0]\n\n\ndef get_stderr(cmd):\n \"\"\"\n Fork the specified command, returning stderr\n \"\"\"\n return popen(cmd)[1]\n","subject":"Switch to using popen as the function name to stick more to subprocess naming","message":"Switch to using popen as the function name to stick more to subprocess naming\n","lang":"Python","license":"mit","repos":"IanLee1521\/utilities"} {"commit":"d147d8865dc4b82eaff87d0d4dd65ba7f4622a90","old_file":"django\/contrib\/admin\/__init__.py","new_file":"django\/contrib\/admin\/__init__.py","old_contents":"from django.contrib.admin.options import ModelAdmin, HORIZONTAL, VERTICAL\nfrom django.contrib.admin.options import StackedInline, TabularInline\nfrom django.contrib.admin.sites import AdminSite, site\n\n\ndef autodiscover():\n \"\"\"\n Auto-discover INSTALLED_APPS admin.py modules and fail silently when\n not present. This forces an import on them to register any admin bits they\n may want.\n \"\"\"\n\n import copy\n from django.conf import settings\n from django.utils.importlib import import_module\n from django.utils.module_loading import module_has_submodule\n\n for app in settings.INSTALLED_APPS:\n mod = import_module(app)\n # Attempt to import the app's admin module.\n try:\n before_import_registry = copy.copy(site._registry)\n import_module('%s.admin' % app)\n except:\n # Reset the model registry to the state before the last import as\n # this import will have to reoccur on the next request and this\n # could raise NotRegistered and AlreadyRegistered exceptions\n # (see #8245).\n site._registry = before_import_registry\n\n # Decide whether to bubble up this error. If the app just\n # doesn't have an admin module, we can ignore the error\n # attempting to import it, otherwise we want it to bubble up.\n if module_has_submodule(mod, 'admin'):\n raise\n","new_contents":"# ACTION_CHECKBOX_NAME is unused, but should stay since its import from here\n# has been referenced in documentation.\nfrom django.contrib.admin.helpers import ACTION_CHECKBOX_NAME\nfrom django.contrib.admin.options import ModelAdmin, HORIZONTAL, VERTICAL\nfrom django.contrib.admin.options import StackedInline, TabularInline\nfrom django.contrib.admin.sites import AdminSite, site\n\n\ndef autodiscover():\n \"\"\"\n Auto-discover INSTALLED_APPS admin.py modules and fail silently when\n not present. This forces an import on them to register any admin bits they\n may want.\n \"\"\"\n\n import copy\n from django.conf import settings\n from django.utils.importlib import import_module\n from django.utils.module_loading import module_has_submodule\n\n for app in settings.INSTALLED_APPS:\n mod = import_module(app)\n # Attempt to import the app's admin module.\n try:\n before_import_registry = copy.copy(site._registry)\n import_module('%s.admin' % app)\n except:\n # Reset the model registry to the state before the last import as\n # this import will have to reoccur on the next request and this\n # could raise NotRegistered and AlreadyRegistered exceptions\n # (see #8245).\n site._registry = before_import_registry\n\n # Decide whether to bubble up this error. If the app just\n # doesn't have an admin module, we can ignore the error\n # attempting to import it, otherwise we want it to bubble up.\n if module_has_submodule(mod, 'admin'):\n raise\n","subject":"Revert the removal of an unused import (in [14175]) that was referenced in documentation. Thanks for noticing, clong.","message":"Revert the removal of an unused import (in [14175]) that was referenced in documentation. Thanks for noticing, clong.\n\ngit-svn-id: http:\/\/code.djangoproject.com\/svn\/django\/trunk@14359 bcc190cf-cafb-0310-a4f2-bffc1f526a37\n\n--HG--\nextra : convert_revision : e026073455a73c9fe9a9f026b76ac783b2a12d23\n","lang":"Python","license":"bsd-3-clause","repos":"adieu\/django-nonrel,heracek\/django-nonrel,adieu\/django-nonrel,heracek\/django-nonrel,adieu\/django-nonrel,heracek\/django-nonrel"} {"commit":"4f05805c0ec31da0b978cdccc0d79336272859fe","old_file":"node\/multi_var.py","new_file":"node\/multi_var.py","old_contents":"\nfrom nodes import Node\n\nclass MultiVar(Node):\n char = \"'\"\n args = 0\n results = None\n contents = -1\n \n def __init__(self, node_1: Node.NodeSingle, node_2: Node.NodeSingle):\n self.node_1 = node_1\n self.node_2 = node_2\n self.args = max([node_1.args, node_2.args])\n \n def prepare(self, stack):\n if len(stack) == 0:\n self.add_arg(stack)\n\n @Node.is_func\n def apply(self, *stack):\n self.node_2.prepare(stack)\n rtn = self.node_2(stack[:self.node_2.args])\n self.node_1.prepare(stack)\n rtn.extend(self.node_1(stack[:self.node_1.args]))\n return rtn\n","new_contents":"\nfrom nodes import Node\n\nclass MultiVar(Node):\n char = \"'\"\n args = 0\n results = None\n contents = -1\n \n def __init__(self, node_1: Node.NodeSingle, node_2: Node.NodeSingle):\n self.node_1 = node_1\n self.node_2 = node_2\n \n def prepare(self, stack):\n self.node_1.prepare(stack)\n self.node_2.prepare(stack)\n self.args = max([self.node_1.args,self.node_2.args])\n\n @Node.is_func\n def apply(self, *stack):\n rtn = self.node_2(stack[:self.node_2.args])\n rtn.extend(self.node_1(stack[:self.node_1.args]))\n return rtn\n","subject":"Fix multivar for nodes with variable length stacks","message":"Fix multivar for nodes with variable length stacks\n","lang":"Python","license":"mit","repos":"muddyfish\/PYKE,muddyfish\/PYKE"} {"commit":"3518e9088ecbbc273f922ba418d2962d6af2dda5","old_file":"feature_extraction\/measurements\/texture_haralick.py","new_file":"feature_extraction\/measurements\/texture_haralick.py","old_contents":"from . import Measurement\nimport feature_extraction.util.cleanup as cleanup\n\nclass HaralickTexture(Measurement):\n\tdef compute(self, image):\n\t\treturn []\n","new_contents":"from . import Measurement\nimport feature_extraction.util.cleanup as cleanup\nfrom skimage.morphology import binary_erosion, disk\n\nclass HaralickTexture(Measurement):\n\tdefault_options = {\n\t\t'clip_cell_borders': True,\n\t\t'erode_cell': False,\n\t\t'erode_cell_amount': False,\n\t}\n\tdef __init__(self, options=None):\n\t\tsuper(HaralickTexture, self).__init__(options)\n\n\tdef compute(self, image):\n\t\t# -- preprocessing\n\t\tif self.options.clip_cell_borders:\n\t\t\t# get the cell boundary mask\n\t\t\tmask = cleanup.cell_boundary_mask(image)\n\n\t\t\t# if we're told to, erode the mask with a disk by some amount\n\t\t\tif self.options.erode_cell:\n\t\t\t\tmask = binary_erosion(cleanup.cell_boundary_mask(), disk(self.options.erode_cell_amount))\n\n\t\t\t# mask the image\n\t\t\timage = image[mask]\n\n\t\t# -- haralick setup and run\n\n\t\treturn []\n","subject":"Add cell-boundary preprocessing to HaralickTexture measurement","message":"Add cell-boundary preprocessing to HaralickTexture measurement\n","lang":"Python","license":"apache-2.0","repos":"widoptimization-willett\/feature-extraction"} {"commit":"b08315337e71737a36e3e79da99ce167620711b9","old_file":"photodaemon.py","new_file":"photodaemon.py","old_contents":"#!\/bin\/env python\n\nimport picamera\nimport redis\nimport time\nimport json\nimport os\n\ndef take_photo():\n print \"%s Capture photo\" % (time.strftime('%Y.%m.%d %H:%M:%S %Z'))\n camera = picamera.PiCamera()\n camera.vflip = True\n camera.resolution = (1280, 720)\n time.sleep(1)\n camera.capture('static\/photo.jpg')\n camera.close()\n r.publish('photo', time.time())\n print \"%s Capture done\" % (time.strftime('%Y.%m.%d %H:%M:%S %Z'))\n\ndef get_config():\n with open(os.path.join(os.path.dirname(__file__), 'config.json')) as data_file:\n return json.load(data_file)\n\ndef main():\n config = get_config()\n\n r = redis.StrictRedis(host=config['host'], port=config['port'], db=config['db'])\n p = r.pubsub()\n p.subscribe('take-photo')\n\n while True:\n message = p.get_message()\n if message and message['type'] == 'message':\n take_photo()\n\n time.sleep(0.1)\n\nif __name__ == '__main__':\n main()\n","new_contents":"#!\/bin\/env python\n\nimport picamera\nimport redis\nimport time\nimport json\nimport os\n\ndef take_photo():\n print \"%s Capture photo\" % (time.strftime('%Y.%m.%d %H:%M:%S %Z'))\n camera = picamera.PiCamera()\n camera.vflip = True\n camera.resolution = (1280, 720)\n time.sleep(1)\n camera.capture('static\/photo.jpg')\n camera.close()\n print \"%s Capture done\" % (time.strftime('%Y.%m.%d %H:%M:%S %Z'))\n\ndef get_config():\n with open(os.path.join(os.path.dirname(__file__), 'config.json')) as data_file:\n return json.load(data_file)\n\ndef main():\n config = get_config()\n\n r = redis.StrictRedis(host=config['host'], port=config['port'], db=config['db'])\n p = r.pubsub()\n p.subscribe('take-photo')\n\n while True:\n message = p.get_message()\n if message and message['type'] == 'message':\n take_photo()\n r.publish('photo', time.time())\n\n time.sleep(0.1)\n\nif __name__ == '__main__':\n main()\n","subject":"Fix publishing photo creation event","message":"Fix publishing photo creation event\n","lang":"Python","license":"mit","repos":"Ajnasz\/pippo,Ajnasz\/pippo,Ajnasz\/pippo"} {"commit":"b2fbb48049abbfff7f1636059f8ad7eda07667c7","old_file":"test\/single_system\/all.py","new_file":"test\/single_system\/all.py","old_contents":"import sys, unittest\nimport bmc_test\nimport power_test\nimport xmlrunner\n\ntests = []\ntests.extend(bmc_test.tests)\n#tests.extend(power_test.tests)\n\nif __name__ == '__main__':\n for test in tests:\n test.system = sys.argv[1]\n suite = unittest.TestLoader().loadTestsFromTestCase(test)\n xmlrunner.XMLTestRunner(verbose = 1, output='test-reports').run(suite)\n","new_contents":"import sys, unittest, os\nimport bmc_test\nimport power_test\nimport xmlrunner\n\ntests = []\ntests.extend(bmc_test.tests)\n#tests.extend(power_test.tests)\n\nif __name__ == '__main__':\n for test in tests:\n test.system = sys.argv[1]\n suite = unittest.TestLoader().loadTestsFromTestCase(test)\n result = xmlrunner.XMLTestRunner(verbose = 1, output='test-reports').run(suite)\n if result.failures or result.errors:\n os.sys.exit(1)\n","subject":"Return a bad error code when a test fails","message":"Return a bad error code when a test fails\n","lang":"Python","license":"bsd-3-clause","repos":"Cynerva\/pyipmi,emaadmanzoor\/pyipmi"}