repo
stringlengths 7
55
| path
stringlengths 4
127
| func_name
stringlengths 1
88
| original_string
stringlengths 75
19.8k
| language
stringclasses 1
value | code
stringlengths 75
19.8k
| code_tokens
list | docstring
stringlengths 3
17.3k
| docstring_tokens
list | sha
stringlengths 40
40
| url
stringlengths 87
242
| partition
stringclasses 1
value |
---|---|---|---|---|---|---|---|---|---|---|---|
cyface/django-termsandconditions | termsandconditions/views.py | TermsView.get_context_data | def get_context_data(self, **kwargs):
"""Pass additional context data"""
context = super(TermsView, self).get_context_data(**kwargs)
context['terms_base_template'] = getattr(settings, 'TERMS_BASE_TEMPLATE', DEFAULT_TERMS_BASE_TEMPLATE)
return context | python | def get_context_data(self, **kwargs):
"""Pass additional context data"""
context = super(TermsView, self).get_context_data(**kwargs)
context['terms_base_template'] = getattr(settings, 'TERMS_BASE_TEMPLATE', DEFAULT_TERMS_BASE_TEMPLATE)
return context | [
"def",
"get_context_data",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"context",
"=",
"super",
"(",
"TermsView",
",",
"self",
")",
".",
"get_context_data",
"(",
"*",
"*",
"kwargs",
")",
"context",
"[",
"'terms_base_template'",
"]",
"=",
"getattr",
"(",
"settings",
",",
"'TERMS_BASE_TEMPLATE'",
",",
"DEFAULT_TERMS_BASE_TEMPLATE",
")",
"return",
"context"
] | Pass additional context data | [
"Pass",
"additional",
"context",
"data"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/views.py#L53-L57 | train |
cyface/django-termsandconditions | termsandconditions/views.py | AcceptTermsView.get_initial | def get_initial(self):
"""Override of CreateView method, queries for which T&C to accept and catches returnTo from URL"""
LOGGER.debug('termsandconditions.views.AcceptTermsView.get_initial')
terms = self.get_terms(self.kwargs)
return_to = self.request.GET.get('returnTo', '/')
return {'terms': terms, 'returnTo': return_to} | python | def get_initial(self):
"""Override of CreateView method, queries for which T&C to accept and catches returnTo from URL"""
LOGGER.debug('termsandconditions.views.AcceptTermsView.get_initial')
terms = self.get_terms(self.kwargs)
return_to = self.request.GET.get('returnTo', '/')
return {'terms': terms, 'returnTo': return_to} | [
"def",
"get_initial",
"(",
"self",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'termsandconditions.views.AcceptTermsView.get_initial'",
")",
"terms",
"=",
"self",
".",
"get_terms",
"(",
"self",
".",
"kwargs",
")",
"return_to",
"=",
"self",
".",
"request",
".",
"GET",
".",
"get",
"(",
"'returnTo'",
",",
"'/'",
")",
"return",
"{",
"'terms'",
":",
"terms",
",",
"'returnTo'",
":",
"return_to",
"}"
] | Override of CreateView method, queries for which T&C to accept and catches returnTo from URL | [
"Override",
"of",
"CreateView",
"method",
"queries",
"for",
"which",
"T&C",
"to",
"accept",
"and",
"catches",
"returnTo",
"from",
"URL"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/views.py#L82-L89 | train |
cyface/django-termsandconditions | termsandconditions/views.py | AcceptTermsView.post | def post(self, request, *args, **kwargs):
"""
Handles POST request.
"""
return_url = request.POST.get('returnTo', '/')
terms_ids = request.POST.getlist('terms')
if not terms_ids: # pragma: nocover
return HttpResponseRedirect(return_url)
if DJANGO_VERSION <= (2, 0, 0):
user_authenticated = request.user.is_authenticated()
else:
user_authenticated = request.user.is_authenticated
if user_authenticated:
user = request.user
else:
# Get user out of saved pipeline from django-socialauth
if 'partial_pipeline' in request.session:
user_pk = request.session['partial_pipeline']['kwargs']['user']['pk']
user = User.objects.get(id=user_pk)
else:
return HttpResponseRedirect('/')
store_ip_address = getattr(settings, 'TERMS_STORE_IP_ADDRESS', True)
if store_ip_address:
ip_address = request.META.get(getattr(settings, 'TERMS_IP_HEADER_NAME', DEFAULT_TERMS_IP_HEADER_NAME))
else:
ip_address = ""
for terms_id in terms_ids:
try:
new_user_terms = UserTermsAndConditions(
user=user,
terms=TermsAndConditions.objects.get(pk=int(terms_id)),
ip_address=ip_address
)
new_user_terms.save()
except IntegrityError: # pragma: nocover
pass
return HttpResponseRedirect(return_url) | python | def post(self, request, *args, **kwargs):
"""
Handles POST request.
"""
return_url = request.POST.get('returnTo', '/')
terms_ids = request.POST.getlist('terms')
if not terms_ids: # pragma: nocover
return HttpResponseRedirect(return_url)
if DJANGO_VERSION <= (2, 0, 0):
user_authenticated = request.user.is_authenticated()
else:
user_authenticated = request.user.is_authenticated
if user_authenticated:
user = request.user
else:
# Get user out of saved pipeline from django-socialauth
if 'partial_pipeline' in request.session:
user_pk = request.session['partial_pipeline']['kwargs']['user']['pk']
user = User.objects.get(id=user_pk)
else:
return HttpResponseRedirect('/')
store_ip_address = getattr(settings, 'TERMS_STORE_IP_ADDRESS', True)
if store_ip_address:
ip_address = request.META.get(getattr(settings, 'TERMS_IP_HEADER_NAME', DEFAULT_TERMS_IP_HEADER_NAME))
else:
ip_address = ""
for terms_id in terms_ids:
try:
new_user_terms = UserTermsAndConditions(
user=user,
terms=TermsAndConditions.objects.get(pk=int(terms_id)),
ip_address=ip_address
)
new_user_terms.save()
except IntegrityError: # pragma: nocover
pass
return HttpResponseRedirect(return_url) | [
"def",
"post",
"(",
"self",
",",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"return_url",
"=",
"request",
".",
"POST",
".",
"get",
"(",
"'returnTo'",
",",
"'/'",
")",
"terms_ids",
"=",
"request",
".",
"POST",
".",
"getlist",
"(",
"'terms'",
")",
"if",
"not",
"terms_ids",
":",
"# pragma: nocover",
"return",
"HttpResponseRedirect",
"(",
"return_url",
")",
"if",
"DJANGO_VERSION",
"<=",
"(",
"2",
",",
"0",
",",
"0",
")",
":",
"user_authenticated",
"=",
"request",
".",
"user",
".",
"is_authenticated",
"(",
")",
"else",
":",
"user_authenticated",
"=",
"request",
".",
"user",
".",
"is_authenticated",
"if",
"user_authenticated",
":",
"user",
"=",
"request",
".",
"user",
"else",
":",
"# Get user out of saved pipeline from django-socialauth",
"if",
"'partial_pipeline'",
"in",
"request",
".",
"session",
":",
"user_pk",
"=",
"request",
".",
"session",
"[",
"'partial_pipeline'",
"]",
"[",
"'kwargs'",
"]",
"[",
"'user'",
"]",
"[",
"'pk'",
"]",
"user",
"=",
"User",
".",
"objects",
".",
"get",
"(",
"id",
"=",
"user_pk",
")",
"else",
":",
"return",
"HttpResponseRedirect",
"(",
"'/'",
")",
"store_ip_address",
"=",
"getattr",
"(",
"settings",
",",
"'TERMS_STORE_IP_ADDRESS'",
",",
"True",
")",
"if",
"store_ip_address",
":",
"ip_address",
"=",
"request",
".",
"META",
".",
"get",
"(",
"getattr",
"(",
"settings",
",",
"'TERMS_IP_HEADER_NAME'",
",",
"DEFAULT_TERMS_IP_HEADER_NAME",
")",
")",
"else",
":",
"ip_address",
"=",
"\"\"",
"for",
"terms_id",
"in",
"terms_ids",
":",
"try",
":",
"new_user_terms",
"=",
"UserTermsAndConditions",
"(",
"user",
"=",
"user",
",",
"terms",
"=",
"TermsAndConditions",
".",
"objects",
".",
"get",
"(",
"pk",
"=",
"int",
"(",
"terms_id",
")",
")",
",",
"ip_address",
"=",
"ip_address",
")",
"new_user_terms",
".",
"save",
"(",
")",
"except",
"IntegrityError",
":",
"# pragma: nocover",
"pass",
"return",
"HttpResponseRedirect",
"(",
"return_url",
")"
] | Handles POST request. | [
"Handles",
"POST",
"request",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/views.py#L91-L133 | train |
cyface/django-termsandconditions | termsandconditions/views.py | EmailTermsView.form_valid | def form_valid(self, form):
"""Override of CreateView method, sends the email."""
LOGGER.debug('termsandconditions.views.EmailTermsView.form_valid')
template = get_template("termsandconditions/tc_email_terms.html")
template_rendered = template.render({"terms": form.cleaned_data.get('terms')})
LOGGER.debug("Email Terms Body:")
LOGGER.debug(template_rendered)
try:
send_mail(form.cleaned_data.get('email_subject', _('Terms')),
template_rendered,
settings.DEFAULT_FROM_EMAIL,
[form.cleaned_data.get('email_address')],
fail_silently=False)
messages.add_message(self.request, messages.INFO, _("Terms and Conditions Sent."))
except SMTPException: # pragma: no cover
messages.add_message(self.request, messages.ERROR, _("An Error Occurred Sending Your Message."))
self.success_url = form.cleaned_data.get('returnTo', '/') or '/'
return super(EmailTermsView, self).form_valid(form) | python | def form_valid(self, form):
"""Override of CreateView method, sends the email."""
LOGGER.debug('termsandconditions.views.EmailTermsView.form_valid')
template = get_template("termsandconditions/tc_email_terms.html")
template_rendered = template.render({"terms": form.cleaned_data.get('terms')})
LOGGER.debug("Email Terms Body:")
LOGGER.debug(template_rendered)
try:
send_mail(form.cleaned_data.get('email_subject', _('Terms')),
template_rendered,
settings.DEFAULT_FROM_EMAIL,
[form.cleaned_data.get('email_address')],
fail_silently=False)
messages.add_message(self.request, messages.INFO, _("Terms and Conditions Sent."))
except SMTPException: # pragma: no cover
messages.add_message(self.request, messages.ERROR, _("An Error Occurred Sending Your Message."))
self.success_url = form.cleaned_data.get('returnTo', '/') or '/'
return super(EmailTermsView, self).form_valid(form) | [
"def",
"form_valid",
"(",
"self",
",",
"form",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'termsandconditions.views.EmailTermsView.form_valid'",
")",
"template",
"=",
"get_template",
"(",
"\"termsandconditions/tc_email_terms.html\"",
")",
"template_rendered",
"=",
"template",
".",
"render",
"(",
"{",
"\"terms\"",
":",
"form",
".",
"cleaned_data",
".",
"get",
"(",
"'terms'",
")",
"}",
")",
"LOGGER",
".",
"debug",
"(",
"\"Email Terms Body:\"",
")",
"LOGGER",
".",
"debug",
"(",
"template_rendered",
")",
"try",
":",
"send_mail",
"(",
"form",
".",
"cleaned_data",
".",
"get",
"(",
"'email_subject'",
",",
"_",
"(",
"'Terms'",
")",
")",
",",
"template_rendered",
",",
"settings",
".",
"DEFAULT_FROM_EMAIL",
",",
"[",
"form",
".",
"cleaned_data",
".",
"get",
"(",
"'email_address'",
")",
"]",
",",
"fail_silently",
"=",
"False",
")",
"messages",
".",
"add_message",
"(",
"self",
".",
"request",
",",
"messages",
".",
"INFO",
",",
"_",
"(",
"\"Terms and Conditions Sent.\"",
")",
")",
"except",
"SMTPException",
":",
"# pragma: no cover",
"messages",
".",
"add_message",
"(",
"self",
".",
"request",
",",
"messages",
".",
"ERROR",
",",
"_",
"(",
"\"An Error Occurred Sending Your Message.\"",
")",
")",
"self",
".",
"success_url",
"=",
"form",
".",
"cleaned_data",
".",
"get",
"(",
"'returnTo'",
",",
"'/'",
")",
"or",
"'/'",
"return",
"super",
"(",
"EmailTermsView",
",",
"self",
")",
".",
"form_valid",
"(",
"form",
")"
] | Override of CreateView method, sends the email. | [
"Override",
"of",
"CreateView",
"method",
"sends",
"the",
"email",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/views.py#L162-L184 | train |
cyface/django-termsandconditions | termsandconditions/views.py | EmailTermsView.form_invalid | def form_invalid(self, form):
"""Override of CreateView method, logs invalid email form submissions."""
LOGGER.debug("Invalid Email Form Submitted")
messages.add_message(self.request, messages.ERROR, _("Invalid Email Address."))
return super(EmailTermsView, self).form_invalid(form) | python | def form_invalid(self, form):
"""Override of CreateView method, logs invalid email form submissions."""
LOGGER.debug("Invalid Email Form Submitted")
messages.add_message(self.request, messages.ERROR, _("Invalid Email Address."))
return super(EmailTermsView, self).form_invalid(form) | [
"def",
"form_invalid",
"(",
"self",
",",
"form",
")",
":",
"LOGGER",
".",
"debug",
"(",
"\"Invalid Email Form Submitted\"",
")",
"messages",
".",
"add_message",
"(",
"self",
".",
"request",
",",
"messages",
".",
"ERROR",
",",
"_",
"(",
"\"Invalid Email Address.\"",
")",
")",
"return",
"super",
"(",
"EmailTermsView",
",",
"self",
")",
".",
"form_invalid",
"(",
"form",
")"
] | Override of CreateView method, logs invalid email form submissions. | [
"Override",
"of",
"CreateView",
"method",
"logs",
"invalid",
"email",
"form",
"submissions",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/views.py#L186-L190 | train |
cyface/django-termsandconditions | termsandconditions/decorators.py | terms_required | def terms_required(view_func):
"""
This decorator checks to see if the user is logged in, and if so, if they have accepted the site terms.
"""
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view(request, *args, **kwargs):
"""Method to wrap the view passed in"""
# If user has not logged in, or if they have logged in and already agreed to the terms, let the view through
if DJANGO_VERSION <= (2, 0, 0):
user_authenticated = request.user.is_authenticated()
else:
user_authenticated = request.user.is_authenticated
if not user_authenticated or not TermsAndConditions.get_active_terms_not_agreed_to(request.user):
return view_func(request, *args, **kwargs)
# Otherwise, redirect to terms accept
current_path = request.path
login_url_parts = list(urlparse(ACCEPT_TERMS_PATH))
querystring = QueryDict(login_url_parts[4], mutable=True)
querystring['returnTo'] = current_path
login_url_parts[4] = querystring.urlencode(safe='/')
return HttpResponseRedirect(urlunparse(login_url_parts))
return _wrapped_view | python | def terms_required(view_func):
"""
This decorator checks to see if the user is logged in, and if so, if they have accepted the site terms.
"""
@wraps(view_func, assigned=available_attrs(view_func))
def _wrapped_view(request, *args, **kwargs):
"""Method to wrap the view passed in"""
# If user has not logged in, or if they have logged in and already agreed to the terms, let the view through
if DJANGO_VERSION <= (2, 0, 0):
user_authenticated = request.user.is_authenticated()
else:
user_authenticated = request.user.is_authenticated
if not user_authenticated or not TermsAndConditions.get_active_terms_not_agreed_to(request.user):
return view_func(request, *args, **kwargs)
# Otherwise, redirect to terms accept
current_path = request.path
login_url_parts = list(urlparse(ACCEPT_TERMS_PATH))
querystring = QueryDict(login_url_parts[4], mutable=True)
querystring['returnTo'] = current_path
login_url_parts[4] = querystring.urlencode(safe='/')
return HttpResponseRedirect(urlunparse(login_url_parts))
return _wrapped_view | [
"def",
"terms_required",
"(",
"view_func",
")",
":",
"@",
"wraps",
"(",
"view_func",
",",
"assigned",
"=",
"available_attrs",
"(",
"view_func",
")",
")",
"def",
"_wrapped_view",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"\"\"\"Method to wrap the view passed in\"\"\"",
"# If user has not logged in, or if they have logged in and already agreed to the terms, let the view through",
"if",
"DJANGO_VERSION",
"<=",
"(",
"2",
",",
"0",
",",
"0",
")",
":",
"user_authenticated",
"=",
"request",
".",
"user",
".",
"is_authenticated",
"(",
")",
"else",
":",
"user_authenticated",
"=",
"request",
".",
"user",
".",
"is_authenticated",
"if",
"not",
"user_authenticated",
"or",
"not",
"TermsAndConditions",
".",
"get_active_terms_not_agreed_to",
"(",
"request",
".",
"user",
")",
":",
"return",
"view_func",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"# Otherwise, redirect to terms accept",
"current_path",
"=",
"request",
".",
"path",
"login_url_parts",
"=",
"list",
"(",
"urlparse",
"(",
"ACCEPT_TERMS_PATH",
")",
")",
"querystring",
"=",
"QueryDict",
"(",
"login_url_parts",
"[",
"4",
"]",
",",
"mutable",
"=",
"True",
")",
"querystring",
"[",
"'returnTo'",
"]",
"=",
"current_path",
"login_url_parts",
"[",
"4",
"]",
"=",
"querystring",
".",
"urlencode",
"(",
"safe",
"=",
"'/'",
")",
"return",
"HttpResponseRedirect",
"(",
"urlunparse",
"(",
"login_url_parts",
")",
")",
"return",
"_wrapped_view"
] | This decorator checks to see if the user is logged in, and if so, if they have accepted the site terms. | [
"This",
"decorator",
"checks",
"to",
"see",
"if",
"the",
"user",
"is",
"logged",
"in",
"and",
"if",
"so",
"if",
"they",
"have",
"accepted",
"the",
"site",
"terms",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/decorators.py#L11-L36 | train |
cyface/django-termsandconditions | termsandconditions/models.py | TermsAndConditions.get_active | def get_active(slug=DEFAULT_TERMS_SLUG):
"""Finds the latest of a particular terms and conditions"""
active_terms = cache.get('tandc.active_terms_' + slug)
if active_terms is None:
try:
active_terms = TermsAndConditions.objects.filter(
date_active__isnull=False,
date_active__lte=timezone.now(),
slug=slug).latest('date_active')
cache.set('tandc.active_terms_' + slug, active_terms, TERMS_CACHE_SECONDS)
except TermsAndConditions.DoesNotExist: # pragma: nocover
LOGGER.error("Requested Terms and Conditions that Have Not Been Created.")
return None
return active_terms | python | def get_active(slug=DEFAULT_TERMS_SLUG):
"""Finds the latest of a particular terms and conditions"""
active_terms = cache.get('tandc.active_terms_' + slug)
if active_terms is None:
try:
active_terms = TermsAndConditions.objects.filter(
date_active__isnull=False,
date_active__lte=timezone.now(),
slug=slug).latest('date_active')
cache.set('tandc.active_terms_' + slug, active_terms, TERMS_CACHE_SECONDS)
except TermsAndConditions.DoesNotExist: # pragma: nocover
LOGGER.error("Requested Terms and Conditions that Have Not Been Created.")
return None
return active_terms | [
"def",
"get_active",
"(",
"slug",
"=",
"DEFAULT_TERMS_SLUG",
")",
":",
"active_terms",
"=",
"cache",
".",
"get",
"(",
"'tandc.active_terms_'",
"+",
"slug",
")",
"if",
"active_terms",
"is",
"None",
":",
"try",
":",
"active_terms",
"=",
"TermsAndConditions",
".",
"objects",
".",
"filter",
"(",
"date_active__isnull",
"=",
"False",
",",
"date_active__lte",
"=",
"timezone",
".",
"now",
"(",
")",
",",
"slug",
"=",
"slug",
")",
".",
"latest",
"(",
"'date_active'",
")",
"cache",
".",
"set",
"(",
"'tandc.active_terms_'",
"+",
"slug",
",",
"active_terms",
",",
"TERMS_CACHE_SECONDS",
")",
"except",
"TermsAndConditions",
".",
"DoesNotExist",
":",
"# pragma: nocover",
"LOGGER",
".",
"error",
"(",
"\"Requested Terms and Conditions that Have Not Been Created.\"",
")",
"return",
"None",
"return",
"active_terms"
] | Finds the latest of a particular terms and conditions | [
"Finds",
"the",
"latest",
"of",
"a",
"particular",
"terms",
"and",
"conditions"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/models.py#L75-L90 | train |
cyface/django-termsandconditions | termsandconditions/models.py | TermsAndConditions.get_active_terms_ids | def get_active_terms_ids():
"""Returns a list of the IDs of of all terms and conditions"""
active_terms_ids = cache.get('tandc.active_terms_ids')
if active_terms_ids is None:
active_terms_dict = {}
active_terms_ids = []
active_terms_set = TermsAndConditions.objects.filter(date_active__isnull=False, date_active__lte=timezone.now()).order_by('date_active')
for active_terms in active_terms_set:
active_terms_dict[active_terms.slug] = active_terms.id
active_terms_dict = OrderedDict(sorted(active_terms_dict.items(), key=lambda t: t[0]))
for terms in active_terms_dict:
active_terms_ids.append(active_terms_dict[terms])
cache.set('tandc.active_terms_ids', active_terms_ids, TERMS_CACHE_SECONDS)
return active_terms_ids | python | def get_active_terms_ids():
"""Returns a list of the IDs of of all terms and conditions"""
active_terms_ids = cache.get('tandc.active_terms_ids')
if active_terms_ids is None:
active_terms_dict = {}
active_terms_ids = []
active_terms_set = TermsAndConditions.objects.filter(date_active__isnull=False, date_active__lte=timezone.now()).order_by('date_active')
for active_terms in active_terms_set:
active_terms_dict[active_terms.slug] = active_terms.id
active_terms_dict = OrderedDict(sorted(active_terms_dict.items(), key=lambda t: t[0]))
for terms in active_terms_dict:
active_terms_ids.append(active_terms_dict[terms])
cache.set('tandc.active_terms_ids', active_terms_ids, TERMS_CACHE_SECONDS)
return active_terms_ids | [
"def",
"get_active_terms_ids",
"(",
")",
":",
"active_terms_ids",
"=",
"cache",
".",
"get",
"(",
"'tandc.active_terms_ids'",
")",
"if",
"active_terms_ids",
"is",
"None",
":",
"active_terms_dict",
"=",
"{",
"}",
"active_terms_ids",
"=",
"[",
"]",
"active_terms_set",
"=",
"TermsAndConditions",
".",
"objects",
".",
"filter",
"(",
"date_active__isnull",
"=",
"False",
",",
"date_active__lte",
"=",
"timezone",
".",
"now",
"(",
")",
")",
".",
"order_by",
"(",
"'date_active'",
")",
"for",
"active_terms",
"in",
"active_terms_set",
":",
"active_terms_dict",
"[",
"active_terms",
".",
"slug",
"]",
"=",
"active_terms",
".",
"id",
"active_terms_dict",
"=",
"OrderedDict",
"(",
"sorted",
"(",
"active_terms_dict",
".",
"items",
"(",
")",
",",
"key",
"=",
"lambda",
"t",
":",
"t",
"[",
"0",
"]",
")",
")",
"for",
"terms",
"in",
"active_terms_dict",
":",
"active_terms_ids",
".",
"append",
"(",
"active_terms_dict",
"[",
"terms",
"]",
")",
"cache",
".",
"set",
"(",
"'tandc.active_terms_ids'",
",",
"active_terms_ids",
",",
"TERMS_CACHE_SECONDS",
")",
"return",
"active_terms_ids"
] | Returns a list of the IDs of of all terms and conditions | [
"Returns",
"a",
"list",
"of",
"the",
"IDs",
"of",
"of",
"all",
"terms",
"and",
"conditions"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/models.py#L93-L112 | train |
cyface/django-termsandconditions | termsandconditions/models.py | TermsAndConditions.get_active_terms_list | def get_active_terms_list():
"""Returns all the latest active terms and conditions"""
active_terms_list = cache.get('tandc.active_terms_list')
if active_terms_list is None:
active_terms_list = TermsAndConditions.objects.filter(id__in=TermsAndConditions.get_active_terms_ids()).order_by('slug')
cache.set('tandc.active_terms_list', active_terms_list, TERMS_CACHE_SECONDS)
return active_terms_list | python | def get_active_terms_list():
"""Returns all the latest active terms and conditions"""
active_terms_list = cache.get('tandc.active_terms_list')
if active_terms_list is None:
active_terms_list = TermsAndConditions.objects.filter(id__in=TermsAndConditions.get_active_terms_ids()).order_by('slug')
cache.set('tandc.active_terms_list', active_terms_list, TERMS_CACHE_SECONDS)
return active_terms_list | [
"def",
"get_active_terms_list",
"(",
")",
":",
"active_terms_list",
"=",
"cache",
".",
"get",
"(",
"'tandc.active_terms_list'",
")",
"if",
"active_terms_list",
"is",
"None",
":",
"active_terms_list",
"=",
"TermsAndConditions",
".",
"objects",
".",
"filter",
"(",
"id__in",
"=",
"TermsAndConditions",
".",
"get_active_terms_ids",
"(",
")",
")",
".",
"order_by",
"(",
"'slug'",
")",
"cache",
".",
"set",
"(",
"'tandc.active_terms_list'",
",",
"active_terms_list",
",",
"TERMS_CACHE_SECONDS",
")",
"return",
"active_terms_list"
] | Returns all the latest active terms and conditions | [
"Returns",
"all",
"the",
"latest",
"active",
"terms",
"and",
"conditions"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/models.py#L115-L123 | train |
cyface/django-termsandconditions | termsandconditions/models.py | TermsAndConditions.get_active_terms_not_agreed_to | def get_active_terms_not_agreed_to(user):
"""Checks to see if a specified user has agreed to all the latest terms and conditions"""
if TERMS_EXCLUDE_USERS_WITH_PERM is not None:
if user.has_perm(TERMS_EXCLUDE_USERS_WITH_PERM) and not user.is_superuser:
# Django's has_perm() returns True if is_superuser, we don't want that
return []
not_agreed_terms = cache.get('tandc.not_agreed_terms_' + user.get_username())
if not_agreed_terms is None:
try:
LOGGER.debug("Not Agreed Terms")
not_agreed_terms = TermsAndConditions.get_active_terms_list().exclude(
userterms__in=UserTermsAndConditions.objects.filter(user=user)
).order_by('slug')
cache.set('tandc.not_agreed_terms_' + user.get_username(), not_agreed_terms, TERMS_CACHE_SECONDS)
except (TypeError, UserTermsAndConditions.DoesNotExist):
return []
return not_agreed_terms | python | def get_active_terms_not_agreed_to(user):
"""Checks to see if a specified user has agreed to all the latest terms and conditions"""
if TERMS_EXCLUDE_USERS_WITH_PERM is not None:
if user.has_perm(TERMS_EXCLUDE_USERS_WITH_PERM) and not user.is_superuser:
# Django's has_perm() returns True if is_superuser, we don't want that
return []
not_agreed_terms = cache.get('tandc.not_agreed_terms_' + user.get_username())
if not_agreed_terms is None:
try:
LOGGER.debug("Not Agreed Terms")
not_agreed_terms = TermsAndConditions.get_active_terms_list().exclude(
userterms__in=UserTermsAndConditions.objects.filter(user=user)
).order_by('slug')
cache.set('tandc.not_agreed_terms_' + user.get_username(), not_agreed_terms, TERMS_CACHE_SECONDS)
except (TypeError, UserTermsAndConditions.DoesNotExist):
return []
return not_agreed_terms | [
"def",
"get_active_terms_not_agreed_to",
"(",
"user",
")",
":",
"if",
"TERMS_EXCLUDE_USERS_WITH_PERM",
"is",
"not",
"None",
":",
"if",
"user",
".",
"has_perm",
"(",
"TERMS_EXCLUDE_USERS_WITH_PERM",
")",
"and",
"not",
"user",
".",
"is_superuser",
":",
"# Django's has_perm() returns True if is_superuser, we don't want that",
"return",
"[",
"]",
"not_agreed_terms",
"=",
"cache",
".",
"get",
"(",
"'tandc.not_agreed_terms_'",
"+",
"user",
".",
"get_username",
"(",
")",
")",
"if",
"not_agreed_terms",
"is",
"None",
":",
"try",
":",
"LOGGER",
".",
"debug",
"(",
"\"Not Agreed Terms\"",
")",
"not_agreed_terms",
"=",
"TermsAndConditions",
".",
"get_active_terms_list",
"(",
")",
".",
"exclude",
"(",
"userterms__in",
"=",
"UserTermsAndConditions",
".",
"objects",
".",
"filter",
"(",
"user",
"=",
"user",
")",
")",
".",
"order_by",
"(",
"'slug'",
")",
"cache",
".",
"set",
"(",
"'tandc.not_agreed_terms_'",
"+",
"user",
".",
"get_username",
"(",
")",
",",
"not_agreed_terms",
",",
"TERMS_CACHE_SECONDS",
")",
"except",
"(",
"TypeError",
",",
"UserTermsAndConditions",
".",
"DoesNotExist",
")",
":",
"return",
"[",
"]",
"return",
"not_agreed_terms"
] | Checks to see if a specified user has agreed to all the latest terms and conditions | [
"Checks",
"to",
"see",
"if",
"a",
"specified",
"user",
"has",
"agreed",
"to",
"all",
"the",
"latest",
"terms",
"and",
"conditions"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/models.py#L126-L146 | train |
cyface/django-termsandconditions | termsandconditions/templatetags/terms_tags.py | show_terms_if_not_agreed | def show_terms_if_not_agreed(context, field=TERMS_HTTP_PATH_FIELD):
"""Displays a modal on a current page if a user has not yet agreed to the
given terms. If terms are not specified, the default slug is used.
A small snippet is included into your template if a user
who requested the view has not yet agreed the terms. The snippet takes
care of displaying a respective modal.
"""
request = context['request']
url = urlparse(request.META[field])
not_agreed_terms = TermsAndConditions.get_active_terms_not_agreed_to(request.user)
if not_agreed_terms and is_path_protected(url.path):
return {'not_agreed_terms': not_agreed_terms, 'returnTo': url.path}
else:
return {} | python | def show_terms_if_not_agreed(context, field=TERMS_HTTP_PATH_FIELD):
"""Displays a modal on a current page if a user has not yet agreed to the
given terms. If terms are not specified, the default slug is used.
A small snippet is included into your template if a user
who requested the view has not yet agreed the terms. The snippet takes
care of displaying a respective modal.
"""
request = context['request']
url = urlparse(request.META[field])
not_agreed_terms = TermsAndConditions.get_active_terms_not_agreed_to(request.user)
if not_agreed_terms and is_path_protected(url.path):
return {'not_agreed_terms': not_agreed_terms, 'returnTo': url.path}
else:
return {} | [
"def",
"show_terms_if_not_agreed",
"(",
"context",
",",
"field",
"=",
"TERMS_HTTP_PATH_FIELD",
")",
":",
"request",
"=",
"context",
"[",
"'request'",
"]",
"url",
"=",
"urlparse",
"(",
"request",
".",
"META",
"[",
"field",
"]",
")",
"not_agreed_terms",
"=",
"TermsAndConditions",
".",
"get_active_terms_not_agreed_to",
"(",
"request",
".",
"user",
")",
"if",
"not_agreed_terms",
"and",
"is_path_protected",
"(",
"url",
".",
"path",
")",
":",
"return",
"{",
"'not_agreed_terms'",
":",
"not_agreed_terms",
",",
"'returnTo'",
":",
"url",
".",
"path",
"}",
"else",
":",
"return",
"{",
"}"
] | Displays a modal on a current page if a user has not yet agreed to the
given terms. If terms are not specified, the default slug is used.
A small snippet is included into your template if a user
who requested the view has not yet agreed the terms. The snippet takes
care of displaying a respective modal. | [
"Displays",
"a",
"modal",
"on",
"a",
"current",
"page",
"if",
"a",
"user",
"has",
"not",
"yet",
"agreed",
"to",
"the",
"given",
"terms",
".",
"If",
"terms",
"are",
"not",
"specified",
"the",
"default",
"slug",
"is",
"used",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/templatetags/terms_tags.py#L15-L30 | train |
cyface/django-termsandconditions | termsandconditions/pipeline.py | user_accept_terms | def user_accept_terms(backend, user, uid, social_user=None, *args, **kwargs):
"""Check if the user has accepted the terms and conditions after creation."""
LOGGER.debug('user_accept_terms')
if TermsAndConditions.get_active_terms_not_agreed_to(user):
return redirect_to_terms_accept('/')
else:
return {'social_user': social_user, 'user': user} | python | def user_accept_terms(backend, user, uid, social_user=None, *args, **kwargs):
"""Check if the user has accepted the terms and conditions after creation."""
LOGGER.debug('user_accept_terms')
if TermsAndConditions.get_active_terms_not_agreed_to(user):
return redirect_to_terms_accept('/')
else:
return {'social_user': social_user, 'user': user} | [
"def",
"user_accept_terms",
"(",
"backend",
",",
"user",
",",
"uid",
",",
"social_user",
"=",
"None",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"LOGGER",
".",
"debug",
"(",
"'user_accept_terms'",
")",
"if",
"TermsAndConditions",
".",
"get_active_terms_not_agreed_to",
"(",
"user",
")",
":",
"return",
"redirect_to_terms_accept",
"(",
"'/'",
")",
"else",
":",
"return",
"{",
"'social_user'",
":",
"social_user",
",",
"'user'",
":",
"user",
"}"
] | Check if the user has accepted the terms and conditions after creation. | [
"Check",
"if",
"the",
"user",
"has",
"accepted",
"the",
"terms",
"and",
"conditions",
"after",
"creation",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/pipeline.py#L17-L25 | train |
cyface/django-termsandconditions | termsandconditions/pipeline.py | redirect_to_terms_accept | def redirect_to_terms_accept(current_path='/', slug='default'):
"""Redirect the user to the terms and conditions accept page."""
redirect_url_parts = list(urlparse(ACCEPT_TERMS_PATH))
if slug != 'default':
redirect_url_parts[2] += slug
querystring = QueryDict(redirect_url_parts[4], mutable=True)
querystring[TERMS_RETURNTO_PARAM] = current_path
redirect_url_parts[4] = querystring.urlencode(safe='/')
return HttpResponseRedirect(urlunparse(redirect_url_parts)) | python | def redirect_to_terms_accept(current_path='/', slug='default'):
"""Redirect the user to the terms and conditions accept page."""
redirect_url_parts = list(urlparse(ACCEPT_TERMS_PATH))
if slug != 'default':
redirect_url_parts[2] += slug
querystring = QueryDict(redirect_url_parts[4], mutable=True)
querystring[TERMS_RETURNTO_PARAM] = current_path
redirect_url_parts[4] = querystring.urlencode(safe='/')
return HttpResponseRedirect(urlunparse(redirect_url_parts)) | [
"def",
"redirect_to_terms_accept",
"(",
"current_path",
"=",
"'/'",
",",
"slug",
"=",
"'default'",
")",
":",
"redirect_url_parts",
"=",
"list",
"(",
"urlparse",
"(",
"ACCEPT_TERMS_PATH",
")",
")",
"if",
"slug",
"!=",
"'default'",
":",
"redirect_url_parts",
"[",
"2",
"]",
"+=",
"slug",
"querystring",
"=",
"QueryDict",
"(",
"redirect_url_parts",
"[",
"4",
"]",
",",
"mutable",
"=",
"True",
")",
"querystring",
"[",
"TERMS_RETURNTO_PARAM",
"]",
"=",
"current_path",
"redirect_url_parts",
"[",
"4",
"]",
"=",
"querystring",
".",
"urlencode",
"(",
"safe",
"=",
"'/'",
")",
"return",
"HttpResponseRedirect",
"(",
"urlunparse",
"(",
"redirect_url_parts",
")",
")"
] | Redirect the user to the terms and conditions accept page. | [
"Redirect",
"the",
"user",
"to",
"the",
"terms",
"and",
"conditions",
"accept",
"page",
"."
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/pipeline.py#L28-L36 | train |
cyface/django-termsandconditions | termsandconditions/signals.py | user_terms_updated | def user_terms_updated(sender, **kwargs):
"""Called when user terms and conditions is changed - to force cache clearing"""
LOGGER.debug("User T&C Updated Signal Handler")
if kwargs.get('instance').user:
cache.delete('tandc.not_agreed_terms_' + kwargs.get('instance').user.get_username()) | python | def user_terms_updated(sender, **kwargs):
"""Called when user terms and conditions is changed - to force cache clearing"""
LOGGER.debug("User T&C Updated Signal Handler")
if kwargs.get('instance').user:
cache.delete('tandc.not_agreed_terms_' + kwargs.get('instance').user.get_username()) | [
"def",
"user_terms_updated",
"(",
"sender",
",",
"*",
"*",
"kwargs",
")",
":",
"LOGGER",
".",
"debug",
"(",
"\"User T&C Updated Signal Handler\"",
")",
"if",
"kwargs",
".",
"get",
"(",
"'instance'",
")",
".",
"user",
":",
"cache",
".",
"delete",
"(",
"'tandc.not_agreed_terms_'",
"+",
"kwargs",
".",
"get",
"(",
"'instance'",
")",
".",
"user",
".",
"get_username",
"(",
")",
")"
] | Called when user terms and conditions is changed - to force cache clearing | [
"Called",
"when",
"user",
"terms",
"and",
"conditions",
"is",
"changed",
"-",
"to",
"force",
"cache",
"clearing"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/signals.py#L15-L19 | train |
cyface/django-termsandconditions | termsandconditions/signals.py | terms_updated | def terms_updated(sender, **kwargs):
"""Called when terms and conditions is changed - to force cache clearing"""
LOGGER.debug("T&C Updated Signal Handler")
cache.delete('tandc.active_terms_ids')
cache.delete('tandc.active_terms_list')
if kwargs.get('instance').slug:
cache.delete('tandc.active_terms_' + kwargs.get('instance').slug)
for utandc in UserTermsAndConditions.objects.all():
cache.delete('tandc.not_agreed_terms_' + utandc.user.get_username()) | python | def terms_updated(sender, **kwargs):
"""Called when terms and conditions is changed - to force cache clearing"""
LOGGER.debug("T&C Updated Signal Handler")
cache.delete('tandc.active_terms_ids')
cache.delete('tandc.active_terms_list')
if kwargs.get('instance').slug:
cache.delete('tandc.active_terms_' + kwargs.get('instance').slug)
for utandc in UserTermsAndConditions.objects.all():
cache.delete('tandc.not_agreed_terms_' + utandc.user.get_username()) | [
"def",
"terms_updated",
"(",
"sender",
",",
"*",
"*",
"kwargs",
")",
":",
"LOGGER",
".",
"debug",
"(",
"\"T&C Updated Signal Handler\"",
")",
"cache",
".",
"delete",
"(",
"'tandc.active_terms_ids'",
")",
"cache",
".",
"delete",
"(",
"'tandc.active_terms_list'",
")",
"if",
"kwargs",
".",
"get",
"(",
"'instance'",
")",
".",
"slug",
":",
"cache",
".",
"delete",
"(",
"'tandc.active_terms_'",
"+",
"kwargs",
".",
"get",
"(",
"'instance'",
")",
".",
"slug",
")",
"for",
"utandc",
"in",
"UserTermsAndConditions",
".",
"objects",
".",
"all",
"(",
")",
":",
"cache",
".",
"delete",
"(",
"'tandc.not_agreed_terms_'",
"+",
"utandc",
".",
"user",
".",
"get_username",
"(",
")",
")"
] | Called when terms and conditions is changed - to force cache clearing | [
"Called",
"when",
"terms",
"and",
"conditions",
"is",
"changed",
"-",
"to",
"force",
"cache",
"clearing"
] | e18f06d0bad1e047f99222d1153f6e2b3bd5224f | https://github.com/cyface/django-termsandconditions/blob/e18f06d0bad1e047f99222d1153f6e2b3bd5224f/termsandconditions/signals.py#L23-L31 | train |
shtalinberg/django-el-pagination | el_pagination/templatetags/el_pagination_tags.py | paginate | def paginate(parser, token, paginator_class=None):
"""Paginate objects.
Usage:
.. code-block:: html+django
{% paginate entries %}
After this call, the *entries* variable in the template context is replaced
by only the entries of the current page.
You can also keep your *entries* original variable (usually a queryset)
and add to the context another name that refers to entries of the current
page, e.g.:
.. code-block:: html+django
{% paginate entries as page_entries %}
The *as* argument is also useful when a nested context variable is provided
as queryset. In this case, and only in this case, the resulting variable
name is mandatory, e.g.:
.. code-block:: html+django
{% paginate entries.all as entries %}
The number of paginated entries is taken from settings, but you can
override the default locally, e.g.:
.. code-block:: html+django
{% paginate 20 entries %}
Of course you can mix it all:
.. code-block:: html+django
{% paginate 20 entries as paginated_entries %}
By default, the first page is displayed the first time you load the page,
but you can change this, e.g.:
.. code-block:: html+django
{% paginate entries starting from page 3 %}
When changing the default page, it is also possible to reference the last
page (or the second last page, and so on) by using negative indexes, e.g:
.. code-block:: html+django
{% paginate entries starting from page -1 %}
This can be also achieved using a template variable that was passed to the
context, e.g.:
.. code-block:: html+django
{% paginate entries starting from page page_number %}
If the passed page number does not exist, the first page is displayed.
If you have multiple paginations in the same page, you can change the
querydict key for the single pagination, e.g.:
.. code-block:: html+django
{% paginate entries using article_page %}
In this case *article_page* is intended to be a context variable, but you
can hardcode the key using quotes, e.g.:
.. code-block:: html+django
{% paginate entries using 'articles_at_page' %}
Again, you can mix it all (the order of arguments is important):
.. code-block:: html+django
{% paginate 20 entries
starting from page 3 using page_key as paginated_entries %}
Additionally you can pass a path to be used for the pagination:
.. code-block:: html+django
{% paginate 20 entries
using page_key with pagination_url as paginated_entries %}
This way you can easily create views acting as API endpoints, and point
your Ajax calls to that API. In this case *pagination_url* is considered a
context variable, but it is also possible to hardcode the URL, e.g.:
.. code-block:: html+django
{% paginate 20 entries with "/mypage/" %}
If you want the first page to contain a different number of items than
subsequent pages, you can separate the two values with a comma, e.g. if
you want 3 items on the first page and 10 on other pages:
.. code-block:: html+django
{% paginate 3,10 entries %}
You must use this tag before calling the {% show_more %} one.
"""
# Validate arguments.
try:
tag_name, tag_args = token.contents.split(None, 1)
except ValueError:
msg = '%r tag requires arguments' % token.contents.split()[0]
raise template.TemplateSyntaxError(msg)
# Use a regexp to catch args.
match = PAGINATE_EXPRESSION.match(tag_args)
if match is None:
msg = 'Invalid arguments for %r tag' % tag_name
raise template.TemplateSyntaxError(msg)
# Retrieve objects.
kwargs = match.groupdict()
objects = kwargs.pop('objects')
# The variable name must be present if a nested context variable is passed.
if '.' in objects and kwargs['var_name'] is None:
msg = (
'%(tag)r tag requires a variable name `as` argumnent if the '
'queryset is provided as a nested context variable (%(objects)s). '
'You must either pass a direct queryset (e.g. taking advantage '
'of the `with` template tag) or provide a new variable name to '
'store the resulting queryset (e.g. `%(tag)s %(objects)s as '
'objects`).'
) % {'tag': tag_name, 'objects': objects}
raise template.TemplateSyntaxError(msg)
# Call the node.
return PaginateNode(paginator_class, objects, **kwargs) | python | def paginate(parser, token, paginator_class=None):
"""Paginate objects.
Usage:
.. code-block:: html+django
{% paginate entries %}
After this call, the *entries* variable in the template context is replaced
by only the entries of the current page.
You can also keep your *entries* original variable (usually a queryset)
and add to the context another name that refers to entries of the current
page, e.g.:
.. code-block:: html+django
{% paginate entries as page_entries %}
The *as* argument is also useful when a nested context variable is provided
as queryset. In this case, and only in this case, the resulting variable
name is mandatory, e.g.:
.. code-block:: html+django
{% paginate entries.all as entries %}
The number of paginated entries is taken from settings, but you can
override the default locally, e.g.:
.. code-block:: html+django
{% paginate 20 entries %}
Of course you can mix it all:
.. code-block:: html+django
{% paginate 20 entries as paginated_entries %}
By default, the first page is displayed the first time you load the page,
but you can change this, e.g.:
.. code-block:: html+django
{% paginate entries starting from page 3 %}
When changing the default page, it is also possible to reference the last
page (or the second last page, and so on) by using negative indexes, e.g:
.. code-block:: html+django
{% paginate entries starting from page -1 %}
This can be also achieved using a template variable that was passed to the
context, e.g.:
.. code-block:: html+django
{% paginate entries starting from page page_number %}
If the passed page number does not exist, the first page is displayed.
If you have multiple paginations in the same page, you can change the
querydict key for the single pagination, e.g.:
.. code-block:: html+django
{% paginate entries using article_page %}
In this case *article_page* is intended to be a context variable, but you
can hardcode the key using quotes, e.g.:
.. code-block:: html+django
{% paginate entries using 'articles_at_page' %}
Again, you can mix it all (the order of arguments is important):
.. code-block:: html+django
{% paginate 20 entries
starting from page 3 using page_key as paginated_entries %}
Additionally you can pass a path to be used for the pagination:
.. code-block:: html+django
{% paginate 20 entries
using page_key with pagination_url as paginated_entries %}
This way you can easily create views acting as API endpoints, and point
your Ajax calls to that API. In this case *pagination_url* is considered a
context variable, but it is also possible to hardcode the URL, e.g.:
.. code-block:: html+django
{% paginate 20 entries with "/mypage/" %}
If you want the first page to contain a different number of items than
subsequent pages, you can separate the two values with a comma, e.g. if
you want 3 items on the first page and 10 on other pages:
.. code-block:: html+django
{% paginate 3,10 entries %}
You must use this tag before calling the {% show_more %} one.
"""
# Validate arguments.
try:
tag_name, tag_args = token.contents.split(None, 1)
except ValueError:
msg = '%r tag requires arguments' % token.contents.split()[0]
raise template.TemplateSyntaxError(msg)
# Use a regexp to catch args.
match = PAGINATE_EXPRESSION.match(tag_args)
if match is None:
msg = 'Invalid arguments for %r tag' % tag_name
raise template.TemplateSyntaxError(msg)
# Retrieve objects.
kwargs = match.groupdict()
objects = kwargs.pop('objects')
# The variable name must be present if a nested context variable is passed.
if '.' in objects and kwargs['var_name'] is None:
msg = (
'%(tag)r tag requires a variable name `as` argumnent if the '
'queryset is provided as a nested context variable (%(objects)s). '
'You must either pass a direct queryset (e.g. taking advantage '
'of the `with` template tag) or provide a new variable name to '
'store the resulting queryset (e.g. `%(tag)s %(objects)s as '
'objects`).'
) % {'tag': tag_name, 'objects': objects}
raise template.TemplateSyntaxError(msg)
# Call the node.
return PaginateNode(paginator_class, objects, **kwargs) | [
"def",
"paginate",
"(",
"parser",
",",
"token",
",",
"paginator_class",
"=",
"None",
")",
":",
"# Validate arguments.",
"try",
":",
"tag_name",
",",
"tag_args",
"=",
"token",
".",
"contents",
".",
"split",
"(",
"None",
",",
"1",
")",
"except",
"ValueError",
":",
"msg",
"=",
"'%r tag requires arguments'",
"%",
"token",
".",
"contents",
".",
"split",
"(",
")",
"[",
"0",
"]",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"msg",
")",
"# Use a regexp to catch args.",
"match",
"=",
"PAGINATE_EXPRESSION",
".",
"match",
"(",
"tag_args",
")",
"if",
"match",
"is",
"None",
":",
"msg",
"=",
"'Invalid arguments for %r tag'",
"%",
"tag_name",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"msg",
")",
"# Retrieve objects.",
"kwargs",
"=",
"match",
".",
"groupdict",
"(",
")",
"objects",
"=",
"kwargs",
".",
"pop",
"(",
"'objects'",
")",
"# The variable name must be present if a nested context variable is passed.",
"if",
"'.'",
"in",
"objects",
"and",
"kwargs",
"[",
"'var_name'",
"]",
"is",
"None",
":",
"msg",
"=",
"(",
"'%(tag)r tag requires a variable name `as` argumnent if the '",
"'queryset is provided as a nested context variable (%(objects)s). '",
"'You must either pass a direct queryset (e.g. taking advantage '",
"'of the `with` template tag) or provide a new variable name to '",
"'store the resulting queryset (e.g. `%(tag)s %(objects)s as '",
"'objects`).'",
")",
"%",
"{",
"'tag'",
":",
"tag_name",
",",
"'objects'",
":",
"objects",
"}",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"msg",
")",
"# Call the node.",
"return",
"PaginateNode",
"(",
"paginator_class",
",",
"objects",
",",
"*",
"*",
"kwargs",
")"
] | Paginate objects.
Usage:
.. code-block:: html+django
{% paginate entries %}
After this call, the *entries* variable in the template context is replaced
by only the entries of the current page.
You can also keep your *entries* original variable (usually a queryset)
and add to the context another name that refers to entries of the current
page, e.g.:
.. code-block:: html+django
{% paginate entries as page_entries %}
The *as* argument is also useful when a nested context variable is provided
as queryset. In this case, and only in this case, the resulting variable
name is mandatory, e.g.:
.. code-block:: html+django
{% paginate entries.all as entries %}
The number of paginated entries is taken from settings, but you can
override the default locally, e.g.:
.. code-block:: html+django
{% paginate 20 entries %}
Of course you can mix it all:
.. code-block:: html+django
{% paginate 20 entries as paginated_entries %}
By default, the first page is displayed the first time you load the page,
but you can change this, e.g.:
.. code-block:: html+django
{% paginate entries starting from page 3 %}
When changing the default page, it is also possible to reference the last
page (or the second last page, and so on) by using negative indexes, e.g:
.. code-block:: html+django
{% paginate entries starting from page -1 %}
This can be also achieved using a template variable that was passed to the
context, e.g.:
.. code-block:: html+django
{% paginate entries starting from page page_number %}
If the passed page number does not exist, the first page is displayed.
If you have multiple paginations in the same page, you can change the
querydict key for the single pagination, e.g.:
.. code-block:: html+django
{% paginate entries using article_page %}
In this case *article_page* is intended to be a context variable, but you
can hardcode the key using quotes, e.g.:
.. code-block:: html+django
{% paginate entries using 'articles_at_page' %}
Again, you can mix it all (the order of arguments is important):
.. code-block:: html+django
{% paginate 20 entries
starting from page 3 using page_key as paginated_entries %}
Additionally you can pass a path to be used for the pagination:
.. code-block:: html+django
{% paginate 20 entries
using page_key with pagination_url as paginated_entries %}
This way you can easily create views acting as API endpoints, and point
your Ajax calls to that API. In this case *pagination_url* is considered a
context variable, but it is also possible to hardcode the URL, e.g.:
.. code-block:: html+django
{% paginate 20 entries with "/mypage/" %}
If you want the first page to contain a different number of items than
subsequent pages, you can separate the two values with a comma, e.g. if
you want 3 items on the first page and 10 on other pages:
.. code-block:: html+django
{% paginate 3,10 entries %}
You must use this tag before calling the {% show_more %} one. | [
"Paginate",
"objects",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/templatetags/el_pagination_tags.py#L45-L185 | train |
shtalinberg/django-el-pagination | el_pagination/templatetags/el_pagination_tags.py | get_pages | def get_pages(parser, token):
"""Add to context the list of page links.
Usage:
.. code-block:: html+django
{% get_pages %}
This is mostly used for Digg-style pagination.
This call inserts in the template context a *pages* variable, as a sequence
of page links. You can use *pages* in different ways:
- just print *pages.get_rendered* and you will get Digg-style pagination displayed:
.. code-block:: html+django
{{ pages.get_rendered }}
- display pages count:
.. code-block:: html+django
{{ pages|length }}
- check if the page list contains more than one page:
.. code-block:: html+django
{{ pages.paginated }}
{# the following is equivalent #}
{{ pages|length > 1 }}
- get a specific page:
.. code-block:: html+django
{# the current selected page #}
{{ pages.current }}
{# the first page #}
{{ pages.first }}
{# the last page #}
{{ pages.last }}
{# the previous page (or nothing if you are on first page) #}
{{ pages.previous }}
{# the next page (or nothing if you are in last page) #}
{{ pages.next }}
{# the third page #}
{{ pages.3 }}
{# this means page.1 is the same as page.first #}
{# the 1-based index of the first item on the current page #}
{{ pages.current_start_index }}
{# the 1-based index of the last item on the current page #}
{{ pages.current_end_index }}
{# the total number of objects, across all pages #}
{{ pages.total_count }}
{# the first page represented as an arrow #}
{{ pages.first_as_arrow }}
{# the last page represented as an arrow #}
{{ pages.last_as_arrow }}
- iterate over *pages* to get all pages:
.. code-block:: html+django
{% for page in pages %}
{# display page link #}
{{ page.render_link}}
{# the page url (beginning with "?") #}
{{ page.url }}
{# the page path #}
{{ page.path }}
{# the page number #}
{{ page.number }}
{# a string representing the page (commonly the page number) #}
{{ page.label }}
{# check if the page is the current one #}
{{ page.is_current }}
{# check if the page is the first one #}
{{ page.is_first }}
{# check if the page is the last one #}
{{ page.is_last }}
{% endfor %}
You can change the variable name, e.g.:
.. code-block:: html+django
{% get_pages as page_links %}
Must be called after ``{% paginate objects %}``.
"""
# Validate args.
try:
tag_name, args = token.contents.split(None, 1)
except ValueError:
var_name = 'pages'
else:
args = args.split()
if len(args) == 2 and args[0] == 'as':
var_name = args[1]
else:
msg = 'Invalid arguments for %r tag' % tag_name
raise template.TemplateSyntaxError(msg)
# Call the node.
return GetPagesNode(var_name) | python | def get_pages(parser, token):
"""Add to context the list of page links.
Usage:
.. code-block:: html+django
{% get_pages %}
This is mostly used for Digg-style pagination.
This call inserts in the template context a *pages* variable, as a sequence
of page links. You can use *pages* in different ways:
- just print *pages.get_rendered* and you will get Digg-style pagination displayed:
.. code-block:: html+django
{{ pages.get_rendered }}
- display pages count:
.. code-block:: html+django
{{ pages|length }}
- check if the page list contains more than one page:
.. code-block:: html+django
{{ pages.paginated }}
{# the following is equivalent #}
{{ pages|length > 1 }}
- get a specific page:
.. code-block:: html+django
{# the current selected page #}
{{ pages.current }}
{# the first page #}
{{ pages.first }}
{# the last page #}
{{ pages.last }}
{# the previous page (or nothing if you are on first page) #}
{{ pages.previous }}
{# the next page (or nothing if you are in last page) #}
{{ pages.next }}
{# the third page #}
{{ pages.3 }}
{# this means page.1 is the same as page.first #}
{# the 1-based index of the first item on the current page #}
{{ pages.current_start_index }}
{# the 1-based index of the last item on the current page #}
{{ pages.current_end_index }}
{# the total number of objects, across all pages #}
{{ pages.total_count }}
{# the first page represented as an arrow #}
{{ pages.first_as_arrow }}
{# the last page represented as an arrow #}
{{ pages.last_as_arrow }}
- iterate over *pages* to get all pages:
.. code-block:: html+django
{% for page in pages %}
{# display page link #}
{{ page.render_link}}
{# the page url (beginning with "?") #}
{{ page.url }}
{# the page path #}
{{ page.path }}
{# the page number #}
{{ page.number }}
{# a string representing the page (commonly the page number) #}
{{ page.label }}
{# check if the page is the current one #}
{{ page.is_current }}
{# check if the page is the first one #}
{{ page.is_first }}
{# check if the page is the last one #}
{{ page.is_last }}
{% endfor %}
You can change the variable name, e.g.:
.. code-block:: html+django
{% get_pages as page_links %}
Must be called after ``{% paginate objects %}``.
"""
# Validate args.
try:
tag_name, args = token.contents.split(None, 1)
except ValueError:
var_name = 'pages'
else:
args = args.split()
if len(args) == 2 and args[0] == 'as':
var_name = args[1]
else:
msg = 'Invalid arguments for %r tag' % tag_name
raise template.TemplateSyntaxError(msg)
# Call the node.
return GetPagesNode(var_name) | [
"def",
"get_pages",
"(",
"parser",
",",
"token",
")",
":",
"# Validate args.",
"try",
":",
"tag_name",
",",
"args",
"=",
"token",
".",
"contents",
".",
"split",
"(",
"None",
",",
"1",
")",
"except",
"ValueError",
":",
"var_name",
"=",
"'pages'",
"else",
":",
"args",
"=",
"args",
".",
"split",
"(",
")",
"if",
"len",
"(",
"args",
")",
"==",
"2",
"and",
"args",
"[",
"0",
"]",
"==",
"'as'",
":",
"var_name",
"=",
"args",
"[",
"1",
"]",
"else",
":",
"msg",
"=",
"'Invalid arguments for %r tag'",
"%",
"tag_name",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"msg",
")",
"# Call the node.",
"return",
"GetPagesNode",
"(",
"var_name",
")"
] | Add to context the list of page links.
Usage:
.. code-block:: html+django
{% get_pages %}
This is mostly used for Digg-style pagination.
This call inserts in the template context a *pages* variable, as a sequence
of page links. You can use *pages* in different ways:
- just print *pages.get_rendered* and you will get Digg-style pagination displayed:
.. code-block:: html+django
{{ pages.get_rendered }}
- display pages count:
.. code-block:: html+django
{{ pages|length }}
- check if the page list contains more than one page:
.. code-block:: html+django
{{ pages.paginated }}
{# the following is equivalent #}
{{ pages|length > 1 }}
- get a specific page:
.. code-block:: html+django
{# the current selected page #}
{{ pages.current }}
{# the first page #}
{{ pages.first }}
{# the last page #}
{{ pages.last }}
{# the previous page (or nothing if you are on first page) #}
{{ pages.previous }}
{# the next page (or nothing if you are in last page) #}
{{ pages.next }}
{# the third page #}
{{ pages.3 }}
{# this means page.1 is the same as page.first #}
{# the 1-based index of the first item on the current page #}
{{ pages.current_start_index }}
{# the 1-based index of the last item on the current page #}
{{ pages.current_end_index }}
{# the total number of objects, across all pages #}
{{ pages.total_count }}
{# the first page represented as an arrow #}
{{ pages.first_as_arrow }}
{# the last page represented as an arrow #}
{{ pages.last_as_arrow }}
- iterate over *pages* to get all pages:
.. code-block:: html+django
{% for page in pages %}
{# display page link #}
{{ page.render_link}}
{# the page url (beginning with "?") #}
{{ page.url }}
{# the page path #}
{{ page.path }}
{# the page number #}
{{ page.number }}
{# a string representing the page (commonly the page number) #}
{{ page.label }}
{# check if the page is the current one #}
{{ page.is_current }}
{# check if the page is the first one #}
{{ page.is_first }}
{# check if the page is the last one #}
{{ page.is_last }}
{% endfor %}
You can change the variable name, e.g.:
.. code-block:: html+django
{% get_pages as page_links %}
Must be called after ``{% paginate objects %}``. | [
"Add",
"to",
"context",
"the",
"list",
"of",
"page",
"links",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/templatetags/el_pagination_tags.py#L378-L500 | train |
shtalinberg/django-el-pagination | el_pagination/templatetags/el_pagination_tags.py | show_pages | def show_pages(parser, token):
"""Show page links.
Usage:
.. code-block:: html+django
{% show_pages %}
It is just a shortcut for:
.. code-block:: html+django
{% get_pages %}
{{ pages.get_rendered }}
You can set ``ENDLESS_PAGINATION_PAGE_LIST_CALLABLE`` in your *settings.py*
to a callable, or to a dotted path representing a callable, used to
customize the pages that are displayed.
See the *__unicode__* method of ``endless_pagination.models.PageList`` for
a detailed explanation of how the callable can be used.
Must be called after ``{% paginate objects %}``.
"""
# Validate args.
if len(token.contents.split()) != 1:
msg = '%r tag takes no arguments' % token.contents.split()[0]
raise template.TemplateSyntaxError(msg)
# Call the node.
return ShowPagesNode() | python | def show_pages(parser, token):
"""Show page links.
Usage:
.. code-block:: html+django
{% show_pages %}
It is just a shortcut for:
.. code-block:: html+django
{% get_pages %}
{{ pages.get_rendered }}
You can set ``ENDLESS_PAGINATION_PAGE_LIST_CALLABLE`` in your *settings.py*
to a callable, or to a dotted path representing a callable, used to
customize the pages that are displayed.
See the *__unicode__* method of ``endless_pagination.models.PageList`` for
a detailed explanation of how the callable can be used.
Must be called after ``{% paginate objects %}``.
"""
# Validate args.
if len(token.contents.split()) != 1:
msg = '%r tag takes no arguments' % token.contents.split()[0]
raise template.TemplateSyntaxError(msg)
# Call the node.
return ShowPagesNode() | [
"def",
"show_pages",
"(",
"parser",
",",
"token",
")",
":",
"# Validate args.",
"if",
"len",
"(",
"token",
".",
"contents",
".",
"split",
"(",
")",
")",
"!=",
"1",
":",
"msg",
"=",
"'%r tag takes no arguments'",
"%",
"token",
".",
"contents",
".",
"split",
"(",
")",
"[",
"0",
"]",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"msg",
")",
"# Call the node.",
"return",
"ShowPagesNode",
"(",
")"
] | Show page links.
Usage:
.. code-block:: html+django
{% show_pages %}
It is just a shortcut for:
.. code-block:: html+django
{% get_pages %}
{{ pages.get_rendered }}
You can set ``ENDLESS_PAGINATION_PAGE_LIST_CALLABLE`` in your *settings.py*
to a callable, or to a dotted path representing a callable, used to
customize the pages that are displayed.
See the *__unicode__* method of ``endless_pagination.models.PageList`` for
a detailed explanation of how the callable can be used.
Must be called after ``{% paginate objects %}``. | [
"Show",
"page",
"links",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/templatetags/el_pagination_tags.py#L526-L556 | train |
shtalinberg/django-el-pagination | el_pagination/templatetags/el_pagination_tags.py | show_current_number | def show_current_number(parser, token):
"""Show the current page number, or insert it in the context.
This tag can for example be useful to change the page title according to
the current page number.
To just show current page number:
.. code-block:: html+django
{% show_current_number %}
If you use multiple paginations in the same page, you can get the page
number for a specific pagination using the querystring key, e.g.:
.. code-block:: html+django
{% show_current_number using mykey %}
The default page when no querystring is specified is 1. If you changed it
in the `paginate`_ template tag, you have to call ``show_current_number``
according to your choice, e.g.:
.. code-block:: html+django
{% show_current_number starting from page 3 %}
This can be also achieved using a template variable you passed to the
context, e.g.:
.. code-block:: html+django
{% show_current_number starting from page page_number %}
You can of course mix it all (the order of arguments is important):
.. code-block:: html+django
{% show_current_number starting from page 3 using mykey %}
If you want to insert the current page number in the context, without
actually displaying it in the template, use the *as* argument, i.e.:
.. code-block:: html+django
{% show_current_number as page_number %}
{% show_current_number
starting from page 3 using mykey as page_number %}
"""
# Validate args.
try:
tag_name, args = token.contents.split(None, 1)
except ValueError:
key = None
number = None
tag_name = token.contents[0]
var_name = None
else:
# Use a regexp to catch args.
match = SHOW_CURRENT_NUMBER_EXPRESSION.match(args)
if match is None:
msg = 'Invalid arguments for %r tag' % tag_name
raise template.TemplateSyntaxError(msg)
# Retrieve objects.
groupdict = match.groupdict()
key = groupdict['key']
number = groupdict['number']
var_name = groupdict['var_name']
# Call the node.
return ShowCurrentNumberNode(number, key, var_name) | python | def show_current_number(parser, token):
"""Show the current page number, or insert it in the context.
This tag can for example be useful to change the page title according to
the current page number.
To just show current page number:
.. code-block:: html+django
{% show_current_number %}
If you use multiple paginations in the same page, you can get the page
number for a specific pagination using the querystring key, e.g.:
.. code-block:: html+django
{% show_current_number using mykey %}
The default page when no querystring is specified is 1. If you changed it
in the `paginate`_ template tag, you have to call ``show_current_number``
according to your choice, e.g.:
.. code-block:: html+django
{% show_current_number starting from page 3 %}
This can be also achieved using a template variable you passed to the
context, e.g.:
.. code-block:: html+django
{% show_current_number starting from page page_number %}
You can of course mix it all (the order of arguments is important):
.. code-block:: html+django
{% show_current_number starting from page 3 using mykey %}
If you want to insert the current page number in the context, without
actually displaying it in the template, use the *as* argument, i.e.:
.. code-block:: html+django
{% show_current_number as page_number %}
{% show_current_number
starting from page 3 using mykey as page_number %}
"""
# Validate args.
try:
tag_name, args = token.contents.split(None, 1)
except ValueError:
key = None
number = None
tag_name = token.contents[0]
var_name = None
else:
# Use a regexp to catch args.
match = SHOW_CURRENT_NUMBER_EXPRESSION.match(args)
if match is None:
msg = 'Invalid arguments for %r tag' % tag_name
raise template.TemplateSyntaxError(msg)
# Retrieve objects.
groupdict = match.groupdict()
key = groupdict['key']
number = groupdict['number']
var_name = groupdict['var_name']
# Call the node.
return ShowCurrentNumberNode(number, key, var_name) | [
"def",
"show_current_number",
"(",
"parser",
",",
"token",
")",
":",
"# Validate args.",
"try",
":",
"tag_name",
",",
"args",
"=",
"token",
".",
"contents",
".",
"split",
"(",
"None",
",",
"1",
")",
"except",
"ValueError",
":",
"key",
"=",
"None",
"number",
"=",
"None",
"tag_name",
"=",
"token",
".",
"contents",
"[",
"0",
"]",
"var_name",
"=",
"None",
"else",
":",
"# Use a regexp to catch args.",
"match",
"=",
"SHOW_CURRENT_NUMBER_EXPRESSION",
".",
"match",
"(",
"args",
")",
"if",
"match",
"is",
"None",
":",
"msg",
"=",
"'Invalid arguments for %r tag'",
"%",
"tag_name",
"raise",
"template",
".",
"TemplateSyntaxError",
"(",
"msg",
")",
"# Retrieve objects.",
"groupdict",
"=",
"match",
".",
"groupdict",
"(",
")",
"key",
"=",
"groupdict",
"[",
"'key'",
"]",
"number",
"=",
"groupdict",
"[",
"'number'",
"]",
"var_name",
"=",
"groupdict",
"[",
"'var_name'",
"]",
"# Call the node.",
"return",
"ShowCurrentNumberNode",
"(",
"number",
",",
"key",
",",
"var_name",
")"
] | Show the current page number, or insert it in the context.
This tag can for example be useful to change the page title according to
the current page number.
To just show current page number:
.. code-block:: html+django
{% show_current_number %}
If you use multiple paginations in the same page, you can get the page
number for a specific pagination using the querystring key, e.g.:
.. code-block:: html+django
{% show_current_number using mykey %}
The default page when no querystring is specified is 1. If you changed it
in the `paginate`_ template tag, you have to call ``show_current_number``
according to your choice, e.g.:
.. code-block:: html+django
{% show_current_number starting from page 3 %}
This can be also achieved using a template variable you passed to the
context, e.g.:
.. code-block:: html+django
{% show_current_number starting from page page_number %}
You can of course mix it all (the order of arguments is important):
.. code-block:: html+django
{% show_current_number starting from page 3 using mykey %}
If you want to insert the current page number in the context, without
actually displaying it in the template, use the *as* argument, i.e.:
.. code-block:: html+django
{% show_current_number as page_number %}
{% show_current_number
starting from page 3 using mykey as page_number %} | [
"Show",
"the",
"current",
"page",
"number",
"or",
"insert",
"it",
"in",
"the",
"context",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/templatetags/el_pagination_tags.py#L579-L649 | train |
shtalinberg/django-el-pagination | el_pagination/decorators.py | page_template | def page_template(template, key=PAGE_LABEL):
"""Return a view dynamically switching template if the request is Ajax.
Decorate a view that takes a *template* and *extra_context* keyword
arguments (like generic views).
The template is switched to *page_template* if request is ajax and
if *querystring_key* variable passed by the request equals to *key*.
This allows multiple Ajax paginations in the same page.
The name of the page template is given as *page_template* in the
extra context.
"""
def decorator(view):
@wraps(view)
def decorated(request, *args, **kwargs):
# Trust the developer: he wrote ``context.update(extra_context)``
# in his view.
extra_context = kwargs.setdefault('extra_context', {})
extra_context['page_template'] = template
# Switch the template when the request is Ajax.
querystring_key = request.GET.get(QS_KEY,
request.POST.get(QS_KEY, PAGE_LABEL))
if request.is_ajax() and querystring_key == key:
kwargs[TEMPLATE_VARNAME] = template
return view(request, *args, **kwargs)
return decorated
return decorator | python | def page_template(template, key=PAGE_LABEL):
"""Return a view dynamically switching template if the request is Ajax.
Decorate a view that takes a *template* and *extra_context* keyword
arguments (like generic views).
The template is switched to *page_template* if request is ajax and
if *querystring_key* variable passed by the request equals to *key*.
This allows multiple Ajax paginations in the same page.
The name of the page template is given as *page_template* in the
extra context.
"""
def decorator(view):
@wraps(view)
def decorated(request, *args, **kwargs):
# Trust the developer: he wrote ``context.update(extra_context)``
# in his view.
extra_context = kwargs.setdefault('extra_context', {})
extra_context['page_template'] = template
# Switch the template when the request is Ajax.
querystring_key = request.GET.get(QS_KEY,
request.POST.get(QS_KEY, PAGE_LABEL))
if request.is_ajax() and querystring_key == key:
kwargs[TEMPLATE_VARNAME] = template
return view(request, *args, **kwargs)
return decorated
return decorator | [
"def",
"page_template",
"(",
"template",
",",
"key",
"=",
"PAGE_LABEL",
")",
":",
"def",
"decorator",
"(",
"view",
")",
":",
"@",
"wraps",
"(",
"view",
")",
"def",
"decorated",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"# Trust the developer: he wrote ``context.update(extra_context)``",
"# in his view.",
"extra_context",
"=",
"kwargs",
".",
"setdefault",
"(",
"'extra_context'",
",",
"{",
"}",
")",
"extra_context",
"[",
"'page_template'",
"]",
"=",
"template",
"# Switch the template when the request is Ajax.",
"querystring_key",
"=",
"request",
".",
"GET",
".",
"get",
"(",
"QS_KEY",
",",
"request",
".",
"POST",
".",
"get",
"(",
"QS_KEY",
",",
"PAGE_LABEL",
")",
")",
"if",
"request",
".",
"is_ajax",
"(",
")",
"and",
"querystring_key",
"==",
"key",
":",
"kwargs",
"[",
"TEMPLATE_VARNAME",
"]",
"=",
"template",
"return",
"view",
"(",
"request",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
"return",
"decorated",
"return",
"decorator"
] | Return a view dynamically switching template if the request is Ajax.
Decorate a view that takes a *template* and *extra_context* keyword
arguments (like generic views).
The template is switched to *page_template* if request is ajax and
if *querystring_key* variable passed by the request equals to *key*.
This allows multiple Ajax paginations in the same page.
The name of the page template is given as *page_template* in the
extra context. | [
"Return",
"a",
"view",
"dynamically",
"switching",
"template",
"if",
"the",
"request",
"is",
"Ajax",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/decorators.py#L13-L39 | train |
shtalinberg/django-el-pagination | el_pagination/decorators.py | _get_template | def _get_template(querystring_key, mapping):
"""Return the template corresponding to the given ``querystring_key``."""
default = None
try:
template_and_keys = mapping.items()
except AttributeError:
template_and_keys = mapping
for template, key in template_and_keys:
if key is None:
key = PAGE_LABEL
default = template
if key == querystring_key:
return template
return default | python | def _get_template(querystring_key, mapping):
"""Return the template corresponding to the given ``querystring_key``."""
default = None
try:
template_and_keys = mapping.items()
except AttributeError:
template_and_keys = mapping
for template, key in template_and_keys:
if key is None:
key = PAGE_LABEL
default = template
if key == querystring_key:
return template
return default | [
"def",
"_get_template",
"(",
"querystring_key",
",",
"mapping",
")",
":",
"default",
"=",
"None",
"try",
":",
"template_and_keys",
"=",
"mapping",
".",
"items",
"(",
")",
"except",
"AttributeError",
":",
"template_and_keys",
"=",
"mapping",
"for",
"template",
",",
"key",
"in",
"template_and_keys",
":",
"if",
"key",
"is",
"None",
":",
"key",
"=",
"PAGE_LABEL",
"default",
"=",
"template",
"if",
"key",
"==",
"querystring_key",
":",
"return",
"template",
"return",
"default"
] | Return the template corresponding to the given ``querystring_key``. | [
"Return",
"the",
"template",
"corresponding",
"to",
"the",
"given",
"querystring_key",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/decorators.py#L42-L55 | train |
shtalinberg/django-el-pagination | el_pagination/views.py | MultipleObjectMixin.get_queryset | def get_queryset(self):
"""Get the list of items for this view.
This must be an interable, and may be a queryset
(in which qs-specific behavior will be enabled).
See original in ``django.views.generic.list.MultipleObjectMixin``.
"""
if self.queryset is not None:
queryset = self.queryset
if hasattr(queryset, '_clone'):
queryset = queryset._clone()
elif self.model is not None:
queryset = self.model._default_manager.all()
else:
msg = '{0} must define ``queryset`` or ``model``'
raise ImproperlyConfigured(msg.format(self.__class__.__name__))
return queryset | python | def get_queryset(self):
"""Get the list of items for this view.
This must be an interable, and may be a queryset
(in which qs-specific behavior will be enabled).
See original in ``django.views.generic.list.MultipleObjectMixin``.
"""
if self.queryset is not None:
queryset = self.queryset
if hasattr(queryset, '_clone'):
queryset = queryset._clone()
elif self.model is not None:
queryset = self.model._default_manager.all()
else:
msg = '{0} must define ``queryset`` or ``model``'
raise ImproperlyConfigured(msg.format(self.__class__.__name__))
return queryset | [
"def",
"get_queryset",
"(",
"self",
")",
":",
"if",
"self",
".",
"queryset",
"is",
"not",
"None",
":",
"queryset",
"=",
"self",
".",
"queryset",
"if",
"hasattr",
"(",
"queryset",
",",
"'_clone'",
")",
":",
"queryset",
"=",
"queryset",
".",
"_clone",
"(",
")",
"elif",
"self",
".",
"model",
"is",
"not",
"None",
":",
"queryset",
"=",
"self",
".",
"model",
".",
"_default_manager",
".",
"all",
"(",
")",
"else",
":",
"msg",
"=",
"'{0} must define ``queryset`` or ``model``'",
"raise",
"ImproperlyConfigured",
"(",
"msg",
".",
"format",
"(",
"self",
".",
"__class__",
".",
"__name__",
")",
")",
"return",
"queryset"
] | Get the list of items for this view.
This must be an interable, and may be a queryset
(in which qs-specific behavior will be enabled).
See original in ``django.views.generic.list.MultipleObjectMixin``. | [
"Get",
"the",
"list",
"of",
"items",
"for",
"this",
"view",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/views.py#L22-L39 | train |
shtalinberg/django-el-pagination | el_pagination/views.py | MultipleObjectMixin.get_context_object_name | def get_context_object_name(self, object_list):
"""Get the name of the item to be used in the context.
See original in ``django.views.generic.list.MultipleObjectMixin``.
"""
if self.context_object_name:
return self.context_object_name
elif hasattr(object_list, 'model'):
object_name = object_list.model._meta.object_name.lower()
return smart_str('{0}_list'.format(object_name))
else:
return None | python | def get_context_object_name(self, object_list):
"""Get the name of the item to be used in the context.
See original in ``django.views.generic.list.MultipleObjectMixin``.
"""
if self.context_object_name:
return self.context_object_name
elif hasattr(object_list, 'model'):
object_name = object_list.model._meta.object_name.lower()
return smart_str('{0}_list'.format(object_name))
else:
return None | [
"def",
"get_context_object_name",
"(",
"self",
",",
"object_list",
")",
":",
"if",
"self",
".",
"context_object_name",
":",
"return",
"self",
".",
"context_object_name",
"elif",
"hasattr",
"(",
"object_list",
",",
"'model'",
")",
":",
"object_name",
"=",
"object_list",
".",
"model",
".",
"_meta",
".",
"object_name",
".",
"lower",
"(",
")",
"return",
"smart_str",
"(",
"'{0}_list'",
".",
"format",
"(",
"object_name",
")",
")",
"else",
":",
"return",
"None"
] | Get the name of the item to be used in the context.
See original in ``django.views.generic.list.MultipleObjectMixin``. | [
"Get",
"the",
"name",
"of",
"the",
"item",
"to",
"be",
"used",
"in",
"the",
"context",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/views.py#L50-L61 | train |
shtalinberg/django-el-pagination | el_pagination/views.py | AjaxMultipleObjectTemplateResponseMixin.get_page_template | def get_page_template(self, **kwargs):
"""Return the template name used for this request.
Only called if *page_template* is not given as a kwarg of
*self.as_view*.
"""
opts = self.object_list.model._meta
return '{0}/{1}{2}{3}.html'.format(
opts.app_label,
opts.object_name.lower(),
self.template_name_suffix,
self.page_template_suffix,
) | python | def get_page_template(self, **kwargs):
"""Return the template name used for this request.
Only called if *page_template* is not given as a kwarg of
*self.as_view*.
"""
opts = self.object_list.model._meta
return '{0}/{1}{2}{3}.html'.format(
opts.app_label,
opts.object_name.lower(),
self.template_name_suffix,
self.page_template_suffix,
) | [
"def",
"get_page_template",
"(",
"self",
",",
"*",
"*",
"kwargs",
")",
":",
"opts",
"=",
"self",
".",
"object_list",
".",
"model",
".",
"_meta",
"return",
"'{0}/{1}{2}{3}.html'",
".",
"format",
"(",
"opts",
".",
"app_label",
",",
"opts",
".",
"object_name",
".",
"lower",
"(",
")",
",",
"self",
".",
"template_name_suffix",
",",
"self",
".",
"page_template_suffix",
",",
")"
] | Return the template name used for this request.
Only called if *page_template* is not given as a kwarg of
*self.as_view*. | [
"Return",
"the",
"template",
"name",
"used",
"for",
"this",
"request",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/views.py#L135-L147 | train |
shtalinberg/django-el-pagination | el_pagination/models.py | ELPage.render_link | def render_link(self):
"""Render the page as a link."""
extra_context = {
'add_nofollow': settings.ADD_NOFOLLOW,
'page': self,
'querystring_key': self.querystring_key,
}
if self.is_current:
template_name = 'el_pagination/current_link.html'
else:
template_name = 'el_pagination/page_link.html'
if settings.USE_NEXT_PREVIOUS_LINKS:
if self.is_previous:
template_name = 'el_pagination/previous_link.html'
if self.is_next:
template_name = 'el_pagination/next_link.html'
if template_name not in _template_cache:
_template_cache[template_name] = loader.get_template(template_name)
template = _template_cache[template_name]
with self.context.push(**extra_context):
return template.render(self.context.flatten()) | python | def render_link(self):
"""Render the page as a link."""
extra_context = {
'add_nofollow': settings.ADD_NOFOLLOW,
'page': self,
'querystring_key': self.querystring_key,
}
if self.is_current:
template_name = 'el_pagination/current_link.html'
else:
template_name = 'el_pagination/page_link.html'
if settings.USE_NEXT_PREVIOUS_LINKS:
if self.is_previous:
template_name = 'el_pagination/previous_link.html'
if self.is_next:
template_name = 'el_pagination/next_link.html'
if template_name not in _template_cache:
_template_cache[template_name] = loader.get_template(template_name)
template = _template_cache[template_name]
with self.context.push(**extra_context):
return template.render(self.context.flatten()) | [
"def",
"render_link",
"(",
"self",
")",
":",
"extra_context",
"=",
"{",
"'add_nofollow'",
":",
"settings",
".",
"ADD_NOFOLLOW",
",",
"'page'",
":",
"self",
",",
"'querystring_key'",
":",
"self",
".",
"querystring_key",
",",
"}",
"if",
"self",
".",
"is_current",
":",
"template_name",
"=",
"'el_pagination/current_link.html'",
"else",
":",
"template_name",
"=",
"'el_pagination/page_link.html'",
"if",
"settings",
".",
"USE_NEXT_PREVIOUS_LINKS",
":",
"if",
"self",
".",
"is_previous",
":",
"template_name",
"=",
"'el_pagination/previous_link.html'",
"if",
"self",
".",
"is_next",
":",
"template_name",
"=",
"'el_pagination/next_link.html'",
"if",
"template_name",
"not",
"in",
"_template_cache",
":",
"_template_cache",
"[",
"template_name",
"]",
"=",
"loader",
".",
"get_template",
"(",
"template_name",
")",
"template",
"=",
"_template_cache",
"[",
"template_name",
"]",
"with",
"self",
".",
"context",
".",
"push",
"(",
"*",
"*",
"extra_context",
")",
":",
"return",
"template",
".",
"render",
"(",
"self",
".",
"context",
".",
"flatten",
"(",
")",
")"
] | Render the page as a link. | [
"Render",
"the",
"page",
"as",
"a",
"link",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/models.py#L63-L83 | train |
shtalinberg/django-el-pagination | el_pagination/models.py | PageList.previous | def previous(self):
"""Return the previous page.
The page label is defined in ``settings.PREVIOUS_LABEL``.
Return an empty string if current page is the first.
"""
if self._page.has_previous():
return self._endless_page(
self._page.previous_page_number(),
label=settings.PREVIOUS_LABEL)
return '' | python | def previous(self):
"""Return the previous page.
The page label is defined in ``settings.PREVIOUS_LABEL``.
Return an empty string if current page is the first.
"""
if self._page.has_previous():
return self._endless_page(
self._page.previous_page_number(),
label=settings.PREVIOUS_LABEL)
return '' | [
"def",
"previous",
"(",
"self",
")",
":",
"if",
"self",
".",
"_page",
".",
"has_previous",
"(",
")",
":",
"return",
"self",
".",
"_endless_page",
"(",
"self",
".",
"_page",
".",
"previous_page_number",
"(",
")",
",",
"label",
"=",
"settings",
".",
"PREVIOUS_LABEL",
")",
"return",
"''"
] | Return the previous page.
The page label is defined in ``settings.PREVIOUS_LABEL``.
Return an empty string if current page is the first. | [
"Return",
"the",
"previous",
"page",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/models.py#L243-L253 | train |
shtalinberg/django-el-pagination | el_pagination/models.py | PageList.next | def next(self):
"""Return the next page.
The page label is defined in ``settings.NEXT_LABEL``.
Return an empty string if current page is the last.
"""
if self._page.has_next():
return self._endless_page(
self._page.next_page_number(),
label=settings.NEXT_LABEL)
return '' | python | def next(self):
"""Return the next page.
The page label is defined in ``settings.NEXT_LABEL``.
Return an empty string if current page is the last.
"""
if self._page.has_next():
return self._endless_page(
self._page.next_page_number(),
label=settings.NEXT_LABEL)
return '' | [
"def",
"next",
"(",
"self",
")",
":",
"if",
"self",
".",
"_page",
".",
"has_next",
"(",
")",
":",
"return",
"self",
".",
"_endless_page",
"(",
"self",
".",
"_page",
".",
"next_page_number",
"(",
")",
",",
"label",
"=",
"settings",
".",
"NEXT_LABEL",
")",
"return",
"''"
] | Return the next page.
The page label is defined in ``settings.NEXT_LABEL``.
Return an empty string if current page is the last. | [
"Return",
"the",
"next",
"page",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/models.py#L255-L265 | train |
shtalinberg/django-el-pagination | el_pagination/paginators.py | CustomPage.start_index | def start_index(self):
"""Return the 1-based index of the first item on this page."""
paginator = self.paginator
# Special case, return zero if no items.
if paginator.count == 0:
return 0
elif self.number == 1:
return 1
return (
(self.number - 2) * paginator.per_page + paginator.first_page + 1) | python | def start_index(self):
"""Return the 1-based index of the first item on this page."""
paginator = self.paginator
# Special case, return zero if no items.
if paginator.count == 0:
return 0
elif self.number == 1:
return 1
return (
(self.number - 2) * paginator.per_page + paginator.first_page + 1) | [
"def",
"start_index",
"(",
"self",
")",
":",
"paginator",
"=",
"self",
".",
"paginator",
"# Special case, return zero if no items.",
"if",
"paginator",
".",
"count",
"==",
"0",
":",
"return",
"0",
"elif",
"self",
".",
"number",
"==",
"1",
":",
"return",
"1",
"return",
"(",
"(",
"self",
".",
"number",
"-",
"2",
")",
"*",
"paginator",
".",
"per_page",
"+",
"paginator",
".",
"first_page",
"+",
"1",
")"
] | Return the 1-based index of the first item on this page. | [
"Return",
"the",
"1",
"-",
"based",
"index",
"of",
"the",
"first",
"item",
"on",
"this",
"page",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/paginators.py#L17-L26 | train |
shtalinberg/django-el-pagination | el_pagination/paginators.py | CustomPage.end_index | def end_index(self):
"""Return the 1-based index of the last item on this page."""
paginator = self.paginator
# Special case for the last page because there can be orphans.
if self.number == paginator.num_pages:
return paginator.count
return (self.number - 1) * paginator.per_page + paginator.first_page | python | def end_index(self):
"""Return the 1-based index of the last item on this page."""
paginator = self.paginator
# Special case for the last page because there can be orphans.
if self.number == paginator.num_pages:
return paginator.count
return (self.number - 1) * paginator.per_page + paginator.first_page | [
"def",
"end_index",
"(",
"self",
")",
":",
"paginator",
"=",
"self",
".",
"paginator",
"# Special case for the last page because there can be orphans.",
"if",
"self",
".",
"number",
"==",
"paginator",
".",
"num_pages",
":",
"return",
"paginator",
".",
"count",
"return",
"(",
"self",
".",
"number",
"-",
"1",
")",
"*",
"paginator",
".",
"per_page",
"+",
"paginator",
".",
"first_page"
] | Return the 1-based index of the last item on this page. | [
"Return",
"the",
"1",
"-",
"based",
"index",
"of",
"the",
"last",
"item",
"on",
"this",
"page",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/paginators.py#L28-L34 | train |
shtalinberg/django-el-pagination | el_pagination/utils.py | get_page_numbers | def get_page_numbers(
current_page, num_pages,
extremes=DEFAULT_CALLABLE_EXTREMES,
arounds=DEFAULT_CALLABLE_AROUNDS,
arrows=DEFAULT_CALLABLE_ARROWS):
"""Default callable for page listing.
Produce a Digg-style pagination.
"""
page_range = range(1, num_pages + 1)
pages = []
if current_page != 1:
if arrows:
pages.append('first')
pages.append('previous')
# Get first and last pages (extremes).
first = page_range[:extremes]
pages.extend(first)
last = page_range[-extremes:]
# Get the current pages (arounds).
current_start = current_page - 1 - arounds
if current_start < 0:
current_start = 0
current_end = current_page + arounds
if current_end > num_pages:
current_end = num_pages
current = page_range[current_start:current_end]
# Mix first with current pages.
to_add = current
if extremes:
diff = current[0] - first[-1]
if diff > 1:
pages.append(None)
elif diff < 1:
to_add = current[abs(diff) + 1:]
pages.extend(to_add)
# Mix current with last pages.
if extremes:
diff = last[0] - current[-1]
to_add = last
if diff > 1:
pages.append(None)
elif diff < 1:
to_add = last[abs(diff) + 1:]
pages.extend(to_add)
if current_page != num_pages:
pages.append('next')
if arrows:
pages.append('last')
return pages | python | def get_page_numbers(
current_page, num_pages,
extremes=DEFAULT_CALLABLE_EXTREMES,
arounds=DEFAULT_CALLABLE_AROUNDS,
arrows=DEFAULT_CALLABLE_ARROWS):
"""Default callable for page listing.
Produce a Digg-style pagination.
"""
page_range = range(1, num_pages + 1)
pages = []
if current_page != 1:
if arrows:
pages.append('first')
pages.append('previous')
# Get first and last pages (extremes).
first = page_range[:extremes]
pages.extend(first)
last = page_range[-extremes:]
# Get the current pages (arounds).
current_start = current_page - 1 - arounds
if current_start < 0:
current_start = 0
current_end = current_page + arounds
if current_end > num_pages:
current_end = num_pages
current = page_range[current_start:current_end]
# Mix first with current pages.
to_add = current
if extremes:
diff = current[0] - first[-1]
if diff > 1:
pages.append(None)
elif diff < 1:
to_add = current[abs(diff) + 1:]
pages.extend(to_add)
# Mix current with last pages.
if extremes:
diff = last[0] - current[-1]
to_add = last
if diff > 1:
pages.append(None)
elif diff < 1:
to_add = last[abs(diff) + 1:]
pages.extend(to_add)
if current_page != num_pages:
pages.append('next')
if arrows:
pages.append('last')
return pages | [
"def",
"get_page_numbers",
"(",
"current_page",
",",
"num_pages",
",",
"extremes",
"=",
"DEFAULT_CALLABLE_EXTREMES",
",",
"arounds",
"=",
"DEFAULT_CALLABLE_AROUNDS",
",",
"arrows",
"=",
"DEFAULT_CALLABLE_ARROWS",
")",
":",
"page_range",
"=",
"range",
"(",
"1",
",",
"num_pages",
"+",
"1",
")",
"pages",
"=",
"[",
"]",
"if",
"current_page",
"!=",
"1",
":",
"if",
"arrows",
":",
"pages",
".",
"append",
"(",
"'first'",
")",
"pages",
".",
"append",
"(",
"'previous'",
")",
"# Get first and last pages (extremes).",
"first",
"=",
"page_range",
"[",
":",
"extremes",
"]",
"pages",
".",
"extend",
"(",
"first",
")",
"last",
"=",
"page_range",
"[",
"-",
"extremes",
":",
"]",
"# Get the current pages (arounds).",
"current_start",
"=",
"current_page",
"-",
"1",
"-",
"arounds",
"if",
"current_start",
"<",
"0",
":",
"current_start",
"=",
"0",
"current_end",
"=",
"current_page",
"+",
"arounds",
"if",
"current_end",
">",
"num_pages",
":",
"current_end",
"=",
"num_pages",
"current",
"=",
"page_range",
"[",
"current_start",
":",
"current_end",
"]",
"# Mix first with current pages.",
"to_add",
"=",
"current",
"if",
"extremes",
":",
"diff",
"=",
"current",
"[",
"0",
"]",
"-",
"first",
"[",
"-",
"1",
"]",
"if",
"diff",
">",
"1",
":",
"pages",
".",
"append",
"(",
"None",
")",
"elif",
"diff",
"<",
"1",
":",
"to_add",
"=",
"current",
"[",
"abs",
"(",
"diff",
")",
"+",
"1",
":",
"]",
"pages",
".",
"extend",
"(",
"to_add",
")",
"# Mix current with last pages.",
"if",
"extremes",
":",
"diff",
"=",
"last",
"[",
"0",
"]",
"-",
"current",
"[",
"-",
"1",
"]",
"to_add",
"=",
"last",
"if",
"diff",
">",
"1",
":",
"pages",
".",
"append",
"(",
"None",
")",
"elif",
"diff",
"<",
"1",
":",
"to_add",
"=",
"last",
"[",
"abs",
"(",
"diff",
")",
"+",
"1",
":",
"]",
"pages",
".",
"extend",
"(",
"to_add",
")",
"if",
"current_page",
"!=",
"num_pages",
":",
"pages",
".",
"append",
"(",
"'next'",
")",
"if",
"arrows",
":",
"pages",
".",
"append",
"(",
"'last'",
")",
"return",
"pages"
] | Default callable for page listing.
Produce a Digg-style pagination. | [
"Default",
"callable",
"for",
"page",
"listing",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/utils.py#L50-L104 | train |
shtalinberg/django-el-pagination | el_pagination/utils.py | _make_elastic_range | def _make_elastic_range(begin, end):
"""Generate an S-curved range of pages.
Start from both left and right, adding exponentially growing indexes,
until the two trends collide.
"""
# Limit growth for huge numbers of pages.
starting_factor = max(1, (end - begin) // 100)
factor = _iter_factors(starting_factor)
left_half, right_half = [], []
left_val, right_val = begin, end
right_val = end
while left_val < right_val:
left_half.append(left_val)
right_half.append(right_val)
next_factor = next(factor)
left_val = begin + next_factor
right_val = end - next_factor
# If the trends happen to meet exactly at one point, retain it.
if left_val == right_val:
left_half.append(left_val)
right_half.reverse()
return left_half + right_half | python | def _make_elastic_range(begin, end):
"""Generate an S-curved range of pages.
Start from both left and right, adding exponentially growing indexes,
until the two trends collide.
"""
# Limit growth for huge numbers of pages.
starting_factor = max(1, (end - begin) // 100)
factor = _iter_factors(starting_factor)
left_half, right_half = [], []
left_val, right_val = begin, end
right_val = end
while left_val < right_val:
left_half.append(left_val)
right_half.append(right_val)
next_factor = next(factor)
left_val = begin + next_factor
right_val = end - next_factor
# If the trends happen to meet exactly at one point, retain it.
if left_val == right_val:
left_half.append(left_val)
right_half.reverse()
return left_half + right_half | [
"def",
"_make_elastic_range",
"(",
"begin",
",",
"end",
")",
":",
"# Limit growth for huge numbers of pages.",
"starting_factor",
"=",
"max",
"(",
"1",
",",
"(",
"end",
"-",
"begin",
")",
"//",
"100",
")",
"factor",
"=",
"_iter_factors",
"(",
"starting_factor",
")",
"left_half",
",",
"right_half",
"=",
"[",
"]",
",",
"[",
"]",
"left_val",
",",
"right_val",
"=",
"begin",
",",
"end",
"right_val",
"=",
"end",
"while",
"left_val",
"<",
"right_val",
":",
"left_half",
".",
"append",
"(",
"left_val",
")",
"right_half",
".",
"append",
"(",
"right_val",
")",
"next_factor",
"=",
"next",
"(",
"factor",
")",
"left_val",
"=",
"begin",
"+",
"next_factor",
"right_val",
"=",
"end",
"-",
"next_factor",
"# If the trends happen to meet exactly at one point, retain it.",
"if",
"left_val",
"==",
"right_val",
":",
"left_half",
".",
"append",
"(",
"left_val",
")",
"right_half",
".",
"reverse",
"(",
")",
"return",
"left_half",
"+",
"right_half"
] | Generate an S-curved range of pages.
Start from both left and right, adding exponentially growing indexes,
until the two trends collide. | [
"Generate",
"an",
"S",
"-",
"curved",
"range",
"of",
"pages",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/utils.py#L118-L140 | train |
shtalinberg/django-el-pagination | el_pagination/utils.py | get_elastic_page_numbers | def get_elastic_page_numbers(current_page, num_pages):
"""Alternative callable for page listing.
Produce an adaptive pagination, useful for big numbers of pages, by
splitting the num_pages ranges in two parts at current_page. Each part
will have its own S-curve.
"""
if num_pages <= 10:
return list(range(1, num_pages + 1))
if current_page == 1:
pages = [1]
else:
pages = ['first', 'previous']
pages.extend(_make_elastic_range(1, current_page))
if current_page != num_pages:
pages.extend(_make_elastic_range(current_page, num_pages)[1:])
pages.extend(['next', 'last'])
return pages | python | def get_elastic_page_numbers(current_page, num_pages):
"""Alternative callable for page listing.
Produce an adaptive pagination, useful for big numbers of pages, by
splitting the num_pages ranges in two parts at current_page. Each part
will have its own S-curve.
"""
if num_pages <= 10:
return list(range(1, num_pages + 1))
if current_page == 1:
pages = [1]
else:
pages = ['first', 'previous']
pages.extend(_make_elastic_range(1, current_page))
if current_page != num_pages:
pages.extend(_make_elastic_range(current_page, num_pages)[1:])
pages.extend(['next', 'last'])
return pages | [
"def",
"get_elastic_page_numbers",
"(",
"current_page",
",",
"num_pages",
")",
":",
"if",
"num_pages",
"<=",
"10",
":",
"return",
"list",
"(",
"range",
"(",
"1",
",",
"num_pages",
"+",
"1",
")",
")",
"if",
"current_page",
"==",
"1",
":",
"pages",
"=",
"[",
"1",
"]",
"else",
":",
"pages",
"=",
"[",
"'first'",
",",
"'previous'",
"]",
"pages",
".",
"extend",
"(",
"_make_elastic_range",
"(",
"1",
",",
"current_page",
")",
")",
"if",
"current_page",
"!=",
"num_pages",
":",
"pages",
".",
"extend",
"(",
"_make_elastic_range",
"(",
"current_page",
",",
"num_pages",
")",
"[",
"1",
":",
"]",
")",
"pages",
".",
"extend",
"(",
"[",
"'next'",
",",
"'last'",
"]",
")",
"return",
"pages"
] | Alternative callable for page listing.
Produce an adaptive pagination, useful for big numbers of pages, by
splitting the num_pages ranges in two parts at current_page. Each part
will have its own S-curve. | [
"Alternative",
"callable",
"for",
"page",
"listing",
"."
] | 889ba62b46cb58292d554753a0bfda0b0a6d57da | https://github.com/shtalinberg/django-el-pagination/blob/889ba62b46cb58292d554753a0bfda0b0a6d57da/el_pagination/utils.py#L143-L160 | train |
justinmayer/django-autoslug | autoslug/utils.py | get_prepopulated_value | def get_prepopulated_value(field, instance):
"""
Returns preliminary value based on `populate_from`.
"""
if hasattr(field.populate_from, '__call__'):
# AutoSlugField(populate_from=lambda instance: ...)
return field.populate_from(instance)
else:
# AutoSlugField(populate_from='foo')
attr = getattr(instance, field.populate_from)
return callable(attr) and attr() or attr | python | def get_prepopulated_value(field, instance):
"""
Returns preliminary value based on `populate_from`.
"""
if hasattr(field.populate_from, '__call__'):
# AutoSlugField(populate_from=lambda instance: ...)
return field.populate_from(instance)
else:
# AutoSlugField(populate_from='foo')
attr = getattr(instance, field.populate_from)
return callable(attr) and attr() or attr | [
"def",
"get_prepopulated_value",
"(",
"field",
",",
"instance",
")",
":",
"if",
"hasattr",
"(",
"field",
".",
"populate_from",
",",
"'__call__'",
")",
":",
"# AutoSlugField(populate_from=lambda instance: ...)",
"return",
"field",
".",
"populate_from",
"(",
"instance",
")",
"else",
":",
"# AutoSlugField(populate_from='foo')",
"attr",
"=",
"getattr",
"(",
"instance",
",",
"field",
".",
"populate_from",
")",
"return",
"callable",
"(",
"attr",
")",
"and",
"attr",
"(",
")",
"or",
"attr"
] | Returns preliminary value based on `populate_from`. | [
"Returns",
"preliminary",
"value",
"based",
"on",
"populate_from",
"."
] | b3991daddf5a476a829b48c28afad4ae08a18179 | https://github.com/justinmayer/django-autoslug/blob/b3991daddf5a476a829b48c28afad4ae08a18179/autoslug/utils.py#L35-L45 | train |
justinmayer/django-autoslug | autoslug/utils.py | get_uniqueness_lookups | def get_uniqueness_lookups(field, instance, unique_with):
"""
Returns a dict'able tuple of lookups to ensure uniqueness of a slug.
"""
for original_lookup_name in unique_with:
if '__' in original_lookup_name:
field_name, inner_lookup = original_lookup_name.split('__', 1)
else:
field_name, inner_lookup = original_lookup_name, None
try:
other_field = instance._meta.get_field(field_name)
except FieldDoesNotExist:
raise ValueError('Could not find attribute %s.%s referenced'
' by %s.%s (see constraint `unique_with`)'
% (instance._meta.object_name, field_name,
instance._meta.object_name, field.name))
if field == other_field:
raise ValueError('Attribute %s.%s references itself in `unique_with`.'
' Please use "unique=True" for this case.'
% (instance._meta.object_name, field_name))
value = getattr(instance, field_name)
if not value:
if other_field.blank:
field_object = instance._meta.get_field(field_name)
if isinstance(field_object, ForeignKey):
lookup = '%s__isnull' % field_name
yield lookup, True
break
raise ValueError('Could not check uniqueness of %s.%s with'
' respect to %s.%s because the latter is empty.'
' Please ensure that "%s" is declared *after*'
' all fields listed in unique_with.'
% (instance._meta.object_name, field.name,
instance._meta.object_name, field_name,
field.name))
if isinstance(other_field, DateField): # DateTimeField is a DateField subclass
inner_lookup = inner_lookup or 'day'
if '__' in inner_lookup:
raise ValueError('The `unique_with` constraint in %s.%s'
' is set to "%s", but AutoSlugField only'
' accepts one level of nesting for dates'
' (e.g. "date__month").'
% (instance._meta.object_name, field.name,
original_lookup_name))
parts = ['year', 'month', 'day']
try:
granularity = parts.index(inner_lookup) + 1
except ValueError:
raise ValueError('expected one of %s, got "%s" in "%s"'
% (parts, inner_lookup, original_lookup_name))
else:
for part in parts[:granularity]:
lookup = '%s__%s' % (field_name, part)
yield lookup, getattr(value, part)
else:
# TODO: this part should be documented as it involves recursion
if inner_lookup:
if not hasattr(value, '_meta'):
raise ValueError('Could not resolve lookup "%s" in `unique_with` of %s.%s'
% (original_lookup_name, instance._meta.object_name, field.name))
for inner_name, inner_value in get_uniqueness_lookups(field, value, [inner_lookup]):
yield original_lookup_name, inner_value
else:
yield field_name, value | python | def get_uniqueness_lookups(field, instance, unique_with):
"""
Returns a dict'able tuple of lookups to ensure uniqueness of a slug.
"""
for original_lookup_name in unique_with:
if '__' in original_lookup_name:
field_name, inner_lookup = original_lookup_name.split('__', 1)
else:
field_name, inner_lookup = original_lookup_name, None
try:
other_field = instance._meta.get_field(field_name)
except FieldDoesNotExist:
raise ValueError('Could not find attribute %s.%s referenced'
' by %s.%s (see constraint `unique_with`)'
% (instance._meta.object_name, field_name,
instance._meta.object_name, field.name))
if field == other_field:
raise ValueError('Attribute %s.%s references itself in `unique_with`.'
' Please use "unique=True" for this case.'
% (instance._meta.object_name, field_name))
value = getattr(instance, field_name)
if not value:
if other_field.blank:
field_object = instance._meta.get_field(field_name)
if isinstance(field_object, ForeignKey):
lookup = '%s__isnull' % field_name
yield lookup, True
break
raise ValueError('Could not check uniqueness of %s.%s with'
' respect to %s.%s because the latter is empty.'
' Please ensure that "%s" is declared *after*'
' all fields listed in unique_with.'
% (instance._meta.object_name, field.name,
instance._meta.object_name, field_name,
field.name))
if isinstance(other_field, DateField): # DateTimeField is a DateField subclass
inner_lookup = inner_lookup or 'day'
if '__' in inner_lookup:
raise ValueError('The `unique_with` constraint in %s.%s'
' is set to "%s", but AutoSlugField only'
' accepts one level of nesting for dates'
' (e.g. "date__month").'
% (instance._meta.object_name, field.name,
original_lookup_name))
parts = ['year', 'month', 'day']
try:
granularity = parts.index(inner_lookup) + 1
except ValueError:
raise ValueError('expected one of %s, got "%s" in "%s"'
% (parts, inner_lookup, original_lookup_name))
else:
for part in parts[:granularity]:
lookup = '%s__%s' % (field_name, part)
yield lookup, getattr(value, part)
else:
# TODO: this part should be documented as it involves recursion
if inner_lookup:
if not hasattr(value, '_meta'):
raise ValueError('Could not resolve lookup "%s" in `unique_with` of %s.%s'
% (original_lookup_name, instance._meta.object_name, field.name))
for inner_name, inner_value in get_uniqueness_lookups(field, value, [inner_lookup]):
yield original_lookup_name, inner_value
else:
yield field_name, value | [
"def",
"get_uniqueness_lookups",
"(",
"field",
",",
"instance",
",",
"unique_with",
")",
":",
"for",
"original_lookup_name",
"in",
"unique_with",
":",
"if",
"'__'",
"in",
"original_lookup_name",
":",
"field_name",
",",
"inner_lookup",
"=",
"original_lookup_name",
".",
"split",
"(",
"'__'",
",",
"1",
")",
"else",
":",
"field_name",
",",
"inner_lookup",
"=",
"original_lookup_name",
",",
"None",
"try",
":",
"other_field",
"=",
"instance",
".",
"_meta",
".",
"get_field",
"(",
"field_name",
")",
"except",
"FieldDoesNotExist",
":",
"raise",
"ValueError",
"(",
"'Could not find attribute %s.%s referenced'",
"' by %s.%s (see constraint `unique_with`)'",
"%",
"(",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field_name",
",",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field",
".",
"name",
")",
")",
"if",
"field",
"==",
"other_field",
":",
"raise",
"ValueError",
"(",
"'Attribute %s.%s references itself in `unique_with`.'",
"' Please use \"unique=True\" for this case.'",
"%",
"(",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field_name",
")",
")",
"value",
"=",
"getattr",
"(",
"instance",
",",
"field_name",
")",
"if",
"not",
"value",
":",
"if",
"other_field",
".",
"blank",
":",
"field_object",
"=",
"instance",
".",
"_meta",
".",
"get_field",
"(",
"field_name",
")",
"if",
"isinstance",
"(",
"field_object",
",",
"ForeignKey",
")",
":",
"lookup",
"=",
"'%s__isnull'",
"%",
"field_name",
"yield",
"lookup",
",",
"True",
"break",
"raise",
"ValueError",
"(",
"'Could not check uniqueness of %s.%s with'",
"' respect to %s.%s because the latter is empty.'",
"' Please ensure that \"%s\" is declared *after*'",
"' all fields listed in unique_with.'",
"%",
"(",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field",
".",
"name",
",",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field_name",
",",
"field",
".",
"name",
")",
")",
"if",
"isinstance",
"(",
"other_field",
",",
"DateField",
")",
":",
"# DateTimeField is a DateField subclass",
"inner_lookup",
"=",
"inner_lookup",
"or",
"'day'",
"if",
"'__'",
"in",
"inner_lookup",
":",
"raise",
"ValueError",
"(",
"'The `unique_with` constraint in %s.%s'",
"' is set to \"%s\", but AutoSlugField only'",
"' accepts one level of nesting for dates'",
"' (e.g. \"date__month\").'",
"%",
"(",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field",
".",
"name",
",",
"original_lookup_name",
")",
")",
"parts",
"=",
"[",
"'year'",
",",
"'month'",
",",
"'day'",
"]",
"try",
":",
"granularity",
"=",
"parts",
".",
"index",
"(",
"inner_lookup",
")",
"+",
"1",
"except",
"ValueError",
":",
"raise",
"ValueError",
"(",
"'expected one of %s, got \"%s\" in \"%s\"'",
"%",
"(",
"parts",
",",
"inner_lookup",
",",
"original_lookup_name",
")",
")",
"else",
":",
"for",
"part",
"in",
"parts",
"[",
":",
"granularity",
"]",
":",
"lookup",
"=",
"'%s__%s'",
"%",
"(",
"field_name",
",",
"part",
")",
"yield",
"lookup",
",",
"getattr",
"(",
"value",
",",
"part",
")",
"else",
":",
"# TODO: this part should be documented as it involves recursion",
"if",
"inner_lookup",
":",
"if",
"not",
"hasattr",
"(",
"value",
",",
"'_meta'",
")",
":",
"raise",
"ValueError",
"(",
"'Could not resolve lookup \"%s\" in `unique_with` of %s.%s'",
"%",
"(",
"original_lookup_name",
",",
"instance",
".",
"_meta",
".",
"object_name",
",",
"field",
".",
"name",
")",
")",
"for",
"inner_name",
",",
"inner_value",
"in",
"get_uniqueness_lookups",
"(",
"field",
",",
"value",
",",
"[",
"inner_lookup",
"]",
")",
":",
"yield",
"original_lookup_name",
",",
"inner_value",
"else",
":",
"yield",
"field_name",
",",
"value"
] | Returns a dict'able tuple of lookups to ensure uniqueness of a slug. | [
"Returns",
"a",
"dict",
"able",
"tuple",
"of",
"lookups",
"to",
"ensure",
"uniqueness",
"of",
"a",
"slug",
"."
] | b3991daddf5a476a829b48c28afad4ae08a18179 | https://github.com/justinmayer/django-autoslug/blob/b3991daddf5a476a829b48c28afad4ae08a18179/autoslug/utils.py#L94-L162 | train |
erikrose/blessings | blessings/__init__.py | derivative_colors | def derivative_colors(colors):
"""Return the names of valid color variants, given the base colors."""
return set([('on_' + c) for c in colors] +
[('bright_' + c) for c in colors] +
[('on_bright_' + c) for c in colors]) | python | def derivative_colors(colors):
"""Return the names of valid color variants, given the base colors."""
return set([('on_' + c) for c in colors] +
[('bright_' + c) for c in colors] +
[('on_bright_' + c) for c in colors]) | [
"def",
"derivative_colors",
"(",
"colors",
")",
":",
"return",
"set",
"(",
"[",
"(",
"'on_'",
"+",
"c",
")",
"for",
"c",
"in",
"colors",
"]",
"+",
"[",
"(",
"'bright_'",
"+",
"c",
")",
"for",
"c",
"in",
"colors",
"]",
"+",
"[",
"(",
"'on_bright_'",
"+",
"c",
")",
"for",
"c",
"in",
"colors",
"]",
")"
] | Return the names of valid color variants, given the base colors. | [
"Return",
"the",
"names",
"of",
"valid",
"color",
"variants",
"given",
"the",
"base",
"colors",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L414-L418 | train |
erikrose/blessings | blessings/__init__.py | split_into_formatters | def split_into_formatters(compound):
"""Split a possibly compound format string into segments.
>>> split_into_formatters('bold_underline_bright_blue_on_red')
['bold', 'underline', 'bright_blue', 'on_red']
>>> split_into_formatters('red_no_italic_shadow_on_bright_cyan')
['red', 'no_italic', 'shadow', 'on_bright_cyan']
"""
merged_segs = []
# These occur only as prefixes, so they can always be merged:
mergeable_prefixes = ['no', 'on', 'bright', 'on_bright']
for s in compound.split('_'):
if merged_segs and merged_segs[-1] in mergeable_prefixes:
merged_segs[-1] += '_' + s
else:
merged_segs.append(s)
return merged_segs | python | def split_into_formatters(compound):
"""Split a possibly compound format string into segments.
>>> split_into_formatters('bold_underline_bright_blue_on_red')
['bold', 'underline', 'bright_blue', 'on_red']
>>> split_into_formatters('red_no_italic_shadow_on_bright_cyan')
['red', 'no_italic', 'shadow', 'on_bright_cyan']
"""
merged_segs = []
# These occur only as prefixes, so they can always be merged:
mergeable_prefixes = ['no', 'on', 'bright', 'on_bright']
for s in compound.split('_'):
if merged_segs and merged_segs[-1] in mergeable_prefixes:
merged_segs[-1] += '_' + s
else:
merged_segs.append(s)
return merged_segs | [
"def",
"split_into_formatters",
"(",
"compound",
")",
":",
"merged_segs",
"=",
"[",
"]",
"# These occur only as prefixes, so they can always be merged:",
"mergeable_prefixes",
"=",
"[",
"'no'",
",",
"'on'",
",",
"'bright'",
",",
"'on_bright'",
"]",
"for",
"s",
"in",
"compound",
".",
"split",
"(",
"'_'",
")",
":",
"if",
"merged_segs",
"and",
"merged_segs",
"[",
"-",
"1",
"]",
"in",
"mergeable_prefixes",
":",
"merged_segs",
"[",
"-",
"1",
"]",
"+=",
"'_'",
"+",
"s",
"else",
":",
"merged_segs",
".",
"append",
"(",
"s",
")",
"return",
"merged_segs"
] | Split a possibly compound format string into segments.
>>> split_into_formatters('bold_underline_bright_blue_on_red')
['bold', 'underline', 'bright_blue', 'on_red']
>>> split_into_formatters('red_no_italic_shadow_on_bright_cyan')
['red', 'no_italic', 'shadow', 'on_bright_cyan'] | [
"Split",
"a",
"possibly",
"compound",
"format",
"string",
"into",
"segments",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L548-L564 | train |
erikrose/blessings | blessings/__init__.py | Terminal.location | def location(self, x=None, y=None):
"""Return a context manager for temporarily moving the cursor.
Move the cursor to a certain position on entry, let you print stuff
there, then return the cursor to its original position::
term = Terminal()
with term.location(2, 5):
print('Hello, world!')
for x in xrange(10):
print('I can do it %i times!' % x)
Specify ``x`` to move to a certain column, ``y`` to move to a certain
row, both, or neither. If you specify neither, only the saving and
restoration of cursor position will happen. This can be useful if you
simply want to restore your place after doing some manual cursor
movement.
"""
# Save position and move to the requested column, row, or both:
self.stream.write(self.save)
if x is not None and y is not None:
self.stream.write(self.move(y, x))
elif x is not None:
self.stream.write(self.move_x(x))
elif y is not None:
self.stream.write(self.move_y(y))
try:
yield
finally:
# Restore original cursor position:
self.stream.write(self.restore) | python | def location(self, x=None, y=None):
"""Return a context manager for temporarily moving the cursor.
Move the cursor to a certain position on entry, let you print stuff
there, then return the cursor to its original position::
term = Terminal()
with term.location(2, 5):
print('Hello, world!')
for x in xrange(10):
print('I can do it %i times!' % x)
Specify ``x`` to move to a certain column, ``y`` to move to a certain
row, both, or neither. If you specify neither, only the saving and
restoration of cursor position will happen. This can be useful if you
simply want to restore your place after doing some manual cursor
movement.
"""
# Save position and move to the requested column, row, or both:
self.stream.write(self.save)
if x is not None and y is not None:
self.stream.write(self.move(y, x))
elif x is not None:
self.stream.write(self.move_x(x))
elif y is not None:
self.stream.write(self.move_y(y))
try:
yield
finally:
# Restore original cursor position:
self.stream.write(self.restore) | [
"def",
"location",
"(",
"self",
",",
"x",
"=",
"None",
",",
"y",
"=",
"None",
")",
":",
"# Save position and move to the requested column, row, or both:",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"save",
")",
"if",
"x",
"is",
"not",
"None",
"and",
"y",
"is",
"not",
"None",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"move",
"(",
"y",
",",
"x",
")",
")",
"elif",
"x",
"is",
"not",
"None",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"move_x",
"(",
"x",
")",
")",
"elif",
"y",
"is",
"not",
"None",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"move_y",
"(",
"y",
")",
")",
"try",
":",
"yield",
"finally",
":",
"# Restore original cursor position:",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"restore",
")"
] | Return a context manager for temporarily moving the cursor.
Move the cursor to a certain position on entry, let you print stuff
there, then return the cursor to its original position::
term = Terminal()
with term.location(2, 5):
print('Hello, world!')
for x in xrange(10):
print('I can do it %i times!' % x)
Specify ``x`` to move to a certain column, ``y`` to move to a certain
row, both, or neither. If you specify neither, only the saving and
restoration of cursor position will happen. This can be useful if you
simply want to restore your place after doing some manual cursor
movement. | [
"Return",
"a",
"context",
"manager",
"for",
"temporarily",
"moving",
"the",
"cursor",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L242-L273 | train |
erikrose/blessings | blessings/__init__.py | Terminal.fullscreen | def fullscreen(self):
"""Return a context manager that enters fullscreen mode while inside it
and restores normal mode on leaving."""
self.stream.write(self.enter_fullscreen)
try:
yield
finally:
self.stream.write(self.exit_fullscreen) | python | def fullscreen(self):
"""Return a context manager that enters fullscreen mode while inside it
and restores normal mode on leaving."""
self.stream.write(self.enter_fullscreen)
try:
yield
finally:
self.stream.write(self.exit_fullscreen) | [
"def",
"fullscreen",
"(",
"self",
")",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"enter_fullscreen",
")",
"try",
":",
"yield",
"finally",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"exit_fullscreen",
")"
] | Return a context manager that enters fullscreen mode while inside it
and restores normal mode on leaving. | [
"Return",
"a",
"context",
"manager",
"that",
"enters",
"fullscreen",
"mode",
"while",
"inside",
"it",
"and",
"restores",
"normal",
"mode",
"on",
"leaving",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L276-L283 | train |
erikrose/blessings | blessings/__init__.py | Terminal.hidden_cursor | def hidden_cursor(self):
"""Return a context manager that hides the cursor while inside it and
makes it visible on leaving."""
self.stream.write(self.hide_cursor)
try:
yield
finally:
self.stream.write(self.normal_cursor) | python | def hidden_cursor(self):
"""Return a context manager that hides the cursor while inside it and
makes it visible on leaving."""
self.stream.write(self.hide_cursor)
try:
yield
finally:
self.stream.write(self.normal_cursor) | [
"def",
"hidden_cursor",
"(",
"self",
")",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"hide_cursor",
")",
"try",
":",
"yield",
"finally",
":",
"self",
".",
"stream",
".",
"write",
"(",
"self",
".",
"normal_cursor",
")"
] | Return a context manager that hides the cursor while inside it and
makes it visible on leaving. | [
"Return",
"a",
"context",
"manager",
"that",
"hides",
"the",
"cursor",
"while",
"inside",
"it",
"and",
"makes",
"it",
"visible",
"on",
"leaving",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L286-L293 | train |
erikrose/blessings | blessings/__init__.py | Terminal._resolve_formatter | def _resolve_formatter(self, attr):
"""Resolve a sugary or plain capability name, color, or compound
formatting function name into a callable capability.
Return a ``ParametrizingString`` or a ``FormattingString``.
"""
if attr in COLORS:
return self._resolve_color(attr)
elif attr in COMPOUNDABLES:
# Bold, underline, or something that takes no parameters
return self._formatting_string(self._resolve_capability(attr))
else:
formatters = split_into_formatters(attr)
if all(f in COMPOUNDABLES for f in formatters):
# It's a compound formatter, like "bold_green_on_red". Future
# optimization: combine all formatting into a single escape
# sequence.
return self._formatting_string(
u''.join(self._resolve_formatter(s) for s in formatters))
else:
return ParametrizingString(self._resolve_capability(attr)) | python | def _resolve_formatter(self, attr):
"""Resolve a sugary or plain capability name, color, or compound
formatting function name into a callable capability.
Return a ``ParametrizingString`` or a ``FormattingString``.
"""
if attr in COLORS:
return self._resolve_color(attr)
elif attr in COMPOUNDABLES:
# Bold, underline, or something that takes no parameters
return self._formatting_string(self._resolve_capability(attr))
else:
formatters = split_into_formatters(attr)
if all(f in COMPOUNDABLES for f in formatters):
# It's a compound formatter, like "bold_green_on_red". Future
# optimization: combine all formatting into a single escape
# sequence.
return self._formatting_string(
u''.join(self._resolve_formatter(s) for s in formatters))
else:
return ParametrizingString(self._resolve_capability(attr)) | [
"def",
"_resolve_formatter",
"(",
"self",
",",
"attr",
")",
":",
"if",
"attr",
"in",
"COLORS",
":",
"return",
"self",
".",
"_resolve_color",
"(",
"attr",
")",
"elif",
"attr",
"in",
"COMPOUNDABLES",
":",
"# Bold, underline, or something that takes no parameters",
"return",
"self",
".",
"_formatting_string",
"(",
"self",
".",
"_resolve_capability",
"(",
"attr",
")",
")",
"else",
":",
"formatters",
"=",
"split_into_formatters",
"(",
"attr",
")",
"if",
"all",
"(",
"f",
"in",
"COMPOUNDABLES",
"for",
"f",
"in",
"formatters",
")",
":",
"# It's a compound formatter, like \"bold_green_on_red\". Future",
"# optimization: combine all formatting into a single escape",
"# sequence.",
"return",
"self",
".",
"_formatting_string",
"(",
"u''",
".",
"join",
"(",
"self",
".",
"_resolve_formatter",
"(",
"s",
")",
"for",
"s",
"in",
"formatters",
")",
")",
"else",
":",
"return",
"ParametrizingString",
"(",
"self",
".",
"_resolve_capability",
"(",
"attr",
")",
")"
] | Resolve a sugary or plain capability name, color, or compound
formatting function name into a callable capability.
Return a ``ParametrizingString`` or a ``FormattingString``. | [
"Resolve",
"a",
"sugary",
"or",
"plain",
"capability",
"name",
"color",
"or",
"compound",
"formatting",
"function",
"name",
"into",
"a",
"callable",
"capability",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L347-L368 | train |
erikrose/blessings | blessings/__init__.py | Terminal._resolve_capability | def _resolve_capability(self, atom):
"""Return a terminal code for a capname or a sugary name, or an empty
Unicode.
The return value is always Unicode, because otherwise it is clumsy
(especially in Python 3) to concatenate with real (Unicode) strings.
"""
code = tigetstr(self._sugar.get(atom, atom))
if code:
# See the comment in ParametrizingString for why this is latin1.
return code.decode('latin1')
return u'' | python | def _resolve_capability(self, atom):
"""Return a terminal code for a capname or a sugary name, or an empty
Unicode.
The return value is always Unicode, because otherwise it is clumsy
(especially in Python 3) to concatenate with real (Unicode) strings.
"""
code = tigetstr(self._sugar.get(atom, atom))
if code:
# See the comment in ParametrizingString for why this is latin1.
return code.decode('latin1')
return u'' | [
"def",
"_resolve_capability",
"(",
"self",
",",
"atom",
")",
":",
"code",
"=",
"tigetstr",
"(",
"self",
".",
"_sugar",
".",
"get",
"(",
"atom",
",",
"atom",
")",
")",
"if",
"code",
":",
"# See the comment in ParametrizingString for why this is latin1.",
"return",
"code",
".",
"decode",
"(",
"'latin1'",
")",
"return",
"u''"
] | Return a terminal code for a capname or a sugary name, or an empty
Unicode.
The return value is always Unicode, because otherwise it is clumsy
(especially in Python 3) to concatenate with real (Unicode) strings. | [
"Return",
"a",
"terminal",
"code",
"for",
"a",
"capname",
"or",
"a",
"sugary",
"name",
"or",
"an",
"empty",
"Unicode",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L370-L382 | train |
erikrose/blessings | blessings/__init__.py | Terminal._resolve_color | def _resolve_color(self, color):
"""Resolve a color like red or on_bright_green into a callable
capability."""
# TODO: Does curses automatically exchange red and blue and cyan and
# yellow when a terminal supports setf/setb rather than setaf/setab?
# I'll be blasted if I can find any documentation. The following
# assumes it does.
color_cap = (self._background_color if 'on_' in color else
self._foreground_color)
# curses constants go up to only 7, so add an offset to get at the
# bright colors at 8-15:
offset = 8 if 'bright_' in color else 0
base_color = color.rsplit('_', 1)[-1]
return self._formatting_string(
color_cap(getattr(curses, 'COLOR_' + base_color.upper()) + offset)) | python | def _resolve_color(self, color):
"""Resolve a color like red or on_bright_green into a callable
capability."""
# TODO: Does curses automatically exchange red and blue and cyan and
# yellow when a terminal supports setf/setb rather than setaf/setab?
# I'll be blasted if I can find any documentation. The following
# assumes it does.
color_cap = (self._background_color if 'on_' in color else
self._foreground_color)
# curses constants go up to only 7, so add an offset to get at the
# bright colors at 8-15:
offset = 8 if 'bright_' in color else 0
base_color = color.rsplit('_', 1)[-1]
return self._formatting_string(
color_cap(getattr(curses, 'COLOR_' + base_color.upper()) + offset)) | [
"def",
"_resolve_color",
"(",
"self",
",",
"color",
")",
":",
"# TODO: Does curses automatically exchange red and blue and cyan and",
"# yellow when a terminal supports setf/setb rather than setaf/setab?",
"# I'll be blasted if I can find any documentation. The following",
"# assumes it does.",
"color_cap",
"=",
"(",
"self",
".",
"_background_color",
"if",
"'on_'",
"in",
"color",
"else",
"self",
".",
"_foreground_color",
")",
"# curses constants go up to only 7, so add an offset to get at the",
"# bright colors at 8-15:",
"offset",
"=",
"8",
"if",
"'bright_'",
"in",
"color",
"else",
"0",
"base_color",
"=",
"color",
".",
"rsplit",
"(",
"'_'",
",",
"1",
")",
"[",
"-",
"1",
"]",
"return",
"self",
".",
"_formatting_string",
"(",
"color_cap",
"(",
"getattr",
"(",
"curses",
",",
"'COLOR_'",
"+",
"base_color",
".",
"upper",
"(",
")",
")",
"+",
"offset",
")",
")"
] | Resolve a color like red or on_bright_green into a callable
capability. | [
"Resolve",
"a",
"color",
"like",
"red",
"or",
"on_bright_green",
"into",
"a",
"callable",
"capability",
"."
] | b1d4daf948d1db8455af64836906785204d09055 | https://github.com/erikrose/blessings/blob/b1d4daf948d1db8455af64836906785204d09055/blessings/__init__.py#L384-L398 | train |
relekang/django-nopassword | nopassword/backends/sms.py | TwilioBackend.send_login_code | def send_login_code(self, code, context, **kwargs):
"""
Send a login code via SMS
"""
from_number = self.from_number or getattr(settings, 'DEFAULT_FROM_NUMBER')
sms_content = render_to_string(self.template_name, context)
self.twilio_client.messages.create(
to=code.user.phone_number,
from_=from_number,
body=sms_content
) | python | def send_login_code(self, code, context, **kwargs):
"""
Send a login code via SMS
"""
from_number = self.from_number or getattr(settings, 'DEFAULT_FROM_NUMBER')
sms_content = render_to_string(self.template_name, context)
self.twilio_client.messages.create(
to=code.user.phone_number,
from_=from_number,
body=sms_content
) | [
"def",
"send_login_code",
"(",
"self",
",",
"code",
",",
"context",
",",
"*",
"*",
"kwargs",
")",
":",
"from_number",
"=",
"self",
".",
"from_number",
"or",
"getattr",
"(",
"settings",
",",
"'DEFAULT_FROM_NUMBER'",
")",
"sms_content",
"=",
"render_to_string",
"(",
"self",
".",
"template_name",
",",
"context",
")",
"self",
".",
"twilio_client",
".",
"messages",
".",
"create",
"(",
"to",
"=",
"code",
".",
"user",
".",
"phone_number",
",",
"from_",
"=",
"from_number",
",",
"body",
"=",
"sms_content",
")"
] | Send a login code via SMS | [
"Send",
"a",
"login",
"code",
"via",
"SMS"
] | d1d0f99617b1394c860864852326be673f9b935f | https://github.com/relekang/django-nopassword/blob/d1d0f99617b1394c860864852326be673f9b935f/nopassword/backends/sms.py#L20-L31 | train |
bshillingford/python-torchfile | torchfile.py | load | def load(filename, **kwargs):
"""
Loads the given t7 file using default settings; kwargs are forwarded
to `T7Reader`.
"""
with open(filename, 'rb') as f:
reader = T7Reader(f, **kwargs)
return reader.read_obj() | python | def load(filename, **kwargs):
"""
Loads the given t7 file using default settings; kwargs are forwarded
to `T7Reader`.
"""
with open(filename, 'rb') as f:
reader = T7Reader(f, **kwargs)
return reader.read_obj() | [
"def",
"load",
"(",
"filename",
",",
"*",
"*",
"kwargs",
")",
":",
"with",
"open",
"(",
"filename",
",",
"'rb'",
")",
"as",
"f",
":",
"reader",
"=",
"T7Reader",
"(",
"f",
",",
"*",
"*",
"kwargs",
")",
"return",
"reader",
".",
"read_obj",
"(",
")"
] | Loads the given t7 file using default settings; kwargs are forwarded
to `T7Reader`. | [
"Loads",
"the",
"given",
"t7",
"file",
"using",
"default",
"settings",
";",
"kwargs",
"are",
"forwarded",
"to",
"T7Reader",
"."
] | 20b3e13b6267c254e9df67446844010629f48d61 | https://github.com/bshillingford/python-torchfile/blob/20b3e13b6267c254e9df67446844010629f48d61/torchfile.py#L417-L424 | train |
slackapi/python-rtmbot | rtmbot/core.py | Job.check | def check(self):
''' Returns True if `interval` seconds have passed since it last ran '''
if self.lastrun + self.interval < time.time():
return True
else:
return False | python | def check(self):
''' Returns True if `interval` seconds have passed since it last ran '''
if self.lastrun + self.interval < time.time():
return True
else:
return False | [
"def",
"check",
"(",
"self",
")",
":",
"if",
"self",
".",
"lastrun",
"+",
"self",
".",
"interval",
"<",
"time",
".",
"time",
"(",
")",
":",
"return",
"True",
"else",
":",
"return",
"False"
] | Returns True if `interval` seconds have passed since it last ran | [
"Returns",
"True",
"if",
"interval",
"seconds",
"have",
"passed",
"since",
"it",
"last",
"ran"
] | 1adc5df6791a9f6e7070ab1ff0d1c95db7dbb0ab | https://github.com/slackapi/python-rtmbot/blob/1adc5df6791a9f6e7070ab1ff0d1c95db7dbb0ab/rtmbot/core.py#L293-L298 | train |
anttttti/Wordbatch | wordbatch/feature_union.py | make_union | def make_union(*transformers, **kwargs):
"""Construct a FeatureUnion from the given transformers.
This is a shorthand for the FeatureUnion constructor; it does not require,
and does not permit, naming the transformers. Instead, they will be given
names automatically based on their types. It also does not allow weighting.
Parameters
----------
*transformers : list of estimators
n_jobs : int, optional
Number of jobs to run in parallel (default 1).
Returns
-------
f : FeatureUnion
Examples
--------
>>> from sklearn.decomposition import PCA, TruncatedSVD
>>> from sklearn.pipeline import make_union
>>> make_union(PCA(), TruncatedSVD()) # doctest: +NORMALIZE_WHITESPACE
FeatureUnion(n_jobs=1,
transformer_list=[('pca',
PCA(copy=True, iterated_power='auto',
n_components=None, random_state=None,
svd_solver='auto', tol=0.0, whiten=False)),
('truncatedsvd',
TruncatedSVD(algorithm='randomized',
n_components=2, n_iter=5,
random_state=None, tol=0.0))],
transformer_weights=None)
"""
n_jobs = kwargs.pop('n_jobs', 1)
concatenate = kwargs.pop('concatenate', True)
if kwargs:
# We do not currently support `transformer_weights` as we may want to
# change its type spec in make_union
raise TypeError('Unknown keyword arguments: "{}"'
.format(list(kwargs.keys())[0]))
return FeatureUnion(_name_estimators(transformers), n_jobs= n_jobs, concatenate= concatenate) | python | def make_union(*transformers, **kwargs):
"""Construct a FeatureUnion from the given transformers.
This is a shorthand for the FeatureUnion constructor; it does not require,
and does not permit, naming the transformers. Instead, they will be given
names automatically based on their types. It also does not allow weighting.
Parameters
----------
*transformers : list of estimators
n_jobs : int, optional
Number of jobs to run in parallel (default 1).
Returns
-------
f : FeatureUnion
Examples
--------
>>> from sklearn.decomposition import PCA, TruncatedSVD
>>> from sklearn.pipeline import make_union
>>> make_union(PCA(), TruncatedSVD()) # doctest: +NORMALIZE_WHITESPACE
FeatureUnion(n_jobs=1,
transformer_list=[('pca',
PCA(copy=True, iterated_power='auto',
n_components=None, random_state=None,
svd_solver='auto', tol=0.0, whiten=False)),
('truncatedsvd',
TruncatedSVD(algorithm='randomized',
n_components=2, n_iter=5,
random_state=None, tol=0.0))],
transformer_weights=None)
"""
n_jobs = kwargs.pop('n_jobs', 1)
concatenate = kwargs.pop('concatenate', True)
if kwargs:
# We do not currently support `transformer_weights` as we may want to
# change its type spec in make_union
raise TypeError('Unknown keyword arguments: "{}"'
.format(list(kwargs.keys())[0]))
return FeatureUnion(_name_estimators(transformers), n_jobs= n_jobs, concatenate= concatenate) | [
"def",
"make_union",
"(",
"*",
"transformers",
",",
"*",
"*",
"kwargs",
")",
":",
"n_jobs",
"=",
"kwargs",
".",
"pop",
"(",
"'n_jobs'",
",",
"1",
")",
"concatenate",
"=",
"kwargs",
".",
"pop",
"(",
"'concatenate'",
",",
"True",
")",
"if",
"kwargs",
":",
"# We do not currently support `transformer_weights` as we may want to",
"# change its type spec in make_union",
"raise",
"TypeError",
"(",
"'Unknown keyword arguments: \"{}\"'",
".",
"format",
"(",
"list",
"(",
"kwargs",
".",
"keys",
"(",
")",
")",
"[",
"0",
"]",
")",
")",
"return",
"FeatureUnion",
"(",
"_name_estimators",
"(",
"transformers",
")",
",",
"n_jobs",
"=",
"n_jobs",
",",
"concatenate",
"=",
"concatenate",
")"
] | Construct a FeatureUnion from the given transformers.
This is a shorthand for the FeatureUnion constructor; it does not require,
and does not permit, naming the transformers. Instead, they will be given
names automatically based on their types. It also does not allow weighting.
Parameters
----------
*transformers : list of estimators
n_jobs : int, optional
Number of jobs to run in parallel (default 1).
Returns
-------
f : FeatureUnion
Examples
--------
>>> from sklearn.decomposition import PCA, TruncatedSVD
>>> from sklearn.pipeline import make_union
>>> make_union(PCA(), TruncatedSVD()) # doctest: +NORMALIZE_WHITESPACE
FeatureUnion(n_jobs=1,
transformer_list=[('pca',
PCA(copy=True, iterated_power='auto',
n_components=None, random_state=None,
svd_solver='auto', tol=0.0, whiten=False)),
('truncatedsvd',
TruncatedSVD(algorithm='randomized',
n_components=2, n_iter=5,
random_state=None, tol=0.0))],
transformer_weights=None) | [
"Construct",
"a",
"FeatureUnion",
"from",
"the",
"given",
"transformers",
"."
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/feature_union.py#L208-L249 | train |
anttttti/Wordbatch | wordbatch/feature_union.py | FeatureUnion.get_feature_names | def get_feature_names(self):
"""Get feature names from all transformers.
Returns
-------
feature_names : list of strings
Names of the features produced by transform.
"""
feature_names = []
for name, trans, weight in self._iter():
if not hasattr(trans, 'get_feature_names'):
raise AttributeError("Transformer %s (type %s) does not "
"provide get_feature_names."
% (str(name), type(trans).__name__))
feature_names.extend([name + "__" + f for f in
trans.get_feature_names()])
return feature_names | python | def get_feature_names(self):
"""Get feature names from all transformers.
Returns
-------
feature_names : list of strings
Names of the features produced by transform.
"""
feature_names = []
for name, trans, weight in self._iter():
if not hasattr(trans, 'get_feature_names'):
raise AttributeError("Transformer %s (type %s) does not "
"provide get_feature_names."
% (str(name), type(trans).__name__))
feature_names.extend([name + "__" + f for f in
trans.get_feature_names()])
return feature_names | [
"def",
"get_feature_names",
"(",
"self",
")",
":",
"feature_names",
"=",
"[",
"]",
"for",
"name",
",",
"trans",
",",
"weight",
"in",
"self",
".",
"_iter",
"(",
")",
":",
"if",
"not",
"hasattr",
"(",
"trans",
",",
"'get_feature_names'",
")",
":",
"raise",
"AttributeError",
"(",
"\"Transformer %s (type %s) does not \"",
"\"provide get_feature_names.\"",
"%",
"(",
"str",
"(",
"name",
")",
",",
"type",
"(",
"trans",
")",
".",
"__name__",
")",
")",
"feature_names",
".",
"extend",
"(",
"[",
"name",
"+",
"\"__\"",
"+",
"f",
"for",
"f",
"in",
"trans",
".",
"get_feature_names",
"(",
")",
"]",
")",
"return",
"feature_names"
] | Get feature names from all transformers.
Returns
-------
feature_names : list of strings
Names of the features produced by transform. | [
"Get",
"feature",
"names",
"from",
"all",
"transformers",
"."
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/feature_union.py#L97-L113 | train |
anttttti/Wordbatch | wordbatch/feature_union.py | FeatureUnion.fit | def fit(self, X, y=None):
"""Fit all transformers using X.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data, used to fit transformers.
y : array-like, shape (n_samples, ...), optional
Targets for supervised learning.
Returns
-------
self : FeatureUnion
This estimator
"""
self.transformer_list = list(self.transformer_list)
self._validate_transformers()
with Pool(self.n_jobs) as pool:
transformers = pool.starmap(_fit_one_transformer,
((trans, X[trans['col_pick']] if hasattr(trans, 'col_pick') else X, y) for _, trans, _ in self._iter()))
self._update_transformer_list(transformers)
return self | python | def fit(self, X, y=None):
"""Fit all transformers using X.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data, used to fit transformers.
y : array-like, shape (n_samples, ...), optional
Targets for supervised learning.
Returns
-------
self : FeatureUnion
This estimator
"""
self.transformer_list = list(self.transformer_list)
self._validate_transformers()
with Pool(self.n_jobs) as pool:
transformers = pool.starmap(_fit_one_transformer,
((trans, X[trans['col_pick']] if hasattr(trans, 'col_pick') else X, y) for _, trans, _ in self._iter()))
self._update_transformer_list(transformers)
return self | [
"def",
"fit",
"(",
"self",
",",
"X",
",",
"y",
"=",
"None",
")",
":",
"self",
".",
"transformer_list",
"=",
"list",
"(",
"self",
".",
"transformer_list",
")",
"self",
".",
"_validate_transformers",
"(",
")",
"with",
"Pool",
"(",
"self",
".",
"n_jobs",
")",
"as",
"pool",
":",
"transformers",
"=",
"pool",
".",
"starmap",
"(",
"_fit_one_transformer",
",",
"(",
"(",
"trans",
",",
"X",
"[",
"trans",
"[",
"'col_pick'",
"]",
"]",
"if",
"hasattr",
"(",
"trans",
",",
"'col_pick'",
")",
"else",
"X",
",",
"y",
")",
"for",
"_",
",",
"trans",
",",
"_",
"in",
"self",
".",
"_iter",
"(",
")",
")",
")",
"self",
".",
"_update_transformer_list",
"(",
"transformers",
")",
"return",
"self"
] | Fit all transformers using X.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data, used to fit transformers.
y : array-like, shape (n_samples, ...), optional
Targets for supervised learning.
Returns
-------
self : FeatureUnion
This estimator | [
"Fit",
"all",
"transformers",
"using",
"X",
"."
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/feature_union.py#L115-L137 | train |
anttttti/Wordbatch | wordbatch/feature_union.py | FeatureUnion.fit_transform | def fit_transform(self, X, y=None, **fit_params):
"""Fit all transformers, transform the data and concatenate results.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data to be transformed.
y : array-like, shape (n_samples, ...), optional
Targets for supervised learning.
Returns
-------
X_t : array-like or sparse matrix, shape (n_samples, sum_n_components)
hstack of results of transformers. sum_n_components is the
sum of n_components (output dimension) over transformers.
"""
self._validate_transformers()
with Pool(self.n_jobs) as pool:
result = pool.starmap(_fit_transform_one,
((trans, weight, X[trans['col_pick']] if hasattr(trans, 'col_pick') else X, y)
for name, trans, weight in self._iter()))
if not result:
# All transformers are None
return np.zeros((X.shape[0], 0))
Xs, transformers = zip(*result)
self._update_transformer_list(transformers)
if self.concatenate:
if any(sparse.issparse(f) for f in Xs):
Xs = sparse.hstack(Xs).tocsr()
else:
Xs = np.hstack(Xs)
return Xs | python | def fit_transform(self, X, y=None, **fit_params):
"""Fit all transformers, transform the data and concatenate results.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data to be transformed.
y : array-like, shape (n_samples, ...), optional
Targets for supervised learning.
Returns
-------
X_t : array-like or sparse matrix, shape (n_samples, sum_n_components)
hstack of results of transformers. sum_n_components is the
sum of n_components (output dimension) over transformers.
"""
self._validate_transformers()
with Pool(self.n_jobs) as pool:
result = pool.starmap(_fit_transform_one,
((trans, weight, X[trans['col_pick']] if hasattr(trans, 'col_pick') else X, y)
for name, trans, weight in self._iter()))
if not result:
# All transformers are None
return np.zeros((X.shape[0], 0))
Xs, transformers = zip(*result)
self._update_transformer_list(transformers)
if self.concatenate:
if any(sparse.issparse(f) for f in Xs):
Xs = sparse.hstack(Xs).tocsr()
else:
Xs = np.hstack(Xs)
return Xs | [
"def",
"fit_transform",
"(",
"self",
",",
"X",
",",
"y",
"=",
"None",
",",
"*",
"*",
"fit_params",
")",
":",
"self",
".",
"_validate_transformers",
"(",
")",
"with",
"Pool",
"(",
"self",
".",
"n_jobs",
")",
"as",
"pool",
":",
"result",
"=",
"pool",
".",
"starmap",
"(",
"_fit_transform_one",
",",
"(",
"(",
"trans",
",",
"weight",
",",
"X",
"[",
"trans",
"[",
"'col_pick'",
"]",
"]",
"if",
"hasattr",
"(",
"trans",
",",
"'col_pick'",
")",
"else",
"X",
",",
"y",
")",
"for",
"name",
",",
"trans",
",",
"weight",
"in",
"self",
".",
"_iter",
"(",
")",
")",
")",
"if",
"not",
"result",
":",
"# All transformers are None",
"return",
"np",
".",
"zeros",
"(",
"(",
"X",
".",
"shape",
"[",
"0",
"]",
",",
"0",
")",
")",
"Xs",
",",
"transformers",
"=",
"zip",
"(",
"*",
"result",
")",
"self",
".",
"_update_transformer_list",
"(",
"transformers",
")",
"if",
"self",
".",
"concatenate",
":",
"if",
"any",
"(",
"sparse",
".",
"issparse",
"(",
"f",
")",
"for",
"f",
"in",
"Xs",
")",
":",
"Xs",
"=",
"sparse",
".",
"hstack",
"(",
"Xs",
")",
".",
"tocsr",
"(",
")",
"else",
":",
"Xs",
"=",
"np",
".",
"hstack",
"(",
"Xs",
")",
"return",
"Xs"
] | Fit all transformers, transform the data and concatenate results.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data to be transformed.
y : array-like, shape (n_samples, ...), optional
Targets for supervised learning.
Returns
-------
X_t : array-like or sparse matrix, shape (n_samples, sum_n_components)
hstack of results of transformers. sum_n_components is the
sum of n_components (output dimension) over transformers. | [
"Fit",
"all",
"transformers",
"transform",
"the",
"data",
"and",
"concatenate",
"results",
"."
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/feature_union.py#L139-L171 | train |
anttttti/Wordbatch | wordbatch/feature_union.py | FeatureUnion.transform | def transform(self, X):
"""Transform X separately by each transformer, concatenate results.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data to be transformed.
Returns
-------
X_t : array-like or sparse matrix, shape (n_samples, sum_n_components)
hstack of results of transformers. sum_n_components is the
sum of n_components (output dimension) over transformers.
"""
with Pool(self.n_jobs) as pool:
Xs = pool.starmap(_transform_one, ((trans, weight, X[trans['col_pick']] if hasattr(trans, 'col_pick')
else X) for name, trans, weight in self._iter()))
if not Xs:
# All transformers are None
return np.zeros((X.shape[0], 0))
if self.concatenate:
if any(sparse.issparse(f) for f in Xs):
Xs = sparse.hstack(Xs).tocsr()
else:
Xs = np.hstack(Xs)
return Xs | python | def transform(self, X):
"""Transform X separately by each transformer, concatenate results.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data to be transformed.
Returns
-------
X_t : array-like or sparse matrix, shape (n_samples, sum_n_components)
hstack of results of transformers. sum_n_components is the
sum of n_components (output dimension) over transformers.
"""
with Pool(self.n_jobs) as pool:
Xs = pool.starmap(_transform_one, ((trans, weight, X[trans['col_pick']] if hasattr(trans, 'col_pick')
else X) for name, trans, weight in self._iter()))
if not Xs:
# All transformers are None
return np.zeros((X.shape[0], 0))
if self.concatenate:
if any(sparse.issparse(f) for f in Xs):
Xs = sparse.hstack(Xs).tocsr()
else:
Xs = np.hstack(Xs)
return Xs | [
"def",
"transform",
"(",
"self",
",",
"X",
")",
":",
"with",
"Pool",
"(",
"self",
".",
"n_jobs",
")",
"as",
"pool",
":",
"Xs",
"=",
"pool",
".",
"starmap",
"(",
"_transform_one",
",",
"(",
"(",
"trans",
",",
"weight",
",",
"X",
"[",
"trans",
"[",
"'col_pick'",
"]",
"]",
"if",
"hasattr",
"(",
"trans",
",",
"'col_pick'",
")",
"else",
"X",
")",
"for",
"name",
",",
"trans",
",",
"weight",
"in",
"self",
".",
"_iter",
"(",
")",
")",
")",
"if",
"not",
"Xs",
":",
"# All transformers are None",
"return",
"np",
".",
"zeros",
"(",
"(",
"X",
".",
"shape",
"[",
"0",
"]",
",",
"0",
")",
")",
"if",
"self",
".",
"concatenate",
":",
"if",
"any",
"(",
"sparse",
".",
"issparse",
"(",
"f",
")",
"for",
"f",
"in",
"Xs",
")",
":",
"Xs",
"=",
"sparse",
".",
"hstack",
"(",
"Xs",
")",
".",
"tocsr",
"(",
")",
"else",
":",
"Xs",
"=",
"np",
".",
"hstack",
"(",
"Xs",
")",
"return",
"Xs"
] | Transform X separately by each transformer, concatenate results.
Parameters
----------
X : iterable or array-like, depending on transformers
Input data to be transformed.
Returns
-------
X_t : array-like or sparse matrix, shape (n_samples, sum_n_components)
hstack of results of transformers. sum_n_components is the
sum of n_components (output dimension) over transformers. | [
"Transform",
"X",
"separately",
"by",
"each",
"transformer",
"concatenate",
"results",
"."
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/feature_union.py#L173-L198 | train |
anttttti/Wordbatch | wordbatch/batcher.py | Batcher.split_batches | def split_batches(self, data, minibatch_size= None):
"""Split data into minibatches with a specified size
Parameters
----------
data: iterable and indexable
List-like data to be split into batches. Includes spark_contextipy matrices and Pandas DataFrames.
minibatch_size: int
Expected sizes of minibatches split from the data.
Returns
-------
data_split: list
List of minibatches, each entry is a list-like object representing the data subset in a batch.
"""
if minibatch_size==None: minibatch_size= self.minibatch_size
if isinstance(data, list) or isinstance(data, tuple): len_data= len(data)
else: len_data= data.shape[0]
if isinstance(data,pd.DataFrame):
data_split = [data.iloc[x * minibatch_size:(x + 1) * minibatch_size] for x in
range(int(ceil(len_data / minibatch_size)))]
else:
data_split= [data[x* minibatch_size:min(len_data, (x+1)*minibatch_size)]
for x in range(int(ceil(len_data/minibatch_size)))]
return data_split | python | def split_batches(self, data, minibatch_size= None):
"""Split data into minibatches with a specified size
Parameters
----------
data: iterable and indexable
List-like data to be split into batches. Includes spark_contextipy matrices and Pandas DataFrames.
minibatch_size: int
Expected sizes of minibatches split from the data.
Returns
-------
data_split: list
List of minibatches, each entry is a list-like object representing the data subset in a batch.
"""
if minibatch_size==None: minibatch_size= self.minibatch_size
if isinstance(data, list) or isinstance(data, tuple): len_data= len(data)
else: len_data= data.shape[0]
if isinstance(data,pd.DataFrame):
data_split = [data.iloc[x * minibatch_size:(x + 1) * minibatch_size] for x in
range(int(ceil(len_data / minibatch_size)))]
else:
data_split= [data[x* minibatch_size:min(len_data, (x+1)*minibatch_size)]
for x in range(int(ceil(len_data/minibatch_size)))]
return data_split | [
"def",
"split_batches",
"(",
"self",
",",
"data",
",",
"minibatch_size",
"=",
"None",
")",
":",
"if",
"minibatch_size",
"==",
"None",
":",
"minibatch_size",
"=",
"self",
".",
"minibatch_size",
"if",
"isinstance",
"(",
"data",
",",
"list",
")",
"or",
"isinstance",
"(",
"data",
",",
"tuple",
")",
":",
"len_data",
"=",
"len",
"(",
"data",
")",
"else",
":",
"len_data",
"=",
"data",
".",
"shape",
"[",
"0",
"]",
"if",
"isinstance",
"(",
"data",
",",
"pd",
".",
"DataFrame",
")",
":",
"data_split",
"=",
"[",
"data",
".",
"iloc",
"[",
"x",
"*",
"minibatch_size",
":",
"(",
"x",
"+",
"1",
")",
"*",
"minibatch_size",
"]",
"for",
"x",
"in",
"range",
"(",
"int",
"(",
"ceil",
"(",
"len_data",
"/",
"minibatch_size",
")",
")",
")",
"]",
"else",
":",
"data_split",
"=",
"[",
"data",
"[",
"x",
"*",
"minibatch_size",
":",
"min",
"(",
"len_data",
",",
"(",
"x",
"+",
"1",
")",
"*",
"minibatch_size",
")",
"]",
"for",
"x",
"in",
"range",
"(",
"int",
"(",
"ceil",
"(",
"len_data",
"/",
"minibatch_size",
")",
")",
")",
"]",
"return",
"data_split"
] | Split data into minibatches with a specified size
Parameters
----------
data: iterable and indexable
List-like data to be split into batches. Includes spark_contextipy matrices and Pandas DataFrames.
minibatch_size: int
Expected sizes of minibatches split from the data.
Returns
-------
data_split: list
List of minibatches, each entry is a list-like object representing the data subset in a batch. | [
"Split",
"data",
"into",
"minibatches",
"with",
"a",
"specified",
"size"
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/batcher.py#L77-L102 | train |
anttttti/Wordbatch | wordbatch/batcher.py | Batcher.merge_batches | def merge_batches(self, data):
"""Merge a list of data minibatches into one single instance representing the data
Parameters
----------
data: list
List of minibatches to merge
Returns
-------
(anonymous): sparse matrix | pd.DataFrame | list
Single complete list-like data merged from given batches
"""
if isinstance(data[0], ssp.csr_matrix): return ssp.vstack(data)
if isinstance(data[0], pd.DataFrame) or isinstance(data[0], pd.Series): return pd.concat(data)
return [item for sublist in data for item in sublist] | python | def merge_batches(self, data):
"""Merge a list of data minibatches into one single instance representing the data
Parameters
----------
data: list
List of minibatches to merge
Returns
-------
(anonymous): sparse matrix | pd.DataFrame | list
Single complete list-like data merged from given batches
"""
if isinstance(data[0], ssp.csr_matrix): return ssp.vstack(data)
if isinstance(data[0], pd.DataFrame) or isinstance(data[0], pd.Series): return pd.concat(data)
return [item for sublist in data for item in sublist] | [
"def",
"merge_batches",
"(",
"self",
",",
"data",
")",
":",
"if",
"isinstance",
"(",
"data",
"[",
"0",
"]",
",",
"ssp",
".",
"csr_matrix",
")",
":",
"return",
"ssp",
".",
"vstack",
"(",
"data",
")",
"if",
"isinstance",
"(",
"data",
"[",
"0",
"]",
",",
"pd",
".",
"DataFrame",
")",
"or",
"isinstance",
"(",
"data",
"[",
"0",
"]",
",",
"pd",
".",
"Series",
")",
":",
"return",
"pd",
".",
"concat",
"(",
"data",
")",
"return",
"[",
"item",
"for",
"sublist",
"in",
"data",
"for",
"item",
"in",
"sublist",
"]"
] | Merge a list of data minibatches into one single instance representing the data
Parameters
----------
data: list
List of minibatches to merge
Returns
-------
(anonymous): sparse matrix | pd.DataFrame | list
Single complete list-like data merged from given batches | [
"Merge",
"a",
"list",
"of",
"data",
"minibatches",
"into",
"one",
"single",
"instance",
"representing",
"the",
"data"
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/batcher.py#L104-L119 | train |
anttttti/Wordbatch | wordbatch/batcher.py | Batcher.shuffle_batch | def shuffle_batch(self, texts, labels= None, seed= None):
"""Shuffle a list of samples, as well as the labels if specified
Parameters
----------
texts: list-like
List of samples to shuffle
labels: list-like (optional)
List of labels to shuffle, should be correspondent to the samples given
seed: int
The seed of the pseudo random number generator to use for shuffling
Returns
-------
texts: list
List of shuffled samples (texts parameters)
labels: list (optional)
List of shuffled labels. This will only be returned when non-None
labels is passed
"""
if seed!=None: random.seed(seed)
index_shuf= list(range(len(texts)))
random.shuffle(index_shuf)
texts= [texts[x] for x in index_shuf]
if labels==None: return texts
labels= [labels[x] for x in index_shuf]
return texts, labels | python | def shuffle_batch(self, texts, labels= None, seed= None):
"""Shuffle a list of samples, as well as the labels if specified
Parameters
----------
texts: list-like
List of samples to shuffle
labels: list-like (optional)
List of labels to shuffle, should be correspondent to the samples given
seed: int
The seed of the pseudo random number generator to use for shuffling
Returns
-------
texts: list
List of shuffled samples (texts parameters)
labels: list (optional)
List of shuffled labels. This will only be returned when non-None
labels is passed
"""
if seed!=None: random.seed(seed)
index_shuf= list(range(len(texts)))
random.shuffle(index_shuf)
texts= [texts[x] for x in index_shuf]
if labels==None: return texts
labels= [labels[x] for x in index_shuf]
return texts, labels | [
"def",
"shuffle_batch",
"(",
"self",
",",
"texts",
",",
"labels",
"=",
"None",
",",
"seed",
"=",
"None",
")",
":",
"if",
"seed",
"!=",
"None",
":",
"random",
".",
"seed",
"(",
"seed",
")",
"index_shuf",
"=",
"list",
"(",
"range",
"(",
"len",
"(",
"texts",
")",
")",
")",
"random",
".",
"shuffle",
"(",
"index_shuf",
")",
"texts",
"=",
"[",
"texts",
"[",
"x",
"]",
"for",
"x",
"in",
"index_shuf",
"]",
"if",
"labels",
"==",
"None",
":",
"return",
"texts",
"labels",
"=",
"[",
"labels",
"[",
"x",
"]",
"for",
"x",
"in",
"index_shuf",
"]",
"return",
"texts",
",",
"labels"
] | Shuffle a list of samples, as well as the labels if specified
Parameters
----------
texts: list-like
List of samples to shuffle
labels: list-like (optional)
List of labels to shuffle, should be correspondent to the samples given
seed: int
The seed of the pseudo random number generator to use for shuffling
Returns
-------
texts: list
List of shuffled samples (texts parameters)
labels: list (optional)
List of shuffled labels. This will only be returned when non-None
labels is passed | [
"Shuffle",
"a",
"list",
"of",
"samples",
"as",
"well",
"as",
"the",
"labels",
"if",
"specified"
] | ef57b5c1d87d9c82fb096598125c2511f9819e4d | https://github.com/anttttti/Wordbatch/blob/ef57b5c1d87d9c82fb096598125c2511f9819e4d/wordbatch/batcher.py#L230-L259 | train |
t-makaro/animatplot | animatplot/util.py | demeshgrid | def demeshgrid(arr):
"""Turns an ndarray created by a meshgrid back into a 1D array
Parameters
----------
arr : array of dimension > 1
This array should have been created by a meshgrid.
"""
dim = len(arr.shape)
for i in range(dim):
Slice1 = [0]*dim
Slice2 = [1]*dim
Slice1[i] = slice(None)
Slice2[i] = slice(None)
if (arr[tuple(Slice1)] == arr[tuple(Slice2)]).all():
return arr[tuple(Slice1)] | python | def demeshgrid(arr):
"""Turns an ndarray created by a meshgrid back into a 1D array
Parameters
----------
arr : array of dimension > 1
This array should have been created by a meshgrid.
"""
dim = len(arr.shape)
for i in range(dim):
Slice1 = [0]*dim
Slice2 = [1]*dim
Slice1[i] = slice(None)
Slice2[i] = slice(None)
if (arr[tuple(Slice1)] == arr[tuple(Slice2)]).all():
return arr[tuple(Slice1)] | [
"def",
"demeshgrid",
"(",
"arr",
")",
":",
"dim",
"=",
"len",
"(",
"arr",
".",
"shape",
")",
"for",
"i",
"in",
"range",
"(",
"dim",
")",
":",
"Slice1",
"=",
"[",
"0",
"]",
"*",
"dim",
"Slice2",
"=",
"[",
"1",
"]",
"*",
"dim",
"Slice1",
"[",
"i",
"]",
"=",
"slice",
"(",
"None",
")",
"Slice2",
"[",
"i",
"]",
"=",
"slice",
"(",
"None",
")",
"if",
"(",
"arr",
"[",
"tuple",
"(",
"Slice1",
")",
"]",
"==",
"arr",
"[",
"tuple",
"(",
"Slice2",
")",
"]",
")",
".",
"all",
"(",
")",
":",
"return",
"arr",
"[",
"tuple",
"(",
"Slice1",
")",
"]"
] | Turns an ndarray created by a meshgrid back into a 1D array
Parameters
----------
arr : array of dimension > 1
This array should have been created by a meshgrid. | [
"Turns",
"an",
"ndarray",
"created",
"by",
"a",
"meshgrid",
"back",
"into",
"a",
"1D",
"array"
] | 632d988687fca7e7415e9fa49fe7eebc3f0991c6 | https://github.com/t-makaro/animatplot/blob/632d988687fca7e7415e9fa49fe7eebc3f0991c6/animatplot/util.py#L23-L38 | train |
t-makaro/animatplot | animatplot/animation.py | Animation.timeline_slider | def timeline_slider(self, text='Time', ax=None, valfmt=None, color=None):
"""Creates a timeline slider.
Parameters
----------
text : str, optional
The text to display for the slider. Defaults to 'Time'
ax : matplotlib.axes.Axes, optional
The matplotlib axes to attach the slider to.
valfmt : str, optional
a format specifier used to print the time
Defaults to '%s' for datetime64, timedelta64 and '%1.2f' otherwise.
color :
The color of the slider.
"""
if ax is None:
adjust_plot = {'bottom': .2}
rect = [.18, .05, .5, .03]
plt.subplots_adjust(**adjust_plot)
self.slider_ax = plt.axes(rect)
else:
self.slider_ax = ax
if valfmt is None:
if (np.issubdtype(self.timeline.t.dtype, np.datetime64)
or np.issubdtype(self.timeline.t.dtype, np.timedelta64)):
valfmt = '%s'
else:
valfmt = '%1.2f'
if self.timeline.log:
valfmt = '$10^{%s}$' % valfmt
self.slider = Slider(
self.slider_ax, text, 0, self.timeline._len-1,
valinit=0,
valfmt=(valfmt+self.timeline.units),
valstep=1, color=color
)
self._has_slider = True
def set_time(t):
self.timeline.index = int(self.slider.val)
self.slider.valtext.set_text(
self.slider.valfmt % (self.timeline[self.timeline.index]))
if self._pause:
for block in self.blocks:
block._update(self.timeline.index)
self.fig.canvas.draw()
self.slider.on_changed(set_time) | python | def timeline_slider(self, text='Time', ax=None, valfmt=None, color=None):
"""Creates a timeline slider.
Parameters
----------
text : str, optional
The text to display for the slider. Defaults to 'Time'
ax : matplotlib.axes.Axes, optional
The matplotlib axes to attach the slider to.
valfmt : str, optional
a format specifier used to print the time
Defaults to '%s' for datetime64, timedelta64 and '%1.2f' otherwise.
color :
The color of the slider.
"""
if ax is None:
adjust_plot = {'bottom': .2}
rect = [.18, .05, .5, .03]
plt.subplots_adjust(**adjust_plot)
self.slider_ax = plt.axes(rect)
else:
self.slider_ax = ax
if valfmt is None:
if (np.issubdtype(self.timeline.t.dtype, np.datetime64)
or np.issubdtype(self.timeline.t.dtype, np.timedelta64)):
valfmt = '%s'
else:
valfmt = '%1.2f'
if self.timeline.log:
valfmt = '$10^{%s}$' % valfmt
self.slider = Slider(
self.slider_ax, text, 0, self.timeline._len-1,
valinit=0,
valfmt=(valfmt+self.timeline.units),
valstep=1, color=color
)
self._has_slider = True
def set_time(t):
self.timeline.index = int(self.slider.val)
self.slider.valtext.set_text(
self.slider.valfmt % (self.timeline[self.timeline.index]))
if self._pause:
for block in self.blocks:
block._update(self.timeline.index)
self.fig.canvas.draw()
self.slider.on_changed(set_time) | [
"def",
"timeline_slider",
"(",
"self",
",",
"text",
"=",
"'Time'",
",",
"ax",
"=",
"None",
",",
"valfmt",
"=",
"None",
",",
"color",
"=",
"None",
")",
":",
"if",
"ax",
"is",
"None",
":",
"adjust_plot",
"=",
"{",
"'bottom'",
":",
".2",
"}",
"rect",
"=",
"[",
".18",
",",
".05",
",",
".5",
",",
".03",
"]",
"plt",
".",
"subplots_adjust",
"(",
"*",
"*",
"adjust_plot",
")",
"self",
".",
"slider_ax",
"=",
"plt",
".",
"axes",
"(",
"rect",
")",
"else",
":",
"self",
".",
"slider_ax",
"=",
"ax",
"if",
"valfmt",
"is",
"None",
":",
"if",
"(",
"np",
".",
"issubdtype",
"(",
"self",
".",
"timeline",
".",
"t",
".",
"dtype",
",",
"np",
".",
"datetime64",
")",
"or",
"np",
".",
"issubdtype",
"(",
"self",
".",
"timeline",
".",
"t",
".",
"dtype",
",",
"np",
".",
"timedelta64",
")",
")",
":",
"valfmt",
"=",
"'%s'",
"else",
":",
"valfmt",
"=",
"'%1.2f'",
"if",
"self",
".",
"timeline",
".",
"log",
":",
"valfmt",
"=",
"'$10^{%s}$'",
"%",
"valfmt",
"self",
".",
"slider",
"=",
"Slider",
"(",
"self",
".",
"slider_ax",
",",
"text",
",",
"0",
",",
"self",
".",
"timeline",
".",
"_len",
"-",
"1",
",",
"valinit",
"=",
"0",
",",
"valfmt",
"=",
"(",
"valfmt",
"+",
"self",
".",
"timeline",
".",
"units",
")",
",",
"valstep",
"=",
"1",
",",
"color",
"=",
"color",
")",
"self",
".",
"_has_slider",
"=",
"True",
"def",
"set_time",
"(",
"t",
")",
":",
"self",
".",
"timeline",
".",
"index",
"=",
"int",
"(",
"self",
".",
"slider",
".",
"val",
")",
"self",
".",
"slider",
".",
"valtext",
".",
"set_text",
"(",
"self",
".",
"slider",
".",
"valfmt",
"%",
"(",
"self",
".",
"timeline",
"[",
"self",
".",
"timeline",
".",
"index",
"]",
")",
")",
"if",
"self",
".",
"_pause",
":",
"for",
"block",
"in",
"self",
".",
"blocks",
":",
"block",
".",
"_update",
"(",
"self",
".",
"timeline",
".",
"index",
")",
"self",
".",
"fig",
".",
"canvas",
".",
"draw",
"(",
")",
"self",
".",
"slider",
".",
"on_changed",
"(",
"set_time",
")"
] | Creates a timeline slider.
Parameters
----------
text : str, optional
The text to display for the slider. Defaults to 'Time'
ax : matplotlib.axes.Axes, optional
The matplotlib axes to attach the slider to.
valfmt : str, optional
a format specifier used to print the time
Defaults to '%s' for datetime64, timedelta64 and '%1.2f' otherwise.
color :
The color of the slider. | [
"Creates",
"a",
"timeline",
"slider",
"."
] | 632d988687fca7e7415e9fa49fe7eebc3f0991c6 | https://github.com/t-makaro/animatplot/blob/632d988687fca7e7415e9fa49fe7eebc3f0991c6/animatplot/animation.py#L100-L150 | train |
t-makaro/animatplot | animatplot/animation.py | Animation.controls | def controls(self, timeline_slider_args={}, toggle_args={}):
"""Creates interactive controls for the animation
Creates both a play/pause button, and a time slider at once
Parameters
----------
timeline_slider_args : Dict, optional
A dictionary of arguments to be passed to timeline_slider()
toggle_args : Dict, optional
A dictionary of argyments to be passed to toggle()
"""
self.timeline_slider(**timeline_slider_args)
self.toggle(**toggle_args) | python | def controls(self, timeline_slider_args={}, toggle_args={}):
"""Creates interactive controls for the animation
Creates both a play/pause button, and a time slider at once
Parameters
----------
timeline_slider_args : Dict, optional
A dictionary of arguments to be passed to timeline_slider()
toggle_args : Dict, optional
A dictionary of argyments to be passed to toggle()
"""
self.timeline_slider(**timeline_slider_args)
self.toggle(**toggle_args) | [
"def",
"controls",
"(",
"self",
",",
"timeline_slider_args",
"=",
"{",
"}",
",",
"toggle_args",
"=",
"{",
"}",
")",
":",
"self",
".",
"timeline_slider",
"(",
"*",
"*",
"timeline_slider_args",
")",
"self",
".",
"toggle",
"(",
"*",
"*",
"toggle_args",
")"
] | Creates interactive controls for the animation
Creates both a play/pause button, and a time slider at once
Parameters
----------
timeline_slider_args : Dict, optional
A dictionary of arguments to be passed to timeline_slider()
toggle_args : Dict, optional
A dictionary of argyments to be passed to toggle() | [
"Creates",
"interactive",
"controls",
"for",
"the",
"animation"
] | 632d988687fca7e7415e9fa49fe7eebc3f0991c6 | https://github.com/t-makaro/animatplot/blob/632d988687fca7e7415e9fa49fe7eebc3f0991c6/animatplot/animation.py#L152-L165 | train |
t-makaro/animatplot | animatplot/animation.py | Animation.save_gif | def save_gif(self, filename):
"""Saves the animation to a gif
A convience function. Provided to let the user avoid dealing
with writers.
Parameters
----------
filename : str
the name of the file to be created without the file extension
"""
self.timeline.index -= 1 # required for proper starting point for save
self.animation.save(filename+'.gif', writer=PillowWriter(fps=self.timeline.fps)) | python | def save_gif(self, filename):
"""Saves the animation to a gif
A convience function. Provided to let the user avoid dealing
with writers.
Parameters
----------
filename : str
the name of the file to be created without the file extension
"""
self.timeline.index -= 1 # required for proper starting point for save
self.animation.save(filename+'.gif', writer=PillowWriter(fps=self.timeline.fps)) | [
"def",
"save_gif",
"(",
"self",
",",
"filename",
")",
":",
"self",
".",
"timeline",
".",
"index",
"-=",
"1",
"# required for proper starting point for save",
"self",
".",
"animation",
".",
"save",
"(",
"filename",
"+",
"'.gif'",
",",
"writer",
"=",
"PillowWriter",
"(",
"fps",
"=",
"self",
".",
"timeline",
".",
"fps",
")",
")"
] | Saves the animation to a gif
A convience function. Provided to let the user avoid dealing
with writers.
Parameters
----------
filename : str
the name of the file to be created without the file extension | [
"Saves",
"the",
"animation",
"to",
"a",
"gif"
] | 632d988687fca7e7415e9fa49fe7eebc3f0991c6 | https://github.com/t-makaro/animatplot/blob/632d988687fca7e7415e9fa49fe7eebc3f0991c6/animatplot/animation.py#L167-L179 | train |
t-makaro/animatplot | animatplot/animation.py | Animation.save | def save(self, *args, **kwargs):
"""Saves an animation
A wrapper around :meth:`matplotlib.animation.Animation.save`
"""
self.timeline.index -= 1 # required for proper starting point for save
self.animation.save(*args, **kwargs) | python | def save(self, *args, **kwargs):
"""Saves an animation
A wrapper around :meth:`matplotlib.animation.Animation.save`
"""
self.timeline.index -= 1 # required for proper starting point for save
self.animation.save(*args, **kwargs) | [
"def",
"save",
"(",
"self",
",",
"*",
"args",
",",
"*",
"*",
"kwargs",
")",
":",
"self",
".",
"timeline",
".",
"index",
"-=",
"1",
"# required for proper starting point for save",
"self",
".",
"animation",
".",
"save",
"(",
"*",
"args",
",",
"*",
"*",
"kwargs",
")"
] | Saves an animation
A wrapper around :meth:`matplotlib.animation.Animation.save` | [
"Saves",
"an",
"animation"
] | 632d988687fca7e7415e9fa49fe7eebc3f0991c6 | https://github.com/t-makaro/animatplot/blob/632d988687fca7e7415e9fa49fe7eebc3f0991c6/animatplot/animation.py#L181-L187 | train |
dadadel/pyment | pyment/docstring.py | isin_alone | def isin_alone(elems, line):
"""Check if an element from a list is the only element of a string.
:type elems: list
:type line: str
"""
found = False
for e in elems:
if line.strip().lower() == e.lower():
found = True
break
return found | python | def isin_alone(elems, line):
"""Check if an element from a list is the only element of a string.
:type elems: list
:type line: str
"""
found = False
for e in elems:
if line.strip().lower() == e.lower():
found = True
break
return found | [
"def",
"isin_alone",
"(",
"elems",
",",
"line",
")",
":",
"found",
"=",
"False",
"for",
"e",
"in",
"elems",
":",
"if",
"line",
".",
"strip",
"(",
")",
".",
"lower",
"(",
")",
"==",
"e",
".",
"lower",
"(",
")",
":",
"found",
"=",
"True",
"break",
"return",
"found"
] | Check if an element from a list is the only element of a string.
:type elems: list
:type line: str | [
"Check",
"if",
"an",
"element",
"from",
"a",
"list",
"is",
"the",
"only",
"element",
"of",
"a",
"string",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L26-L38 | train |
dadadel/pyment | pyment/docstring.py | isin_start | def isin_start(elems, line):
"""Check if an element from a list starts a string.
:type elems: list
:type line: str
"""
found = False
elems = [elems] if type(elems) is not list else elems
for e in elems:
if line.lstrip().lower().startswith(e):
found = True
break
return found | python | def isin_start(elems, line):
"""Check if an element from a list starts a string.
:type elems: list
:type line: str
"""
found = False
elems = [elems] if type(elems) is not list else elems
for e in elems:
if line.lstrip().lower().startswith(e):
found = True
break
return found | [
"def",
"isin_start",
"(",
"elems",
",",
"line",
")",
":",
"found",
"=",
"False",
"elems",
"=",
"[",
"elems",
"]",
"if",
"type",
"(",
"elems",
")",
"is",
"not",
"list",
"else",
"elems",
"for",
"e",
"in",
"elems",
":",
"if",
"line",
".",
"lstrip",
"(",
")",
".",
"lower",
"(",
")",
".",
"startswith",
"(",
"e",
")",
":",
"found",
"=",
"True",
"break",
"return",
"found"
] | Check if an element from a list starts a string.
:type elems: list
:type line: str | [
"Check",
"if",
"an",
"element",
"from",
"a",
"list",
"starts",
"a",
"string",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L41-L54 | train |
dadadel/pyment | pyment/docstring.py | isin | def isin(elems, line):
"""Check if an element from a list is in a string.
:type elems: list
:type line: str
"""
found = False
for e in elems:
if e in line.lower():
found = True
break
return found | python | def isin(elems, line):
"""Check if an element from a list is in a string.
:type elems: list
:type line: str
"""
found = False
for e in elems:
if e in line.lower():
found = True
break
return found | [
"def",
"isin",
"(",
"elems",
",",
"line",
")",
":",
"found",
"=",
"False",
"for",
"e",
"in",
"elems",
":",
"if",
"e",
"in",
"line",
".",
"lower",
"(",
")",
":",
"found",
"=",
"True",
"break",
"return",
"found"
] | Check if an element from a list is in a string.
:type elems: list
:type line: str | [
"Check",
"if",
"an",
"element",
"from",
"a",
"list",
"is",
"in",
"a",
"string",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L57-L69 | train |
dadadel/pyment | pyment/docstring.py | get_leading_spaces | def get_leading_spaces(data):
"""Get the leading space of a string if it is not empty
:type data: str
"""
spaces = ''
m = re.match(r'^(\s*)', data)
if m:
spaces = m.group(1)
return spaces | python | def get_leading_spaces(data):
"""Get the leading space of a string if it is not empty
:type data: str
"""
spaces = ''
m = re.match(r'^(\s*)', data)
if m:
spaces = m.group(1)
return spaces | [
"def",
"get_leading_spaces",
"(",
"data",
")",
":",
"spaces",
"=",
"''",
"m",
"=",
"re",
".",
"match",
"(",
"r'^(\\s*)'",
",",
"data",
")",
"if",
"m",
":",
"spaces",
"=",
"m",
".",
"group",
"(",
"1",
")",
"return",
"spaces"
] | Get the leading space of a string if it is not empty
:type data: str | [
"Get",
"the",
"leading",
"space",
"of",
"a",
"string",
"if",
"it",
"is",
"not",
"empty"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L72-L82 | train |
dadadel/pyment | pyment/docstring.py | DocToolsBase.get_mandatory_sections | def get_mandatory_sections(self):
"""Get mandatory sections"""
return [s for s in self.opt
if s not in self.optional_sections and
s not in self.excluded_sections] | python | def get_mandatory_sections(self):
"""Get mandatory sections"""
return [s for s in self.opt
if s not in self.optional_sections and
s not in self.excluded_sections] | [
"def",
"get_mandatory_sections",
"(",
"self",
")",
":",
"return",
"[",
"s",
"for",
"s",
"in",
"self",
".",
"opt",
"if",
"s",
"not",
"in",
"self",
".",
"optional_sections",
"and",
"s",
"not",
"in",
"self",
".",
"excluded_sections",
"]"
] | Get mandatory sections | [
"Get",
"mandatory",
"sections"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L133-L137 | train |
dadadel/pyment | pyment/docstring.py | NumpydocTools.get_raw_not_managed | def get_raw_not_managed(self, data):
"""Get elements not managed. They can be used as is.
:param data: the data to proceed
"""
keys = ['also', 'ref', 'note', 'other', 'example', 'method', 'attr']
elems = [self.opt[k] for k in self.opt if k in keys]
data = data.splitlines()
start = 0
init = 0
raw = ''
spaces = None
while start != -1:
start, end = self.get_next_section_lines(data[init:])
if start != -1:
init += start
if isin_alone(elems, data[init]) and \
not isin_alone([self.opt[e] for e in self.excluded_sections], data[init]):
spaces = get_leading_spaces(data[init])
if end != -1:
section = [d.replace(spaces, '', 1).rstrip() for d in data[init:init + end]]
else:
section = [d.replace(spaces, '', 1).rstrip() for d in data[init:]]
raw += '\n'.join(section) + '\n'
init += 2
return raw | python | def get_raw_not_managed(self, data):
"""Get elements not managed. They can be used as is.
:param data: the data to proceed
"""
keys = ['also', 'ref', 'note', 'other', 'example', 'method', 'attr']
elems = [self.opt[k] for k in self.opt if k in keys]
data = data.splitlines()
start = 0
init = 0
raw = ''
spaces = None
while start != -1:
start, end = self.get_next_section_lines(data[init:])
if start != -1:
init += start
if isin_alone(elems, data[init]) and \
not isin_alone([self.opt[e] for e in self.excluded_sections], data[init]):
spaces = get_leading_spaces(data[init])
if end != -1:
section = [d.replace(spaces, '', 1).rstrip() for d in data[init:init + end]]
else:
section = [d.replace(spaces, '', 1).rstrip() for d in data[init:]]
raw += '\n'.join(section) + '\n'
init += 2
return raw | [
"def",
"get_raw_not_managed",
"(",
"self",
",",
"data",
")",
":",
"keys",
"=",
"[",
"'also'",
",",
"'ref'",
",",
"'note'",
",",
"'other'",
",",
"'example'",
",",
"'method'",
",",
"'attr'",
"]",
"elems",
"=",
"[",
"self",
".",
"opt",
"[",
"k",
"]",
"for",
"k",
"in",
"self",
".",
"opt",
"if",
"k",
"in",
"keys",
"]",
"data",
"=",
"data",
".",
"splitlines",
"(",
")",
"start",
"=",
"0",
"init",
"=",
"0",
"raw",
"=",
"''",
"spaces",
"=",
"None",
"while",
"start",
"!=",
"-",
"1",
":",
"start",
",",
"end",
"=",
"self",
".",
"get_next_section_lines",
"(",
"data",
"[",
"init",
":",
"]",
")",
"if",
"start",
"!=",
"-",
"1",
":",
"init",
"+=",
"start",
"if",
"isin_alone",
"(",
"elems",
",",
"data",
"[",
"init",
"]",
")",
"and",
"not",
"isin_alone",
"(",
"[",
"self",
".",
"opt",
"[",
"e",
"]",
"for",
"e",
"in",
"self",
".",
"excluded_sections",
"]",
",",
"data",
"[",
"init",
"]",
")",
":",
"spaces",
"=",
"get_leading_spaces",
"(",
"data",
"[",
"init",
"]",
")",
"if",
"end",
"!=",
"-",
"1",
":",
"section",
"=",
"[",
"d",
".",
"replace",
"(",
"spaces",
",",
"''",
",",
"1",
")",
".",
"rstrip",
"(",
")",
"for",
"d",
"in",
"data",
"[",
"init",
":",
"init",
"+",
"end",
"]",
"]",
"else",
":",
"section",
"=",
"[",
"d",
".",
"replace",
"(",
"spaces",
",",
"''",
",",
"1",
")",
".",
"rstrip",
"(",
")",
"for",
"d",
"in",
"data",
"[",
"init",
":",
"]",
"]",
"raw",
"+=",
"'\\n'",
".",
"join",
"(",
"section",
")",
"+",
"'\\n'",
"init",
"+=",
"2",
"return",
"raw"
] | Get elements not managed. They can be used as is.
:param data: the data to proceed | [
"Get",
"elements",
"not",
"managed",
".",
"They",
"can",
"be",
"used",
"as",
"is",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L411-L437 | train |
dadadel/pyment | pyment/docstring.py | NumpydocTools.get_key_section_header | def get_key_section_header(self, key, spaces):
"""Get the key of the header section
:param key: the key name
:param spaces: spaces to set at the beginning of the header
"""
header = super(NumpydocTools, self).get_key_section_header(key, spaces)
header = spaces + header + '\n' + spaces + '-' * len(header) + '\n'
return header | python | def get_key_section_header(self, key, spaces):
"""Get the key of the header section
:param key: the key name
:param spaces: spaces to set at the beginning of the header
"""
header = super(NumpydocTools, self).get_key_section_header(key, spaces)
header = spaces + header + '\n' + spaces + '-' * len(header) + '\n'
return header | [
"def",
"get_key_section_header",
"(",
"self",
",",
"key",
",",
"spaces",
")",
":",
"header",
"=",
"super",
"(",
"NumpydocTools",
",",
"self",
")",
".",
"get_key_section_header",
"(",
"key",
",",
"spaces",
")",
"header",
"=",
"spaces",
"+",
"header",
"+",
"'\\n'",
"+",
"spaces",
"+",
"'-'",
"*",
"len",
"(",
"header",
")",
"+",
"'\\n'",
"return",
"header"
] | Get the key of the header section
:param key: the key name
:param spaces: spaces to set at the beginning of the header | [
"Get",
"the",
"key",
"of",
"the",
"header",
"section"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L439-L448 | train |
dadadel/pyment | pyment/docstring.py | DocsTools.autodetect_style | def autodetect_style(self, data):
"""Determine the style of a docstring,
and sets it as the default input one for the instance.
:param data: the docstring's data to recognize.
:type data: str
:returns: the style detected else 'unknown'
:rtype: str
"""
# evaluate styles with keys
found_keys = defaultdict(int)
for style in self.tagstyles:
for key in self.opt:
found_keys[style] += data.count(self.opt[key][style]['name'])
fkey = max(found_keys, key=found_keys.get)
detected_style = fkey if found_keys[fkey] else 'unknown'
# evaluate styles with groups
if detected_style == 'unknown':
found_groups = 0
found_googledoc = 0
found_numpydoc = 0
found_numpydocsep = 0
for line in data.strip().splitlines():
for key in self.groups:
found_groups += 1 if isin_start(self.groups[key], line) else 0
for key in self.googledoc:
found_googledoc += 1 if isin_start(self.googledoc[key], line) else 0
for key in self.numpydoc:
found_numpydoc += 1 if isin_start(self.numpydoc[key], line) else 0
if line.strip() and isin_alone(['-' * len(line.strip())], line):
found_numpydocsep += 1
elif isin(self.numpydoc.keywords, line):
found_numpydoc += 1
# TODO: check if not necessary to have > 1??
if found_numpydoc and found_numpydocsep:
detected_style = 'numpydoc'
elif found_googledoc >= found_groups:
detected_style = 'google'
elif found_groups:
detected_style = 'groups'
self.style['in'] = detected_style
return detected_style | python | def autodetect_style(self, data):
"""Determine the style of a docstring,
and sets it as the default input one for the instance.
:param data: the docstring's data to recognize.
:type data: str
:returns: the style detected else 'unknown'
:rtype: str
"""
# evaluate styles with keys
found_keys = defaultdict(int)
for style in self.tagstyles:
for key in self.opt:
found_keys[style] += data.count(self.opt[key][style]['name'])
fkey = max(found_keys, key=found_keys.get)
detected_style = fkey if found_keys[fkey] else 'unknown'
# evaluate styles with groups
if detected_style == 'unknown':
found_groups = 0
found_googledoc = 0
found_numpydoc = 0
found_numpydocsep = 0
for line in data.strip().splitlines():
for key in self.groups:
found_groups += 1 if isin_start(self.groups[key], line) else 0
for key in self.googledoc:
found_googledoc += 1 if isin_start(self.googledoc[key], line) else 0
for key in self.numpydoc:
found_numpydoc += 1 if isin_start(self.numpydoc[key], line) else 0
if line.strip() and isin_alone(['-' * len(line.strip())], line):
found_numpydocsep += 1
elif isin(self.numpydoc.keywords, line):
found_numpydoc += 1
# TODO: check if not necessary to have > 1??
if found_numpydoc and found_numpydocsep:
detected_style = 'numpydoc'
elif found_googledoc >= found_groups:
detected_style = 'google'
elif found_groups:
detected_style = 'groups'
self.style['in'] = detected_style
return detected_style | [
"def",
"autodetect_style",
"(",
"self",
",",
"data",
")",
":",
"# evaluate styles with keys",
"found_keys",
"=",
"defaultdict",
"(",
"int",
")",
"for",
"style",
"in",
"self",
".",
"tagstyles",
":",
"for",
"key",
"in",
"self",
".",
"opt",
":",
"found_keys",
"[",
"style",
"]",
"+=",
"data",
".",
"count",
"(",
"self",
".",
"opt",
"[",
"key",
"]",
"[",
"style",
"]",
"[",
"'name'",
"]",
")",
"fkey",
"=",
"max",
"(",
"found_keys",
",",
"key",
"=",
"found_keys",
".",
"get",
")",
"detected_style",
"=",
"fkey",
"if",
"found_keys",
"[",
"fkey",
"]",
"else",
"'unknown'",
"# evaluate styles with groups",
"if",
"detected_style",
"==",
"'unknown'",
":",
"found_groups",
"=",
"0",
"found_googledoc",
"=",
"0",
"found_numpydoc",
"=",
"0",
"found_numpydocsep",
"=",
"0",
"for",
"line",
"in",
"data",
".",
"strip",
"(",
")",
".",
"splitlines",
"(",
")",
":",
"for",
"key",
"in",
"self",
".",
"groups",
":",
"found_groups",
"+=",
"1",
"if",
"isin_start",
"(",
"self",
".",
"groups",
"[",
"key",
"]",
",",
"line",
")",
"else",
"0",
"for",
"key",
"in",
"self",
".",
"googledoc",
":",
"found_googledoc",
"+=",
"1",
"if",
"isin_start",
"(",
"self",
".",
"googledoc",
"[",
"key",
"]",
",",
"line",
")",
"else",
"0",
"for",
"key",
"in",
"self",
".",
"numpydoc",
":",
"found_numpydoc",
"+=",
"1",
"if",
"isin_start",
"(",
"self",
".",
"numpydoc",
"[",
"key",
"]",
",",
"line",
")",
"else",
"0",
"if",
"line",
".",
"strip",
"(",
")",
"and",
"isin_alone",
"(",
"[",
"'-'",
"*",
"len",
"(",
"line",
".",
"strip",
"(",
")",
")",
"]",
",",
"line",
")",
":",
"found_numpydocsep",
"+=",
"1",
"elif",
"isin",
"(",
"self",
".",
"numpydoc",
".",
"keywords",
",",
"line",
")",
":",
"found_numpydoc",
"+=",
"1",
"# TODO: check if not necessary to have > 1??",
"if",
"found_numpydoc",
"and",
"found_numpydocsep",
":",
"detected_style",
"=",
"'numpydoc'",
"elif",
"found_googledoc",
">=",
"found_groups",
":",
"detected_style",
"=",
"'google'",
"elif",
"found_groups",
":",
"detected_style",
"=",
"'groups'",
"self",
".",
"style",
"[",
"'in'",
"]",
"=",
"detected_style",
"return",
"detected_style"
] | Determine the style of a docstring,
and sets it as the default input one for the instance.
:param data: the docstring's data to recognize.
:type data: str
:returns: the style detected else 'unknown'
:rtype: str | [
"Determine",
"the",
"style",
"of",
"a",
"docstring",
"and",
"sets",
"it",
"as",
"the",
"default",
"input",
"one",
"for",
"the",
"instance",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L647-L693 | train |
dadadel/pyment | pyment/docstring.py | DocsTools._get_options | def _get_options(self, style):
"""Get the list of keywords for a particular style
:param style: the style that the keywords are wanted
"""
return [self.opt[o][style]['name'] for o in self.opt] | python | def _get_options(self, style):
"""Get the list of keywords for a particular style
:param style: the style that the keywords are wanted
"""
return [self.opt[o][style]['name'] for o in self.opt] | [
"def",
"_get_options",
"(",
"self",
",",
"style",
")",
":",
"return",
"[",
"self",
".",
"opt",
"[",
"o",
"]",
"[",
"style",
"]",
"[",
"'name'",
"]",
"for",
"o",
"in",
"self",
".",
"opt",
"]"
] | Get the list of keywords for a particular style
:param style: the style that the keywords are wanted | [
"Get",
"the",
"list",
"of",
"keywords",
"for",
"a",
"particular",
"style"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L713-L719 | train |
dadadel/pyment | pyment/docstring.py | DocsTools.get_group_key_line | def get_group_key_line(self, data, key):
"""Get the next group-style key's line number.
:param data: string to parse
:param key: the key category
:returns: the found line number else -1
"""
idx = -1
for i, line in enumerate(data.splitlines()):
if isin_start(self.groups[key], line):
idx = i
return idx | python | def get_group_key_line(self, data, key):
"""Get the next group-style key's line number.
:param data: string to parse
:param key: the key category
:returns: the found line number else -1
"""
idx = -1
for i, line in enumerate(data.splitlines()):
if isin_start(self.groups[key], line):
idx = i
return idx | [
"def",
"get_group_key_line",
"(",
"self",
",",
"data",
",",
"key",
")",
":",
"idx",
"=",
"-",
"1",
"for",
"i",
",",
"line",
"in",
"enumerate",
"(",
"data",
".",
"splitlines",
"(",
")",
")",
":",
"if",
"isin_start",
"(",
"self",
".",
"groups",
"[",
"key",
"]",
",",
"line",
")",
":",
"idx",
"=",
"i",
"return",
"idx"
] | Get the next group-style key's line number.
:param data: string to parse
:param key: the key category
:returns: the found line number else -1 | [
"Get",
"the",
"next",
"group",
"-",
"style",
"key",
"s",
"line",
"number",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L776-L788 | train |
dadadel/pyment | pyment/docstring.py | DocsTools.get_group_key_index | def get_group_key_index(self, data, key):
"""Get the next groups style's starting line index for a key
:param data: string to parse
:param key: the key category
:returns: the index if found else -1
"""
idx = -1
li = self.get_group_key_line(data, key)
if li != -1:
idx = 0
for line in data.splitlines()[:li]:
idx += len(line) + len('\n')
return idx | python | def get_group_key_index(self, data, key):
"""Get the next groups style's starting line index for a key
:param data: string to parse
:param key: the key category
:returns: the index if found else -1
"""
idx = -1
li = self.get_group_key_line(data, key)
if li != -1:
idx = 0
for line in data.splitlines()[:li]:
idx += len(line) + len('\n')
return idx | [
"def",
"get_group_key_index",
"(",
"self",
",",
"data",
",",
"key",
")",
":",
"idx",
"=",
"-",
"1",
"li",
"=",
"self",
".",
"get_group_key_line",
"(",
"data",
",",
"key",
")",
"if",
"li",
"!=",
"-",
"1",
":",
"idx",
"=",
"0",
"for",
"line",
"in",
"data",
".",
"splitlines",
"(",
")",
"[",
":",
"li",
"]",
":",
"idx",
"+=",
"len",
"(",
"line",
")",
"+",
"len",
"(",
"'\\n'",
")",
"return",
"idx"
] | Get the next groups style's starting line index for a key
:param data: string to parse
:param key: the key category
:returns: the index if found else -1 | [
"Get",
"the",
"next",
"groups",
"style",
"s",
"starting",
"line",
"index",
"for",
"a",
"key"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L794-L808 | train |
dadadel/pyment | pyment/docstring.py | DocsTools.get_group_line | def get_group_line(self, data):
"""Get the next group-style key's line.
:param data: the data to proceed
:returns: the line number
"""
idx = -1
for key in self.groups:
i = self.get_group_key_line(data, key)
if (i < idx and i != -1) or idx == -1:
idx = i
return idx | python | def get_group_line(self, data):
"""Get the next group-style key's line.
:param data: the data to proceed
:returns: the line number
"""
idx = -1
for key in self.groups:
i = self.get_group_key_line(data, key)
if (i < idx and i != -1) or idx == -1:
idx = i
return idx | [
"def",
"get_group_line",
"(",
"self",
",",
"data",
")",
":",
"idx",
"=",
"-",
"1",
"for",
"key",
"in",
"self",
".",
"groups",
":",
"i",
"=",
"self",
".",
"get_group_key_line",
"(",
"data",
",",
"key",
")",
"if",
"(",
"i",
"<",
"idx",
"and",
"i",
"!=",
"-",
"1",
")",
"or",
"idx",
"==",
"-",
"1",
":",
"idx",
"=",
"i",
"return",
"idx"
] | Get the next group-style key's line.
:param data: the data to proceed
:returns: the line number | [
"Get",
"the",
"next",
"group",
"-",
"style",
"key",
"s",
"line",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L810-L822 | train |
dadadel/pyment | pyment/docstring.py | DocsTools.get_group_index | def get_group_index(self, data):
"""Get the next groups style's starting line index
:param data: string to parse
:returns: the index if found else -1
"""
idx = -1
li = self.get_group_line(data)
if li != -1:
idx = 0
for line in data.splitlines()[:li]:
idx += len(line) + len('\n')
return idx | python | def get_group_index(self, data):
"""Get the next groups style's starting line index
:param data: string to parse
:returns: the index if found else -1
"""
idx = -1
li = self.get_group_line(data)
if li != -1:
idx = 0
for line in data.splitlines()[:li]:
idx += len(line) + len('\n')
return idx | [
"def",
"get_group_index",
"(",
"self",
",",
"data",
")",
":",
"idx",
"=",
"-",
"1",
"li",
"=",
"self",
".",
"get_group_line",
"(",
"data",
")",
"if",
"li",
"!=",
"-",
"1",
":",
"idx",
"=",
"0",
"for",
"line",
"in",
"data",
".",
"splitlines",
"(",
")",
"[",
":",
"li",
"]",
":",
"idx",
"+=",
"len",
"(",
"line",
")",
"+",
"len",
"(",
"'\\n'",
")",
"return",
"idx"
] | Get the next groups style's starting line index
:param data: string to parse
:returns: the index if found else -1 | [
"Get",
"the",
"next",
"groups",
"style",
"s",
"starting",
"line",
"index"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L824-L837 | train |
dadadel/pyment | pyment/docstring.py | DocsTools.get_key_index | def get_key_index(self, data, key, starting=True):
"""Get from a docstring the next option with a given key.
:param data: string to parse
:param starting: does the key element must start the line (Default value = True)
:type starting: boolean
:param key: the key category. Can be 'param', 'type', 'return', ...
:returns: index of found element else -1
:rtype: integer
"""
key = self.opt[key][self.style['in']]['name']
if key.startswith(':returns'):
data = data.replace(':return:', ':returns:') # see issue 9
idx = len(data)
ini = 0
loop = True
if key in data:
while loop:
i = data.find(key)
if i != -1:
if starting:
if not data[:i].rstrip(' \t').endswith('\n') and len(data[:i].strip()) > 0:
ini = i + 1
data = data[ini:]
else:
idx = ini + i
loop = False
else:
idx = ini + i
loop = False
else:
loop = False
if idx == len(data):
idx = -1
return idx | python | def get_key_index(self, data, key, starting=True):
"""Get from a docstring the next option with a given key.
:param data: string to parse
:param starting: does the key element must start the line (Default value = True)
:type starting: boolean
:param key: the key category. Can be 'param', 'type', 'return', ...
:returns: index of found element else -1
:rtype: integer
"""
key = self.opt[key][self.style['in']]['name']
if key.startswith(':returns'):
data = data.replace(':return:', ':returns:') # see issue 9
idx = len(data)
ini = 0
loop = True
if key in data:
while loop:
i = data.find(key)
if i != -1:
if starting:
if not data[:i].rstrip(' \t').endswith('\n') and len(data[:i].strip()) > 0:
ini = i + 1
data = data[ini:]
else:
idx = ini + i
loop = False
else:
idx = ini + i
loop = False
else:
loop = False
if idx == len(data):
idx = -1
return idx | [
"def",
"get_key_index",
"(",
"self",
",",
"data",
",",
"key",
",",
"starting",
"=",
"True",
")",
":",
"key",
"=",
"self",
".",
"opt",
"[",
"key",
"]",
"[",
"self",
".",
"style",
"[",
"'in'",
"]",
"]",
"[",
"'name'",
"]",
"if",
"key",
".",
"startswith",
"(",
"':returns'",
")",
":",
"data",
"=",
"data",
".",
"replace",
"(",
"':return:'",
",",
"':returns:'",
")",
"# see issue 9",
"idx",
"=",
"len",
"(",
"data",
")",
"ini",
"=",
"0",
"loop",
"=",
"True",
"if",
"key",
"in",
"data",
":",
"while",
"loop",
":",
"i",
"=",
"data",
".",
"find",
"(",
"key",
")",
"if",
"i",
"!=",
"-",
"1",
":",
"if",
"starting",
":",
"if",
"not",
"data",
"[",
":",
"i",
"]",
".",
"rstrip",
"(",
"' \\t'",
")",
".",
"endswith",
"(",
"'\\n'",
")",
"and",
"len",
"(",
"data",
"[",
":",
"i",
"]",
".",
"strip",
"(",
")",
")",
">",
"0",
":",
"ini",
"=",
"i",
"+",
"1",
"data",
"=",
"data",
"[",
"ini",
":",
"]",
"else",
":",
"idx",
"=",
"ini",
"+",
"i",
"loop",
"=",
"False",
"else",
":",
"idx",
"=",
"ini",
"+",
"i",
"loop",
"=",
"False",
"else",
":",
"loop",
"=",
"False",
"if",
"idx",
"==",
"len",
"(",
"data",
")",
":",
"idx",
"=",
"-",
"1",
"return",
"idx"
] | Get from a docstring the next option with a given key.
:param data: string to parse
:param starting: does the key element must start the line (Default value = True)
:type starting: boolean
:param key: the key category. Can be 'param', 'type', 'return', ...
:returns: index of found element else -1
:rtype: integer | [
"Get",
"from",
"a",
"docstring",
"the",
"next",
"option",
"with",
"a",
"given",
"key",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L839-L874 | train |
dadadel/pyment | pyment/docstring.py | DocString._extract_docs_description | def _extract_docs_description(self):
"""Extract main description from docstring"""
# FIXME: the indentation of descriptions is lost
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
if self.dst.style['in'] == 'groups':
idx = self.dst.get_group_index(data)
elif self.dst.style['in'] == 'google':
lines = data.splitlines()
line_num = self.dst.googledoc.get_next_section_start_line(lines)
if line_num == -1:
idx = -1
else:
idx = len('\n'.join(lines[:line_num]))
elif self.dst.style['in'] == 'numpydoc':
lines = data.splitlines()
line_num = self.dst.numpydoc.get_next_section_start_line(lines)
if line_num == -1:
idx = -1
else:
idx = len('\n'.join(lines[:line_num]))
elif self.dst.style['in'] == 'unknown':
idx = -1
else:
idx = self.dst.get_elem_index(data)
if idx == 0:
self.docs['in']['desc'] = ''
elif idx == -1:
self.docs['in']['desc'] = data
else:
self.docs['in']['desc'] = data[:idx] | python | def _extract_docs_description(self):
"""Extract main description from docstring"""
# FIXME: the indentation of descriptions is lost
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
if self.dst.style['in'] == 'groups':
idx = self.dst.get_group_index(data)
elif self.dst.style['in'] == 'google':
lines = data.splitlines()
line_num = self.dst.googledoc.get_next_section_start_line(lines)
if line_num == -1:
idx = -1
else:
idx = len('\n'.join(lines[:line_num]))
elif self.dst.style['in'] == 'numpydoc':
lines = data.splitlines()
line_num = self.dst.numpydoc.get_next_section_start_line(lines)
if line_num == -1:
idx = -1
else:
idx = len('\n'.join(lines[:line_num]))
elif self.dst.style['in'] == 'unknown':
idx = -1
else:
idx = self.dst.get_elem_index(data)
if idx == 0:
self.docs['in']['desc'] = ''
elif idx == -1:
self.docs['in']['desc'] = data
else:
self.docs['in']['desc'] = data[:idx] | [
"def",
"_extract_docs_description",
"(",
"self",
")",
":",
"# FIXME: the indentation of descriptions is lost",
"data",
"=",
"'\\n'",
".",
"join",
"(",
"[",
"d",
".",
"rstrip",
"(",
")",
".",
"replace",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
",",
"''",
",",
"1",
")",
"for",
"d",
"in",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
".",
"splitlines",
"(",
")",
"]",
")",
"if",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'groups'",
":",
"idx",
"=",
"self",
".",
"dst",
".",
"get_group_index",
"(",
"data",
")",
"elif",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'google'",
":",
"lines",
"=",
"data",
".",
"splitlines",
"(",
")",
"line_num",
"=",
"self",
".",
"dst",
".",
"googledoc",
".",
"get_next_section_start_line",
"(",
"lines",
")",
"if",
"line_num",
"==",
"-",
"1",
":",
"idx",
"=",
"-",
"1",
"else",
":",
"idx",
"=",
"len",
"(",
"'\\n'",
".",
"join",
"(",
"lines",
"[",
":",
"line_num",
"]",
")",
")",
"elif",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'numpydoc'",
":",
"lines",
"=",
"data",
".",
"splitlines",
"(",
")",
"line_num",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_next_section_start_line",
"(",
"lines",
")",
"if",
"line_num",
"==",
"-",
"1",
":",
"idx",
"=",
"-",
"1",
"else",
":",
"idx",
"=",
"len",
"(",
"'\\n'",
".",
"join",
"(",
"lines",
"[",
":",
"line_num",
"]",
")",
")",
"elif",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'unknown'",
":",
"idx",
"=",
"-",
"1",
"else",
":",
"idx",
"=",
"self",
".",
"dst",
".",
"get_elem_index",
"(",
"data",
")",
"if",
"idx",
"==",
"0",
":",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'desc'",
"]",
"=",
"''",
"elif",
"idx",
"==",
"-",
"1",
":",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'desc'",
"]",
"=",
"data",
"else",
":",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'desc'",
"]",
"=",
"data",
"[",
":",
"idx",
"]"
] | Extract main description from docstring | [
"Extract",
"main",
"description",
"from",
"docstring"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1387-L1416 | train |
dadadel/pyment | pyment/docstring.py | DocString._extract_groupstyle_docs_params | def _extract_groupstyle_docs_params(self):
"""Extract group style parameters"""
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
idx = self.dst.get_group_key_line(data, 'param')
if idx >= 0:
data = data.splitlines()[idx + 1:]
end = self.dst.get_group_line('\n'.join(data))
end = end if end != -1 else len(data)
for i in range(end):
# FIXME: see how retrieve multiline param description and how get type
line = data[i]
param = None
desc = ''
ptype = ''
m = re.match(r'^\W*(\w+)[\W\s]+(\w[\s\w]+)', line.strip())
if m:
param = m.group(1).strip()
desc = m.group(2).strip()
else:
m = re.match(r'^\W*(\w+)\W*', line.strip())
if m:
param = m.group(1).strip()
if param:
self.docs['in']['params'].append((param, desc, ptype)) | python | def _extract_groupstyle_docs_params(self):
"""Extract group style parameters"""
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
idx = self.dst.get_group_key_line(data, 'param')
if idx >= 0:
data = data.splitlines()[idx + 1:]
end = self.dst.get_group_line('\n'.join(data))
end = end if end != -1 else len(data)
for i in range(end):
# FIXME: see how retrieve multiline param description and how get type
line = data[i]
param = None
desc = ''
ptype = ''
m = re.match(r'^\W*(\w+)[\W\s]+(\w[\s\w]+)', line.strip())
if m:
param = m.group(1).strip()
desc = m.group(2).strip()
else:
m = re.match(r'^\W*(\w+)\W*', line.strip())
if m:
param = m.group(1).strip()
if param:
self.docs['in']['params'].append((param, desc, ptype)) | [
"def",
"_extract_groupstyle_docs_params",
"(",
"self",
")",
":",
"data",
"=",
"'\\n'",
".",
"join",
"(",
"[",
"d",
".",
"rstrip",
"(",
")",
".",
"replace",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
",",
"''",
",",
"1",
")",
"for",
"d",
"in",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
".",
"splitlines",
"(",
")",
"]",
")",
"idx",
"=",
"self",
".",
"dst",
".",
"get_group_key_line",
"(",
"data",
",",
"'param'",
")",
"if",
"idx",
">=",
"0",
":",
"data",
"=",
"data",
".",
"splitlines",
"(",
")",
"[",
"idx",
"+",
"1",
":",
"]",
"end",
"=",
"self",
".",
"dst",
".",
"get_group_line",
"(",
"'\\n'",
".",
"join",
"(",
"data",
")",
")",
"end",
"=",
"end",
"if",
"end",
"!=",
"-",
"1",
"else",
"len",
"(",
"data",
")",
"for",
"i",
"in",
"range",
"(",
"end",
")",
":",
"# FIXME: see how retrieve multiline param description and how get type",
"line",
"=",
"data",
"[",
"i",
"]",
"param",
"=",
"None",
"desc",
"=",
"''",
"ptype",
"=",
"''",
"m",
"=",
"re",
".",
"match",
"(",
"r'^\\W*(\\w+)[\\W\\s]+(\\w[\\s\\w]+)'",
",",
"line",
".",
"strip",
"(",
")",
")",
"if",
"m",
":",
"param",
"=",
"m",
".",
"group",
"(",
"1",
")",
".",
"strip",
"(",
")",
"desc",
"=",
"m",
".",
"group",
"(",
"2",
")",
".",
"strip",
"(",
")",
"else",
":",
"m",
"=",
"re",
".",
"match",
"(",
"r'^\\W*(\\w+)\\W*'",
",",
"line",
".",
"strip",
"(",
")",
")",
"if",
"m",
":",
"param",
"=",
"m",
".",
"group",
"(",
"1",
")",
".",
"strip",
"(",
")",
"if",
"param",
":",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'params'",
"]",
".",
"append",
"(",
"(",
"param",
",",
"desc",
",",
"ptype",
")",
")"
] | Extract group style parameters | [
"Extract",
"group",
"style",
"parameters"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1418-L1441 | train |
dadadel/pyment | pyment/docstring.py | DocString._extract_docs_return | def _extract_docs_return(self):
"""Extract return description and type"""
if self.dst.style['in'] == 'numpydoc':
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
self.docs['in']['return'] = self.dst.numpydoc.get_return_list(data)
self.docs['in']['rtype'] = None
# TODO: fix this
elif self.dst.style['in'] == 'google':
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
self.docs['in']['return'] = self.dst.googledoc.get_return_list(data)
self.docs['in']['rtype'] = None
elif self.dst.style['in'] == 'groups':
self._extract_groupstyle_docs_return()
elif self.dst.style['in'] in ['javadoc', 'reST']:
self._extract_tagstyle_docs_return() | python | def _extract_docs_return(self):
"""Extract return description and type"""
if self.dst.style['in'] == 'numpydoc':
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
self.docs['in']['return'] = self.dst.numpydoc.get_return_list(data)
self.docs['in']['rtype'] = None
# TODO: fix this
elif self.dst.style['in'] == 'google':
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
self.docs['in']['return'] = self.dst.googledoc.get_return_list(data)
self.docs['in']['rtype'] = None
elif self.dst.style['in'] == 'groups':
self._extract_groupstyle_docs_return()
elif self.dst.style['in'] in ['javadoc', 'reST']:
self._extract_tagstyle_docs_return() | [
"def",
"_extract_docs_return",
"(",
"self",
")",
":",
"if",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'numpydoc'",
":",
"data",
"=",
"'\\n'",
".",
"join",
"(",
"[",
"d",
".",
"rstrip",
"(",
")",
".",
"replace",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
",",
"''",
",",
"1",
")",
"for",
"d",
"in",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
".",
"splitlines",
"(",
")",
"]",
")",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'return'",
"]",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_return_list",
"(",
"data",
")",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'rtype'",
"]",
"=",
"None",
"# TODO: fix this",
"elif",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'google'",
":",
"data",
"=",
"'\\n'",
".",
"join",
"(",
"[",
"d",
".",
"rstrip",
"(",
")",
".",
"replace",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
",",
"''",
",",
"1",
")",
"for",
"d",
"in",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
".",
"splitlines",
"(",
")",
"]",
")",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'return'",
"]",
"=",
"self",
".",
"dst",
".",
"googledoc",
".",
"get_return_list",
"(",
"data",
")",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'rtype'",
"]",
"=",
"None",
"elif",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'groups'",
":",
"self",
".",
"_extract_groupstyle_docs_return",
"(",
")",
"elif",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"in",
"[",
"'javadoc'",
",",
"'reST'",
"]",
":",
"self",
".",
"_extract_tagstyle_docs_return",
"(",
")"
] | Extract return description and type | [
"Extract",
"return",
"description",
"and",
"type"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1585-L1599 | train |
dadadel/pyment | pyment/docstring.py | DocString._extract_docs_other | def _extract_docs_other(self):
"""Extract other specific sections"""
if self.dst.style['in'] == 'numpydoc':
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
lst = self.dst.numpydoc.get_list_key(data, 'also')
lst = self.dst.numpydoc.get_list_key(data, 'ref')
lst = self.dst.numpydoc.get_list_key(data, 'note')
lst = self.dst.numpydoc.get_list_key(data, 'other')
lst = self.dst.numpydoc.get_list_key(data, 'example')
lst = self.dst.numpydoc.get_list_key(data, 'attr') | python | def _extract_docs_other(self):
"""Extract other specific sections"""
if self.dst.style['in'] == 'numpydoc':
data = '\n'.join([d.rstrip().replace(self.docs['out']['spaces'], '', 1) for d in self.docs['in']['raw'].splitlines()])
lst = self.dst.numpydoc.get_list_key(data, 'also')
lst = self.dst.numpydoc.get_list_key(data, 'ref')
lst = self.dst.numpydoc.get_list_key(data, 'note')
lst = self.dst.numpydoc.get_list_key(data, 'other')
lst = self.dst.numpydoc.get_list_key(data, 'example')
lst = self.dst.numpydoc.get_list_key(data, 'attr') | [
"def",
"_extract_docs_other",
"(",
"self",
")",
":",
"if",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'numpydoc'",
":",
"data",
"=",
"'\\n'",
".",
"join",
"(",
"[",
"d",
".",
"rstrip",
"(",
")",
".",
"replace",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
",",
"''",
",",
"1",
")",
"for",
"d",
"in",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
".",
"splitlines",
"(",
")",
"]",
")",
"lst",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_list_key",
"(",
"data",
",",
"'also'",
")",
"lst",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_list_key",
"(",
"data",
",",
"'ref'",
")",
"lst",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_list_key",
"(",
"data",
",",
"'note'",
")",
"lst",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_list_key",
"(",
"data",
",",
"'other'",
")",
"lst",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_list_key",
"(",
"data",
",",
"'example'",
")",
"lst",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_list_key",
"(",
"data",
",",
"'attr'",
")"
] | Extract other specific sections | [
"Extract",
"other",
"specific",
"sections"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1601-L1610 | train |
dadadel/pyment | pyment/docstring.py | DocString._set_desc | def _set_desc(self):
"""Sets the global description if any"""
# TODO: manage different in/out styles
if self.docs['in']['desc']:
self.docs['out']['desc'] = self.docs['in']['desc']
else:
self.docs['out']['desc'] = '' | python | def _set_desc(self):
"""Sets the global description if any"""
# TODO: manage different in/out styles
if self.docs['in']['desc']:
self.docs['out']['desc'] = self.docs['in']['desc']
else:
self.docs['out']['desc'] = '' | [
"def",
"_set_desc",
"(",
"self",
")",
":",
"# TODO: manage different in/out styles",
"if",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'desc'",
"]",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'desc'",
"]",
"=",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'desc'",
"]",
"else",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'desc'",
"]",
"=",
"''"
] | Sets the global description if any | [
"Sets",
"the",
"global",
"description",
"if",
"any"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1638-L1644 | train |
dadadel/pyment | pyment/docstring.py | DocString._set_params | def _set_params(self):
"""Sets the parameters with types, descriptions and default value if any"""
# TODO: manage different in/out styles
if self.docs['in']['params']:
# list of parameters is like: (name, description, type)
self.docs['out']['params'] = list(self.docs['in']['params'])
for e in self.element['params']:
if type(e) is tuple:
# tuple is: (name, default)
param = e[0]
else:
param = e
found = False
for i, p in enumerate(self.docs['out']['params']):
if param == p[0]:
found = True
# add default value if any
if type(e) is tuple:
# param will contain: (name, desc, type, default)
self.docs['out']['params'][i] = (p[0], p[1], p[2], e[1])
if not found:
if type(e) is tuple:
p = (param, '', None, e[1])
else:
p = (param, '', None, None)
self.docs['out']['params'].append(p) | python | def _set_params(self):
"""Sets the parameters with types, descriptions and default value if any"""
# TODO: manage different in/out styles
if self.docs['in']['params']:
# list of parameters is like: (name, description, type)
self.docs['out']['params'] = list(self.docs['in']['params'])
for e in self.element['params']:
if type(e) is tuple:
# tuple is: (name, default)
param = e[0]
else:
param = e
found = False
for i, p in enumerate(self.docs['out']['params']):
if param == p[0]:
found = True
# add default value if any
if type(e) is tuple:
# param will contain: (name, desc, type, default)
self.docs['out']['params'][i] = (p[0], p[1], p[2], e[1])
if not found:
if type(e) is tuple:
p = (param, '', None, e[1])
else:
p = (param, '', None, None)
self.docs['out']['params'].append(p) | [
"def",
"_set_params",
"(",
"self",
")",
":",
"# TODO: manage different in/out styles",
"if",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'params'",
"]",
":",
"# list of parameters is like: (name, description, type)",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'params'",
"]",
"=",
"list",
"(",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'params'",
"]",
")",
"for",
"e",
"in",
"self",
".",
"element",
"[",
"'params'",
"]",
":",
"if",
"type",
"(",
"e",
")",
"is",
"tuple",
":",
"# tuple is: (name, default)",
"param",
"=",
"e",
"[",
"0",
"]",
"else",
":",
"param",
"=",
"e",
"found",
"=",
"False",
"for",
"i",
",",
"p",
"in",
"enumerate",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'params'",
"]",
")",
":",
"if",
"param",
"==",
"p",
"[",
"0",
"]",
":",
"found",
"=",
"True",
"# add default value if any",
"if",
"type",
"(",
"e",
")",
"is",
"tuple",
":",
"# param will contain: (name, desc, type, default)",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'params'",
"]",
"[",
"i",
"]",
"=",
"(",
"p",
"[",
"0",
"]",
",",
"p",
"[",
"1",
"]",
",",
"p",
"[",
"2",
"]",
",",
"e",
"[",
"1",
"]",
")",
"if",
"not",
"found",
":",
"if",
"type",
"(",
"e",
")",
"is",
"tuple",
":",
"p",
"=",
"(",
"param",
",",
"''",
",",
"None",
",",
"e",
"[",
"1",
"]",
")",
"else",
":",
"p",
"=",
"(",
"param",
",",
"''",
",",
"None",
",",
"None",
")",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'params'",
"]",
".",
"append",
"(",
"p",
")"
] | Sets the parameters with types, descriptions and default value if any | [
"Sets",
"the",
"parameters",
"with",
"types",
"descriptions",
"and",
"default",
"value",
"if",
"any"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1646-L1671 | train |
dadadel/pyment | pyment/docstring.py | DocString._set_raises | def _set_raises(self):
"""Sets the raises and descriptions"""
# TODO: manage different in/out styles
# manage setting if not mandatory for numpy but optional
if self.docs['in']['raises']:
if self.dst.style['out'] != 'numpydoc' or self.dst.style['in'] == 'numpydoc' or \
(self.dst.style['out'] == 'numpydoc' and
'raise' not in self.dst.numpydoc.get_excluded_sections()):
# list of parameters is like: (name, description)
self.docs['out']['raises'] = list(self.docs['in']['raises']) | python | def _set_raises(self):
"""Sets the raises and descriptions"""
# TODO: manage different in/out styles
# manage setting if not mandatory for numpy but optional
if self.docs['in']['raises']:
if self.dst.style['out'] != 'numpydoc' or self.dst.style['in'] == 'numpydoc' or \
(self.dst.style['out'] == 'numpydoc' and
'raise' not in self.dst.numpydoc.get_excluded_sections()):
# list of parameters is like: (name, description)
self.docs['out']['raises'] = list(self.docs['in']['raises']) | [
"def",
"_set_raises",
"(",
"self",
")",
":",
"# TODO: manage different in/out styles",
"# manage setting if not mandatory for numpy but optional",
"if",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raises'",
"]",
":",
"if",
"self",
".",
"dst",
".",
"style",
"[",
"'out'",
"]",
"!=",
"'numpydoc'",
"or",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'numpydoc'",
"or",
"(",
"self",
".",
"dst",
".",
"style",
"[",
"'out'",
"]",
"==",
"'numpydoc'",
"and",
"'raise'",
"not",
"in",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_excluded_sections",
"(",
")",
")",
":",
"# list of parameters is like: (name, description)",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'raises'",
"]",
"=",
"list",
"(",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raises'",
"]",
")"
] | Sets the raises and descriptions | [
"Sets",
"the",
"raises",
"and",
"descriptions"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1673-L1682 | train |
dadadel/pyment | pyment/docstring.py | DocString._set_return | def _set_return(self):
"""Sets the return parameter with description and rtype if any"""
# TODO: manage return retrieved from element code (external)
# TODO: manage different in/out styles
if type(self.docs['in']['return']) is list and self.dst.style['out'] not in ['groups', 'numpydoc', 'google']:
# TODO: manage return names
# manage not setting return if not mandatory for numpy
lst = self.docs['in']['return']
if lst:
if lst[0][0] is not None:
self.docs['out']['return'] = "%s-> %s" % (lst[0][0], lst[0][1])
else:
self.docs['out']['return'] = lst[0][1]
self.docs['out']['rtype'] = lst[0][2]
else:
self.docs['out']['return'] = self.docs['in']['return']
self.docs['out']['rtype'] = self.docs['in']['rtype'] | python | def _set_return(self):
"""Sets the return parameter with description and rtype if any"""
# TODO: manage return retrieved from element code (external)
# TODO: manage different in/out styles
if type(self.docs['in']['return']) is list and self.dst.style['out'] not in ['groups', 'numpydoc', 'google']:
# TODO: manage return names
# manage not setting return if not mandatory for numpy
lst = self.docs['in']['return']
if lst:
if lst[0][0] is not None:
self.docs['out']['return'] = "%s-> %s" % (lst[0][0], lst[0][1])
else:
self.docs['out']['return'] = lst[0][1]
self.docs['out']['rtype'] = lst[0][2]
else:
self.docs['out']['return'] = self.docs['in']['return']
self.docs['out']['rtype'] = self.docs['in']['rtype'] | [
"def",
"_set_return",
"(",
"self",
")",
":",
"# TODO: manage return retrieved from element code (external)",
"# TODO: manage different in/out styles",
"if",
"type",
"(",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'return'",
"]",
")",
"is",
"list",
"and",
"self",
".",
"dst",
".",
"style",
"[",
"'out'",
"]",
"not",
"in",
"[",
"'groups'",
",",
"'numpydoc'",
",",
"'google'",
"]",
":",
"# TODO: manage return names",
"# manage not setting return if not mandatory for numpy",
"lst",
"=",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'return'",
"]",
"if",
"lst",
":",
"if",
"lst",
"[",
"0",
"]",
"[",
"0",
"]",
"is",
"not",
"None",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'return'",
"]",
"=",
"\"%s-> %s\"",
"%",
"(",
"lst",
"[",
"0",
"]",
"[",
"0",
"]",
",",
"lst",
"[",
"0",
"]",
"[",
"1",
"]",
")",
"else",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'return'",
"]",
"=",
"lst",
"[",
"0",
"]",
"[",
"1",
"]",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'rtype'",
"]",
"=",
"lst",
"[",
"0",
"]",
"[",
"2",
"]",
"else",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'return'",
"]",
"=",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'return'",
"]",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'rtype'",
"]",
"=",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'rtype'",
"]"
] | Sets the return parameter with description and rtype if any | [
"Sets",
"the",
"return",
"parameter",
"with",
"description",
"and",
"rtype",
"if",
"any"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1684-L1700 | train |
dadadel/pyment | pyment/docstring.py | DocString._set_other | def _set_other(self):
"""Sets other specific sections"""
# manage not setting if not mandatory for numpy
if self.dst.style['in'] == 'numpydoc':
if self.docs['in']['raw'] is not None:
self.docs['out']['post'] = self.dst.numpydoc.get_raw_not_managed(self.docs['in']['raw'])
elif 'post' not in self.docs['out'] or self.docs['out']['post'] is None:
self.docs['out']['post'] = '' | python | def _set_other(self):
"""Sets other specific sections"""
# manage not setting if not mandatory for numpy
if self.dst.style['in'] == 'numpydoc':
if self.docs['in']['raw'] is not None:
self.docs['out']['post'] = self.dst.numpydoc.get_raw_not_managed(self.docs['in']['raw'])
elif 'post' not in self.docs['out'] or self.docs['out']['post'] is None:
self.docs['out']['post'] = '' | [
"def",
"_set_other",
"(",
"self",
")",
":",
"# manage not setting if not mandatory for numpy",
"if",
"self",
".",
"dst",
".",
"style",
"[",
"'in'",
"]",
"==",
"'numpydoc'",
":",
"if",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
"is",
"not",
"None",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'post'",
"]",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"get_raw_not_managed",
"(",
"self",
".",
"docs",
"[",
"'in'",
"]",
"[",
"'raw'",
"]",
")",
"elif",
"'post'",
"not",
"in",
"self",
".",
"docs",
"[",
"'out'",
"]",
"or",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'post'",
"]",
"is",
"None",
":",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'post'",
"]",
"=",
"''"
] | Sets other specific sections | [
"Sets",
"other",
"specific",
"sections"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1702-L1709 | train |
dadadel/pyment | pyment/docstring.py | DocString._set_raw | def _set_raw(self):
"""Sets the output raw docstring"""
sep = self.dst.get_sep(target='out')
sep = sep + ' ' if sep != ' ' else sep
with_space = lambda s: '\n'.join([self.docs['out']['spaces'] + l if i > 0 else l for i, l in enumerate(s.splitlines())])
# sets the description section
raw = self.docs['out']['spaces'] + self.quotes
desc = self.docs['out']['desc'].strip()
if not desc or not desc.count('\n'):
if not self.docs['out']['params'] and not self.docs['out']['return'] and not self.docs['out']['rtype'] and not self.docs['out']['raises']:
raw += desc if desc else self.trailing_space
raw += self.quotes
self.docs['out']['raw'] = raw.rstrip()
return
if not self.first_line:
raw += '\n' + self.docs['out']['spaces']
raw += with_space(self.docs['out']['desc']).strip() + '\n'
# sets the parameters section
raw += self._set_raw_params(sep)
# sets the return section
raw += self._set_raw_return(sep)
# sets the raises section
raw += self._set_raw_raise(sep)
# sets post specific if any
if 'post' in self.docs['out']:
raw += self.docs['out']['spaces'] + with_space(self.docs['out']['post']).strip() + '\n'
# sets the doctests if any
if 'doctests' in self.docs['out']:
raw += self.docs['out']['spaces'] + with_space(self.docs['out']['doctests']).strip() + '\n'
if raw.count(self.quotes) == 1:
raw += self.docs['out']['spaces'] + self.quotes
self.docs['out']['raw'] = raw.rstrip() | python | def _set_raw(self):
"""Sets the output raw docstring"""
sep = self.dst.get_sep(target='out')
sep = sep + ' ' if sep != ' ' else sep
with_space = lambda s: '\n'.join([self.docs['out']['spaces'] + l if i > 0 else l for i, l in enumerate(s.splitlines())])
# sets the description section
raw = self.docs['out']['spaces'] + self.quotes
desc = self.docs['out']['desc'].strip()
if not desc or not desc.count('\n'):
if not self.docs['out']['params'] and not self.docs['out']['return'] and not self.docs['out']['rtype'] and not self.docs['out']['raises']:
raw += desc if desc else self.trailing_space
raw += self.quotes
self.docs['out']['raw'] = raw.rstrip()
return
if not self.first_line:
raw += '\n' + self.docs['out']['spaces']
raw += with_space(self.docs['out']['desc']).strip() + '\n'
# sets the parameters section
raw += self._set_raw_params(sep)
# sets the return section
raw += self._set_raw_return(sep)
# sets the raises section
raw += self._set_raw_raise(sep)
# sets post specific if any
if 'post' in self.docs['out']:
raw += self.docs['out']['spaces'] + with_space(self.docs['out']['post']).strip() + '\n'
# sets the doctests if any
if 'doctests' in self.docs['out']:
raw += self.docs['out']['spaces'] + with_space(self.docs['out']['doctests']).strip() + '\n'
if raw.count(self.quotes) == 1:
raw += self.docs['out']['spaces'] + self.quotes
self.docs['out']['raw'] = raw.rstrip() | [
"def",
"_set_raw",
"(",
"self",
")",
":",
"sep",
"=",
"self",
".",
"dst",
".",
"get_sep",
"(",
"target",
"=",
"'out'",
")",
"sep",
"=",
"sep",
"+",
"' '",
"if",
"sep",
"!=",
"' '",
"else",
"sep",
"with_space",
"=",
"lambda",
"s",
":",
"'\\n'",
".",
"join",
"(",
"[",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
"+",
"l",
"if",
"i",
">",
"0",
"else",
"l",
"for",
"i",
",",
"l",
"in",
"enumerate",
"(",
"s",
".",
"splitlines",
"(",
")",
")",
"]",
")",
"# sets the description section",
"raw",
"=",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
"+",
"self",
".",
"quotes",
"desc",
"=",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'desc'",
"]",
".",
"strip",
"(",
")",
"if",
"not",
"desc",
"or",
"not",
"desc",
".",
"count",
"(",
"'\\n'",
")",
":",
"if",
"not",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'params'",
"]",
"and",
"not",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'return'",
"]",
"and",
"not",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'rtype'",
"]",
"and",
"not",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'raises'",
"]",
":",
"raw",
"+=",
"desc",
"if",
"desc",
"else",
"self",
".",
"trailing_space",
"raw",
"+=",
"self",
".",
"quotes",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'raw'",
"]",
"=",
"raw",
".",
"rstrip",
"(",
")",
"return",
"if",
"not",
"self",
".",
"first_line",
":",
"raw",
"+=",
"'\\n'",
"+",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
"raw",
"+=",
"with_space",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'desc'",
"]",
")",
".",
"strip",
"(",
")",
"+",
"'\\n'",
"# sets the parameters section",
"raw",
"+=",
"self",
".",
"_set_raw_params",
"(",
"sep",
")",
"# sets the return section",
"raw",
"+=",
"self",
".",
"_set_raw_return",
"(",
"sep",
")",
"# sets the raises section",
"raw",
"+=",
"self",
".",
"_set_raw_raise",
"(",
"sep",
")",
"# sets post specific if any",
"if",
"'post'",
"in",
"self",
".",
"docs",
"[",
"'out'",
"]",
":",
"raw",
"+=",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
"+",
"with_space",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'post'",
"]",
")",
".",
"strip",
"(",
")",
"+",
"'\\n'",
"# sets the doctests if any",
"if",
"'doctests'",
"in",
"self",
".",
"docs",
"[",
"'out'",
"]",
":",
"raw",
"+=",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
"+",
"with_space",
"(",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'doctests'",
"]",
")",
".",
"strip",
"(",
")",
"+",
"'\\n'",
"if",
"raw",
".",
"count",
"(",
"self",
".",
"quotes",
")",
"==",
"1",
":",
"raw",
"+=",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'spaces'",
"]",
"+",
"self",
".",
"quotes",
"self",
".",
"docs",
"[",
"'out'",
"]",
"[",
"'raw'",
"]",
"=",
"raw",
".",
"rstrip",
"(",
")"
] | Sets the output raw docstring | [
"Sets",
"the",
"output",
"raw",
"docstring"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1912-L1950 | train |
dadadel/pyment | pyment/docstring.py | DocString.generate_docs | def generate_docs(self):
"""Generates the output docstring"""
if self.dst.style['out'] == 'numpydoc' and self.dst.numpydoc.first_line is not None:
self.first_line = self.dst.numpydoc.first_line
self._set_desc()
self._set_params()
self._set_return()
self._set_raises()
self._set_other()
self._set_raw()
self.generated_docs = True | python | def generate_docs(self):
"""Generates the output docstring"""
if self.dst.style['out'] == 'numpydoc' and self.dst.numpydoc.first_line is not None:
self.first_line = self.dst.numpydoc.first_line
self._set_desc()
self._set_params()
self._set_return()
self._set_raises()
self._set_other()
self._set_raw()
self.generated_docs = True | [
"def",
"generate_docs",
"(",
"self",
")",
":",
"if",
"self",
".",
"dst",
".",
"style",
"[",
"'out'",
"]",
"==",
"'numpydoc'",
"and",
"self",
".",
"dst",
".",
"numpydoc",
".",
"first_line",
"is",
"not",
"None",
":",
"self",
".",
"first_line",
"=",
"self",
".",
"dst",
".",
"numpydoc",
".",
"first_line",
"self",
".",
"_set_desc",
"(",
")",
"self",
".",
"_set_params",
"(",
")",
"self",
".",
"_set_return",
"(",
")",
"self",
".",
"_set_raises",
"(",
")",
"self",
".",
"_set_other",
"(",
")",
"self",
".",
"_set_raw",
"(",
")",
"self",
".",
"generated_docs",
"=",
"True"
] | Generates the output docstring | [
"Generates",
"the",
"output",
"docstring"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/docstring.py#L1952-L1962 | train |
dadadel/pyment | pyment/pymentapp.py | get_files_from_dir | def get_files_from_dir(path, recursive=True, depth=0, file_ext='.py'):
"""Retrieve the list of files from a folder.
@param path: file or directory where to search files
@param recursive: if True will search also sub-directories
@param depth: if explore recursively, the depth of sub directories to follow
@param file_ext: the files extension to get. Default is '.py'
@return: the file list retrieved. if the input is a file then a one element list.
"""
file_list = []
if os.path.isfile(path) or path == '-':
return [path]
if path[-1] != os.sep:
path = path + os.sep
for f in glob.glob(path + "*"):
if os.path.isdir(f):
if depth < MAX_DEPTH_RECUR: # avoid infinite recursive loop
file_list.extend(get_files_from_dir(f, recursive, depth + 1))
else:
continue
elif f.endswith(file_ext):
file_list.append(f)
return file_list | python | def get_files_from_dir(path, recursive=True, depth=0, file_ext='.py'):
"""Retrieve the list of files from a folder.
@param path: file or directory where to search files
@param recursive: if True will search also sub-directories
@param depth: if explore recursively, the depth of sub directories to follow
@param file_ext: the files extension to get. Default is '.py'
@return: the file list retrieved. if the input is a file then a one element list.
"""
file_list = []
if os.path.isfile(path) or path == '-':
return [path]
if path[-1] != os.sep:
path = path + os.sep
for f in glob.glob(path + "*"):
if os.path.isdir(f):
if depth < MAX_DEPTH_RECUR: # avoid infinite recursive loop
file_list.extend(get_files_from_dir(f, recursive, depth + 1))
else:
continue
elif f.endswith(file_ext):
file_list.append(f)
return file_list | [
"def",
"get_files_from_dir",
"(",
"path",
",",
"recursive",
"=",
"True",
",",
"depth",
"=",
"0",
",",
"file_ext",
"=",
"'.py'",
")",
":",
"file_list",
"=",
"[",
"]",
"if",
"os",
".",
"path",
".",
"isfile",
"(",
"path",
")",
"or",
"path",
"==",
"'-'",
":",
"return",
"[",
"path",
"]",
"if",
"path",
"[",
"-",
"1",
"]",
"!=",
"os",
".",
"sep",
":",
"path",
"=",
"path",
"+",
"os",
".",
"sep",
"for",
"f",
"in",
"glob",
".",
"glob",
"(",
"path",
"+",
"\"*\"",
")",
":",
"if",
"os",
".",
"path",
".",
"isdir",
"(",
"f",
")",
":",
"if",
"depth",
"<",
"MAX_DEPTH_RECUR",
":",
"# avoid infinite recursive loop",
"file_list",
".",
"extend",
"(",
"get_files_from_dir",
"(",
"f",
",",
"recursive",
",",
"depth",
"+",
"1",
")",
")",
"else",
":",
"continue",
"elif",
"f",
".",
"endswith",
"(",
"file_ext",
")",
":",
"file_list",
".",
"append",
"(",
"f",
")",
"return",
"file_list"
] | Retrieve the list of files from a folder.
@param path: file or directory where to search files
@param recursive: if True will search also sub-directories
@param depth: if explore recursively, the depth of sub directories to follow
@param file_ext: the files extension to get. Default is '.py'
@return: the file list retrieved. if the input is a file then a one element list. | [
"Retrieve",
"the",
"list",
"of",
"files",
"from",
"a",
"folder",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pymentapp.py#L17-L40 | train |
dadadel/pyment | pyment/pymentapp.py | get_config | def get_config(config_file):
"""Get the configuration from a file.
@param config_file: the configuration file
@return: the configuration
@rtype: dict
"""
config = {}
tobool = lambda s: True if s.lower() == 'true' else False
if config_file:
try:
f = open(config_file, 'r')
except:
print ("Unable to open configuration file '{0}'".format(config_file))
else:
for line in f.readlines():
if len(line.strip()):
key, value = line.split("=", 1)
key, value = key.strip(), value.strip()
if key in ['init2class', 'first_line', 'convert_only']:
value = tobool(value)
config[key] = value
return config | python | def get_config(config_file):
"""Get the configuration from a file.
@param config_file: the configuration file
@return: the configuration
@rtype: dict
"""
config = {}
tobool = lambda s: True if s.lower() == 'true' else False
if config_file:
try:
f = open(config_file, 'r')
except:
print ("Unable to open configuration file '{0}'".format(config_file))
else:
for line in f.readlines():
if len(line.strip()):
key, value = line.split("=", 1)
key, value = key.strip(), value.strip()
if key in ['init2class', 'first_line', 'convert_only']:
value = tobool(value)
config[key] = value
return config | [
"def",
"get_config",
"(",
"config_file",
")",
":",
"config",
"=",
"{",
"}",
"tobool",
"=",
"lambda",
"s",
":",
"True",
"if",
"s",
".",
"lower",
"(",
")",
"==",
"'true'",
"else",
"False",
"if",
"config_file",
":",
"try",
":",
"f",
"=",
"open",
"(",
"config_file",
",",
"'r'",
")",
"except",
":",
"print",
"(",
"\"Unable to open configuration file '{0}'\"",
".",
"format",
"(",
"config_file",
")",
")",
"else",
":",
"for",
"line",
"in",
"f",
".",
"readlines",
"(",
")",
":",
"if",
"len",
"(",
"line",
".",
"strip",
"(",
")",
")",
":",
"key",
",",
"value",
"=",
"line",
".",
"split",
"(",
"\"=\"",
",",
"1",
")",
"key",
",",
"value",
"=",
"key",
".",
"strip",
"(",
")",
",",
"value",
".",
"strip",
"(",
")",
"if",
"key",
"in",
"[",
"'init2class'",
",",
"'first_line'",
",",
"'convert_only'",
"]",
":",
"value",
"=",
"tobool",
"(",
"value",
")",
"config",
"[",
"key",
"]",
"=",
"value",
"return",
"config"
] | Get the configuration from a file.
@param config_file: the configuration file
@return: the configuration
@rtype: dict | [
"Get",
"the",
"configuration",
"from",
"a",
"file",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pymentapp.py#L43-L66 | train |
dadadel/pyment | pyment/pyment.py | PyComment.get_output_docs | def get_output_docs(self):
"""Return the output docstrings once formatted
:returns: the formatted docstrings
:rtype: list
"""
if not self.parsed:
self._parse()
lst = []
for e in self.docs_list:
lst.append(e['docs'].get_raw_docs())
return lst | python | def get_output_docs(self):
"""Return the output docstrings once formatted
:returns: the formatted docstrings
:rtype: list
"""
if not self.parsed:
self._parse()
lst = []
for e in self.docs_list:
lst.append(e['docs'].get_raw_docs())
return lst | [
"def",
"get_output_docs",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"parsed",
":",
"self",
".",
"_parse",
"(",
")",
"lst",
"=",
"[",
"]",
"for",
"e",
"in",
"self",
".",
"docs_list",
":",
"lst",
".",
"append",
"(",
"e",
"[",
"'docs'",
"]",
".",
"get_raw_docs",
"(",
")",
")",
"return",
"lst"
] | Return the output docstrings once formatted
:returns: the formatted docstrings
:rtype: list | [
"Return",
"the",
"output",
"docstrings",
"once",
"formatted"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pyment.py#L222-L234 | train |
dadadel/pyment | pyment/pyment.py | PyComment.compute_before_after | def compute_before_after(self):
"""Compute the list of lines before and after the proposed docstring changes.
:return: tuple of before,after where each is a list of lines of python code.
"""
if not self.parsed:
self._parse()
list_from = self.input_lines
list_to = []
last = 0
for e in self.docs_list:
start, end = e['location']
if start <= 0:
start, end = -start, -end
list_to.extend(list_from[last:start + 1])
else:
list_to.extend(list_from[last:start])
docs = e['docs'].get_raw_docs()
list_docs = [l + '\n' for l in docs.splitlines()]
list_to.extend(list_docs)
last = end + 1
if last < len(list_from):
list_to.extend(list_from[last:])
return list_from, list_to | python | def compute_before_after(self):
"""Compute the list of lines before and after the proposed docstring changes.
:return: tuple of before,after where each is a list of lines of python code.
"""
if not self.parsed:
self._parse()
list_from = self.input_lines
list_to = []
last = 0
for e in self.docs_list:
start, end = e['location']
if start <= 0:
start, end = -start, -end
list_to.extend(list_from[last:start + 1])
else:
list_to.extend(list_from[last:start])
docs = e['docs'].get_raw_docs()
list_docs = [l + '\n' for l in docs.splitlines()]
list_to.extend(list_docs)
last = end + 1
if last < len(list_from):
list_to.extend(list_from[last:])
return list_from, list_to | [
"def",
"compute_before_after",
"(",
"self",
")",
":",
"if",
"not",
"self",
".",
"parsed",
":",
"self",
".",
"_parse",
"(",
")",
"list_from",
"=",
"self",
".",
"input_lines",
"list_to",
"=",
"[",
"]",
"last",
"=",
"0",
"for",
"e",
"in",
"self",
".",
"docs_list",
":",
"start",
",",
"end",
"=",
"e",
"[",
"'location'",
"]",
"if",
"start",
"<=",
"0",
":",
"start",
",",
"end",
"=",
"-",
"start",
",",
"-",
"end",
"list_to",
".",
"extend",
"(",
"list_from",
"[",
"last",
":",
"start",
"+",
"1",
"]",
")",
"else",
":",
"list_to",
".",
"extend",
"(",
"list_from",
"[",
"last",
":",
"start",
"]",
")",
"docs",
"=",
"e",
"[",
"'docs'",
"]",
".",
"get_raw_docs",
"(",
")",
"list_docs",
"=",
"[",
"l",
"+",
"'\\n'",
"for",
"l",
"in",
"docs",
".",
"splitlines",
"(",
")",
"]",
"list_to",
".",
"extend",
"(",
"list_docs",
")",
"last",
"=",
"end",
"+",
"1",
"if",
"last",
"<",
"len",
"(",
"list_from",
")",
":",
"list_to",
".",
"extend",
"(",
"list_from",
"[",
"last",
":",
"]",
")",
"return",
"list_from",
",",
"list_to"
] | Compute the list of lines before and after the proposed docstring changes.
:return: tuple of before,after where each is a list of lines of python code. | [
"Compute",
"the",
"list",
"of",
"lines",
"before",
"and",
"after",
"the",
"proposed",
"docstring",
"changes",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pyment.py#L236-L260 | train |
dadadel/pyment | pyment/pyment.py | PyComment.diff | def diff(self, source_path='', target_path='', which=-1):
"""Build the diff between original docstring and proposed docstring.
:type which: int
-> -1 means all the dosctrings of the file
-> >=0 means the index of the docstring to proceed (Default value = -1)
:param source_path: (Default value = '')
:param target_path: (Default value = '')
:returns: the resulted diff
:rtype: List[str]
"""
list_from, list_to = self.compute_before_after()
if source_path.startswith(os.sep):
source_path = source_path[1:]
if source_path and not source_path.endswith(os.sep):
source_path += os.sep
if target_path.startswith(os.sep):
target_path = target_path[1:]
if target_path and not target_path.endswith(os.sep):
target_path += os.sep
fromfile = 'a/' + source_path + os.path.basename(self.input_file)
tofile = 'b/' + target_path + os.path.basename(self.input_file)
diff_list = difflib.unified_diff(list_from, list_to, fromfile, tofile)
return [d for d in diff_list] | python | def diff(self, source_path='', target_path='', which=-1):
"""Build the diff between original docstring and proposed docstring.
:type which: int
-> -1 means all the dosctrings of the file
-> >=0 means the index of the docstring to proceed (Default value = -1)
:param source_path: (Default value = '')
:param target_path: (Default value = '')
:returns: the resulted diff
:rtype: List[str]
"""
list_from, list_to = self.compute_before_after()
if source_path.startswith(os.sep):
source_path = source_path[1:]
if source_path and not source_path.endswith(os.sep):
source_path += os.sep
if target_path.startswith(os.sep):
target_path = target_path[1:]
if target_path and not target_path.endswith(os.sep):
target_path += os.sep
fromfile = 'a/' + source_path + os.path.basename(self.input_file)
tofile = 'b/' + target_path + os.path.basename(self.input_file)
diff_list = difflib.unified_diff(list_from, list_to, fromfile, tofile)
return [d for d in diff_list] | [
"def",
"diff",
"(",
"self",
",",
"source_path",
"=",
"''",
",",
"target_path",
"=",
"''",
",",
"which",
"=",
"-",
"1",
")",
":",
"list_from",
",",
"list_to",
"=",
"self",
".",
"compute_before_after",
"(",
")",
"if",
"source_path",
".",
"startswith",
"(",
"os",
".",
"sep",
")",
":",
"source_path",
"=",
"source_path",
"[",
"1",
":",
"]",
"if",
"source_path",
"and",
"not",
"source_path",
".",
"endswith",
"(",
"os",
".",
"sep",
")",
":",
"source_path",
"+=",
"os",
".",
"sep",
"if",
"target_path",
".",
"startswith",
"(",
"os",
".",
"sep",
")",
":",
"target_path",
"=",
"target_path",
"[",
"1",
":",
"]",
"if",
"target_path",
"and",
"not",
"target_path",
".",
"endswith",
"(",
"os",
".",
"sep",
")",
":",
"target_path",
"+=",
"os",
".",
"sep",
"fromfile",
"=",
"'a/'",
"+",
"source_path",
"+",
"os",
".",
"path",
".",
"basename",
"(",
"self",
".",
"input_file",
")",
"tofile",
"=",
"'b/'",
"+",
"target_path",
"+",
"os",
".",
"path",
".",
"basename",
"(",
"self",
".",
"input_file",
")",
"diff_list",
"=",
"difflib",
".",
"unified_diff",
"(",
"list_from",
",",
"list_to",
",",
"fromfile",
",",
"tofile",
")",
"return",
"[",
"d",
"for",
"d",
"in",
"diff_list",
"]"
] | Build the diff between original docstring and proposed docstring.
:type which: int
-> -1 means all the dosctrings of the file
-> >=0 means the index of the docstring to proceed (Default value = -1)
:param source_path: (Default value = '')
:param target_path: (Default value = '')
:returns: the resulted diff
:rtype: List[str] | [
"Build",
"the",
"diff",
"between",
"original",
"docstring",
"and",
"proposed",
"docstring",
"."
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pyment.py#L262-L287 | train |
dadadel/pyment | pyment/pyment.py | PyComment.get_patch_lines | def get_patch_lines(self, source_path, target_path):
"""Return the diff between source_path and target_path
:param source_path: name of the original file (Default value = '')
:param target_path: name of the final file (Default value = '')
:return: the diff as a list of \n terminated lines
:rtype: List[str]
"""
diff = self.diff(source_path, target_path)
return ["# Patch generated by Pyment v{0}\n\n".format(__version__)] + diff | python | def get_patch_lines(self, source_path, target_path):
"""Return the diff between source_path and target_path
:param source_path: name of the original file (Default value = '')
:param target_path: name of the final file (Default value = '')
:return: the diff as a list of \n terminated lines
:rtype: List[str]
"""
diff = self.diff(source_path, target_path)
return ["# Patch generated by Pyment v{0}\n\n".format(__version__)] + diff | [
"def",
"get_patch_lines",
"(",
"self",
",",
"source_path",
",",
"target_path",
")",
":",
"diff",
"=",
"self",
".",
"diff",
"(",
"source_path",
",",
"target_path",
")",
"return",
"[",
"\"# Patch generated by Pyment v{0}\\n\\n\"",
".",
"format",
"(",
"__version__",
")",
"]",
"+",
"diff"
] | Return the diff between source_path and target_path
:param source_path: name of the original file (Default value = '')
:param target_path: name of the final file (Default value = '')
:return: the diff as a list of \n terminated lines
:rtype: List[str] | [
"Return",
"the",
"diff",
"between",
"source_path",
"and",
"target_path"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pyment.py#L289-L300 | train |
dadadel/pyment | pyment/pyment.py | PyComment.write_patch_file | def write_patch_file(self, patch_file, lines_to_write):
"""Write lines_to_write to a the file called patch_file
:param patch_file: file name of the patch to generate
:param lines_to_write: lines to write to the file - they should be \n terminated
:type lines_to_write: list[str]
:return: None
"""
with open(patch_file, 'w') as f:
f.writelines(lines_to_write) | python | def write_patch_file(self, patch_file, lines_to_write):
"""Write lines_to_write to a the file called patch_file
:param patch_file: file name of the patch to generate
:param lines_to_write: lines to write to the file - they should be \n terminated
:type lines_to_write: list[str]
:return: None
"""
with open(patch_file, 'w') as f:
f.writelines(lines_to_write) | [
"def",
"write_patch_file",
"(",
"self",
",",
"patch_file",
",",
"lines_to_write",
")",
":",
"with",
"open",
"(",
"patch_file",
",",
"'w'",
")",
"as",
"f",
":",
"f",
".",
"writelines",
"(",
"lines_to_write",
")"
] | Write lines_to_write to a the file called patch_file
:param patch_file: file name of the patch to generate
:param lines_to_write: lines to write to the file - they should be \n terminated
:type lines_to_write: list[str]
:return: None | [
"Write",
"lines_to_write",
"to",
"a",
"the",
"file",
"called",
"patch_file"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pyment.py#L302-L312 | train |
dadadel/pyment | pyment/pyment.py | PyComment.overwrite_source_file | def overwrite_source_file(self, lines_to_write):
"""overwrite the file with line_to_write
:param lines_to_write: lines to write to the file - they should be \n terminated
:type lines_to_write: List[str]
:return: None
"""
tmp_filename = '{0}.writing'.format(self.input_file)
ok = False
try:
with open(tmp_filename, 'w') as fh:
fh.writelines(lines_to_write)
ok = True
finally:
if ok:
if platform.system() == 'Windows':
self._windows_rename(tmp_filename)
else:
os.rename(tmp_filename, self.input_file)
else:
os.unlink(tmp_filename) | python | def overwrite_source_file(self, lines_to_write):
"""overwrite the file with line_to_write
:param lines_to_write: lines to write to the file - they should be \n terminated
:type lines_to_write: List[str]
:return: None
"""
tmp_filename = '{0}.writing'.format(self.input_file)
ok = False
try:
with open(tmp_filename, 'w') as fh:
fh.writelines(lines_to_write)
ok = True
finally:
if ok:
if platform.system() == 'Windows':
self._windows_rename(tmp_filename)
else:
os.rename(tmp_filename, self.input_file)
else:
os.unlink(tmp_filename) | [
"def",
"overwrite_source_file",
"(",
"self",
",",
"lines_to_write",
")",
":",
"tmp_filename",
"=",
"'{0}.writing'",
".",
"format",
"(",
"self",
".",
"input_file",
")",
"ok",
"=",
"False",
"try",
":",
"with",
"open",
"(",
"tmp_filename",
",",
"'w'",
")",
"as",
"fh",
":",
"fh",
".",
"writelines",
"(",
"lines_to_write",
")",
"ok",
"=",
"True",
"finally",
":",
"if",
"ok",
":",
"if",
"platform",
".",
"system",
"(",
")",
"==",
"'Windows'",
":",
"self",
".",
"_windows_rename",
"(",
"tmp_filename",
")",
"else",
":",
"os",
".",
"rename",
"(",
"tmp_filename",
",",
"self",
".",
"input_file",
")",
"else",
":",
"os",
".",
"unlink",
"(",
"tmp_filename",
")"
] | overwrite the file with line_to_write
:param lines_to_write: lines to write to the file - they should be \n terminated
:type lines_to_write: List[str]
:return: None | [
"overwrite",
"the",
"file",
"with",
"line_to_write"
] | 3d1bdf87d083ff56230bd0bf7c5252e20552b7b6 | https://github.com/dadadel/pyment/blob/3d1bdf87d083ff56230bd0bf7c5252e20552b7b6/pyment/pyment.py#L314-L335 | train |
what-studio/profiling | profiling/sortkeys.py | by_own_time_per_call | def by_own_time_per_call(stat):
"""Sorting by exclusive elapsed time per call in descending order."""
return (-stat.own_time_per_call if stat.own_hits else -stat.own_time,
by_deep_time_per_call(stat)) | python | def by_own_time_per_call(stat):
"""Sorting by exclusive elapsed time per call in descending order."""
return (-stat.own_time_per_call if stat.own_hits else -stat.own_time,
by_deep_time_per_call(stat)) | [
"def",
"by_own_time_per_call",
"(",
"stat",
")",
":",
"return",
"(",
"-",
"stat",
".",
"own_time_per_call",
"if",
"stat",
".",
"own_hits",
"else",
"-",
"stat",
".",
"own_time",
",",
"by_deep_time_per_call",
"(",
"stat",
")",
")"
] | Sorting by exclusive elapsed time per call in descending order. | [
"Sorting",
"by",
"exclusive",
"elapsed",
"time",
"per",
"call",
"in",
"descending",
"order",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/sortkeys.py#L61-L64 | train |
what-studio/profiling | profiling/profiler.py | Profiler.result | def result(self):
"""Gets the frozen statistics to serialize by Pickle."""
try:
cpu_time = max(0, time.clock() - self._cpu_time_started)
wall_time = max(0, time.time() - self._wall_time_started)
except AttributeError:
cpu_time = wall_time = 0.0
return self.stats, cpu_time, wall_time | python | def result(self):
"""Gets the frozen statistics to serialize by Pickle."""
try:
cpu_time = max(0, time.clock() - self._cpu_time_started)
wall_time = max(0, time.time() - self._wall_time_started)
except AttributeError:
cpu_time = wall_time = 0.0
return self.stats, cpu_time, wall_time | [
"def",
"result",
"(",
"self",
")",
":",
"try",
":",
"cpu_time",
"=",
"max",
"(",
"0",
",",
"time",
".",
"clock",
"(",
")",
"-",
"self",
".",
"_cpu_time_started",
")",
"wall_time",
"=",
"max",
"(",
"0",
",",
"time",
".",
"time",
"(",
")",
"-",
"self",
".",
"_wall_time_started",
")",
"except",
"AttributeError",
":",
"cpu_time",
"=",
"wall_time",
"=",
"0.0",
"return",
"self",
".",
"stats",
",",
"cpu_time",
",",
"wall_time"
] | Gets the frozen statistics to serialize by Pickle. | [
"Gets",
"the",
"frozen",
"statistics",
"to",
"serialize",
"by",
"Pickle",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/profiler.py#L65-L72 | train |
what-studio/profiling | profiling/profiler.py | Profiler.dump | def dump(self, dump_filename, pickle_protocol=pickle.HIGHEST_PROTOCOL):
"""Saves the profiling result to a file
:param dump_filename: path to a file
:type dump_filename: str
:param pickle_protocol: version of pickle protocol
:type pickle_protocol: int
"""
result = self.result()
with open(dump_filename, 'wb') as f:
pickle.dump((self.__class__, result), f, pickle_protocol) | python | def dump(self, dump_filename, pickle_protocol=pickle.HIGHEST_PROTOCOL):
"""Saves the profiling result to a file
:param dump_filename: path to a file
:type dump_filename: str
:param pickle_protocol: version of pickle protocol
:type pickle_protocol: int
"""
result = self.result()
with open(dump_filename, 'wb') as f:
pickle.dump((self.__class__, result), f, pickle_protocol) | [
"def",
"dump",
"(",
"self",
",",
"dump_filename",
",",
"pickle_protocol",
"=",
"pickle",
".",
"HIGHEST_PROTOCOL",
")",
":",
"result",
"=",
"self",
".",
"result",
"(",
")",
"with",
"open",
"(",
"dump_filename",
",",
"'wb'",
")",
"as",
"f",
":",
"pickle",
".",
"dump",
"(",
"(",
"self",
".",
"__class__",
",",
"result",
")",
",",
"f",
",",
"pickle_protocol",
")"
] | Saves the profiling result to a file
:param dump_filename: path to a file
:type dump_filename: str
:param pickle_protocol: version of pickle protocol
:type pickle_protocol: int | [
"Saves",
"the",
"profiling",
"result",
"to",
"a",
"file"
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/profiler.py#L74-L86 | train |
what-studio/profiling | profiling/profiler.py | Profiler.make_viewer | def make_viewer(self, title=None, at=None):
"""Makes a statistics viewer from the profiling result.
"""
viewer = StatisticsViewer()
viewer.set_profiler_class(self.__class__)
stats, cpu_time, wall_time = self.result()
viewer.set_result(stats, cpu_time, wall_time, title=title, at=at)
viewer.activate()
return viewer | python | def make_viewer(self, title=None, at=None):
"""Makes a statistics viewer from the profiling result.
"""
viewer = StatisticsViewer()
viewer.set_profiler_class(self.__class__)
stats, cpu_time, wall_time = self.result()
viewer.set_result(stats, cpu_time, wall_time, title=title, at=at)
viewer.activate()
return viewer | [
"def",
"make_viewer",
"(",
"self",
",",
"title",
"=",
"None",
",",
"at",
"=",
"None",
")",
":",
"viewer",
"=",
"StatisticsViewer",
"(",
")",
"viewer",
".",
"set_profiler_class",
"(",
"self",
".",
"__class__",
")",
"stats",
",",
"cpu_time",
",",
"wall_time",
"=",
"self",
".",
"result",
"(",
")",
"viewer",
".",
"set_result",
"(",
"stats",
",",
"cpu_time",
",",
"wall_time",
",",
"title",
"=",
"title",
",",
"at",
"=",
"at",
")",
"viewer",
".",
"activate",
"(",
")",
"return",
"viewer"
] | Makes a statistics viewer from the profiling result. | [
"Makes",
"a",
"statistics",
"viewer",
"from",
"the",
"profiling",
"result",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/profiler.py#L88-L96 | train |
what-studio/profiling | profiling/remote/__init__.py | pack_msg | def pack_msg(method, msg, pickle_protocol=PICKLE_PROTOCOL):
"""Packs a method and message."""
dump = io.BytesIO()
pickle.dump(msg, dump, pickle_protocol)
size = dump.tell()
return (struct.pack(METHOD_STRUCT_FORMAT, method) +
struct.pack(SIZE_STRUCT_FORMAT, size) + dump.getvalue()) | python | def pack_msg(method, msg, pickle_protocol=PICKLE_PROTOCOL):
"""Packs a method and message."""
dump = io.BytesIO()
pickle.dump(msg, dump, pickle_protocol)
size = dump.tell()
return (struct.pack(METHOD_STRUCT_FORMAT, method) +
struct.pack(SIZE_STRUCT_FORMAT, size) + dump.getvalue()) | [
"def",
"pack_msg",
"(",
"method",
",",
"msg",
",",
"pickle_protocol",
"=",
"PICKLE_PROTOCOL",
")",
":",
"dump",
"=",
"io",
".",
"BytesIO",
"(",
")",
"pickle",
".",
"dump",
"(",
"msg",
",",
"dump",
",",
"pickle_protocol",
")",
"size",
"=",
"dump",
".",
"tell",
"(",
")",
"return",
"(",
"struct",
".",
"pack",
"(",
"METHOD_STRUCT_FORMAT",
",",
"method",
")",
"+",
"struct",
".",
"pack",
"(",
"SIZE_STRUCT_FORMAT",
",",
"size",
")",
"+",
"dump",
".",
"getvalue",
"(",
")",
")"
] | Packs a method and message. | [
"Packs",
"a",
"method",
"and",
"message",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/remote/__init__.py#L60-L66 | train |
what-studio/profiling | profiling/remote/__init__.py | recv | def recv(sock, size):
"""Receives exactly `size` bytes. This function blocks the thread."""
data = sock.recv(size, socket.MSG_WAITALL)
if len(data) < size:
raise socket.error(ECONNRESET, 'Connection closed')
return data | python | def recv(sock, size):
"""Receives exactly `size` bytes. This function blocks the thread."""
data = sock.recv(size, socket.MSG_WAITALL)
if len(data) < size:
raise socket.error(ECONNRESET, 'Connection closed')
return data | [
"def",
"recv",
"(",
"sock",
",",
"size",
")",
":",
"data",
"=",
"sock",
".",
"recv",
"(",
"size",
",",
"socket",
".",
"MSG_WAITALL",
")",
"if",
"len",
"(",
"data",
")",
"<",
"size",
":",
"raise",
"socket",
".",
"error",
"(",
"ECONNRESET",
",",
"'Connection closed'",
")",
"return",
"data"
] | Receives exactly `size` bytes. This function blocks the thread. | [
"Receives",
"exactly",
"size",
"bytes",
".",
"This",
"function",
"blocks",
"the",
"thread",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/remote/__init__.py#L69-L74 | train |
what-studio/profiling | profiling/remote/__init__.py | recv_msg | def recv_msg(sock):
"""Receives a method and message from the socket. This function blocks the
current thread.
"""
data = recv(sock, struct.calcsize(METHOD_STRUCT_FORMAT))
method, = struct.unpack(METHOD_STRUCT_FORMAT, data)
data = recv(sock, struct.calcsize(SIZE_STRUCT_FORMAT))
size, = struct.unpack(SIZE_STRUCT_FORMAT, data)
data = recv(sock, size)
msg = pickle.loads(data)
return method, msg | python | def recv_msg(sock):
"""Receives a method and message from the socket. This function blocks the
current thread.
"""
data = recv(sock, struct.calcsize(METHOD_STRUCT_FORMAT))
method, = struct.unpack(METHOD_STRUCT_FORMAT, data)
data = recv(sock, struct.calcsize(SIZE_STRUCT_FORMAT))
size, = struct.unpack(SIZE_STRUCT_FORMAT, data)
data = recv(sock, size)
msg = pickle.loads(data)
return method, msg | [
"def",
"recv_msg",
"(",
"sock",
")",
":",
"data",
"=",
"recv",
"(",
"sock",
",",
"struct",
".",
"calcsize",
"(",
"METHOD_STRUCT_FORMAT",
")",
")",
"method",
",",
"=",
"struct",
".",
"unpack",
"(",
"METHOD_STRUCT_FORMAT",
",",
"data",
")",
"data",
"=",
"recv",
"(",
"sock",
",",
"struct",
".",
"calcsize",
"(",
"SIZE_STRUCT_FORMAT",
")",
")",
"size",
",",
"=",
"struct",
".",
"unpack",
"(",
"SIZE_STRUCT_FORMAT",
",",
"data",
")",
"data",
"=",
"recv",
"(",
"sock",
",",
"size",
")",
"msg",
"=",
"pickle",
".",
"loads",
"(",
"data",
")",
"return",
"method",
",",
"msg"
] | Receives a method and message from the socket. This function blocks the
current thread. | [
"Receives",
"a",
"method",
"and",
"message",
"from",
"the",
"socket",
".",
"This",
"function",
"blocks",
"the",
"current",
"thread",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/remote/__init__.py#L77-L87 | train |
what-studio/profiling | profiling/remote/__init__.py | ProfilingServer.connected | def connected(self, client):
"""Call this method when a client connected."""
self.clients.add(client)
self._log_connected(client)
self._start_watching(client)
self.send_msg(client, WELCOME, (self.pickle_protocol, __version__),
pickle_protocol=0)
profiler = self.profiler
while True:
try:
profiler = profiler.profiler
except AttributeError:
break
self.send_msg(client, PROFILER, type(profiler))
if self._latest_result_data is not None:
try:
self._send(client, self._latest_result_data)
except socket.error as exc:
if exc.errno in (EBADF, EPIPE):
self.disconnected(client)
return
raise
if len(self.clients) == 1:
self._start_profiling() | python | def connected(self, client):
"""Call this method when a client connected."""
self.clients.add(client)
self._log_connected(client)
self._start_watching(client)
self.send_msg(client, WELCOME, (self.pickle_protocol, __version__),
pickle_protocol=0)
profiler = self.profiler
while True:
try:
profiler = profiler.profiler
except AttributeError:
break
self.send_msg(client, PROFILER, type(profiler))
if self._latest_result_data is not None:
try:
self._send(client, self._latest_result_data)
except socket.error as exc:
if exc.errno in (EBADF, EPIPE):
self.disconnected(client)
return
raise
if len(self.clients) == 1:
self._start_profiling() | [
"def",
"connected",
"(",
"self",
",",
"client",
")",
":",
"self",
".",
"clients",
".",
"add",
"(",
"client",
")",
"self",
".",
"_log_connected",
"(",
"client",
")",
"self",
".",
"_start_watching",
"(",
"client",
")",
"self",
".",
"send_msg",
"(",
"client",
",",
"WELCOME",
",",
"(",
"self",
".",
"pickle_protocol",
",",
"__version__",
")",
",",
"pickle_protocol",
"=",
"0",
")",
"profiler",
"=",
"self",
".",
"profiler",
"while",
"True",
":",
"try",
":",
"profiler",
"=",
"profiler",
".",
"profiler",
"except",
"AttributeError",
":",
"break",
"self",
".",
"send_msg",
"(",
"client",
",",
"PROFILER",
",",
"type",
"(",
"profiler",
")",
")",
"if",
"self",
".",
"_latest_result_data",
"is",
"not",
"None",
":",
"try",
":",
"self",
".",
"_send",
"(",
"client",
",",
"self",
".",
"_latest_result_data",
")",
"except",
"socket",
".",
"error",
"as",
"exc",
":",
"if",
"exc",
".",
"errno",
"in",
"(",
"EBADF",
",",
"EPIPE",
")",
":",
"self",
".",
"disconnected",
"(",
"client",
")",
"return",
"raise",
"if",
"len",
"(",
"self",
".",
"clients",
")",
"==",
"1",
":",
"self",
".",
"_start_profiling",
"(",
")"
] | Call this method when a client connected. | [
"Call",
"this",
"method",
"when",
"a",
"client",
"connected",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/remote/__init__.py#L205-L228 | train |
what-studio/profiling | profiling/remote/__init__.py | ProfilingServer.disconnected | def disconnected(self, client):
"""Call this method when a client disconnected."""
if client not in self.clients:
# already disconnected.
return
self.clients.remove(client)
self._log_disconnected(client)
self._close(client) | python | def disconnected(self, client):
"""Call this method when a client disconnected."""
if client not in self.clients:
# already disconnected.
return
self.clients.remove(client)
self._log_disconnected(client)
self._close(client) | [
"def",
"disconnected",
"(",
"self",
",",
"client",
")",
":",
"if",
"client",
"not",
"in",
"self",
".",
"clients",
":",
"# already disconnected.",
"return",
"self",
".",
"clients",
".",
"remove",
"(",
"client",
")",
"self",
".",
"_log_disconnected",
"(",
"client",
")",
"self",
".",
"_close",
"(",
"client",
")"
] | Call this method when a client disconnected. | [
"Call",
"this",
"method",
"when",
"a",
"client",
"disconnected",
"."
] | 49666ba3ea295eb73782ae6c18a4ec7929d7d8b7 | https://github.com/what-studio/profiling/blob/49666ba3ea295eb73782ae6c18a4ec7929d7d8b7/profiling/remote/__init__.py#L230-L237 | train |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.