"],
- inputs=style_prompt, label="style examples")
-
- neg_prompt = gr.Textbox(placeholder="negative prompts", value=neg_prompt, label="negative prompt")
-
- ins = gr.Slider(1, 60, 30, label="inference steps")
- gs = gr.Slider(1, 10, 2.5, step=1, label="guidance scale")
-
- seed = gr.Slider(0, 10, 2, step=1, label="seed")
- btn1 = gr.Button("실행")
- btn1.click(predict, [ prompt, style_prompt, neg_prompt, ins, gs, seed], out_image)
-
-if __name__ == "__main__":
- demo.launch()
\ No newline at end of file
diff --git a/spaces/candlend/vits-hoshimi/sovits/vdecoder/parallel_wavegan/__init__.py b/spaces/candlend/vits-hoshimi/sovits/vdecoder/parallel_wavegan/__init__.py
deleted file mode 100644
index e69de29bb2d1d6434b8b29ae775ad8c2e48c5391..0000000000000000000000000000000000000000
diff --git a/spaces/candlend/vits-hoshimi/vits/text/mandarin.py b/spaces/candlend/vits-hoshimi/vits/text/mandarin.py
deleted file mode 100644
index a9ce0c4b223cd7fbb00e8332d2dd53de4c7cea09..0000000000000000000000000000000000000000
--- a/spaces/candlend/vits-hoshimi/vits/text/mandarin.py
+++ /dev/null
@@ -1,328 +0,0 @@
-import os
-import sys
-import re
-from pypinyin import lazy_pinyin, BOPOMOFO
-import jieba
-import cn2an
-
-
-# List of (Latin alphabet, bopomofo) pairs:
-_latin_to_bopomofo = [(re.compile('%s' % x[0], re.IGNORECASE), x[1]) for x in [
- ('a', 'ㄟˉ'),
- ('b', 'ㄅㄧˋ'),
- ('c', 'ㄙㄧˉ'),
- ('d', 'ㄉㄧˋ'),
- ('e', 'ㄧˋ'),
- ('f', 'ㄝˊㄈㄨˋ'),
- ('g', 'ㄐㄧˋ'),
- ('h', 'ㄝˇㄑㄩˋ'),
- ('i', 'ㄞˋ'),
- ('j', 'ㄐㄟˋ'),
- ('k', 'ㄎㄟˋ'),
- ('l', 'ㄝˊㄛˋ'),
- ('m', 'ㄝˊㄇㄨˋ'),
- ('n', 'ㄣˉ'),
- ('o', 'ㄡˉ'),
- ('p', 'ㄆㄧˉ'),
- ('q', 'ㄎㄧㄡˉ'),
- ('r', 'ㄚˋ'),
- ('s', 'ㄝˊㄙˋ'),
- ('t', 'ㄊㄧˋ'),
- ('u', 'ㄧㄡˉ'),
- ('v', 'ㄨㄧˉ'),
- ('w', 'ㄉㄚˋㄅㄨˋㄌㄧㄡˋ'),
- ('x', 'ㄝˉㄎㄨˋㄙˋ'),
- ('y', 'ㄨㄞˋ'),
- ('z', 'ㄗㄟˋ')
-]]
-
-# List of (bopomofo, romaji) pairs:
-_bopomofo_to_romaji = [(re.compile('%s' % x[0]), x[1]) for x in [
- ('ㄅㄛ', 'p⁼wo'),
- ('ㄆㄛ', 'pʰwo'),
- ('ㄇㄛ', 'mwo'),
- ('ㄈㄛ', 'fwo'),
- ('ㄅ', 'p⁼'),
- ('ㄆ', 'pʰ'),
- ('ㄇ', 'm'),
- ('ㄈ', 'f'),
- ('ㄉ', 't⁼'),
- ('ㄊ', 'tʰ'),
- ('ㄋ', 'n'),
- ('ㄌ', 'l'),
- ('ㄍ', 'k⁼'),
- ('ㄎ', 'kʰ'),
- ('ㄏ', 'h'),
- ('ㄐ', 'ʧ⁼'),
- ('ㄑ', 'ʧʰ'),
- ('ㄒ', 'ʃ'),
- ('ㄓ', 'ʦ`⁼'),
- ('ㄔ', 'ʦ`ʰ'),
- ('ㄕ', 's`'),
- ('ㄖ', 'ɹ`'),
- ('ㄗ', 'ʦ⁼'),
- ('ㄘ', 'ʦʰ'),
- ('ㄙ', 's'),
- ('ㄚ', 'a'),
- ('ㄛ', 'o'),
- ('ㄜ', 'ə'),
- ('ㄝ', 'e'),
- ('ㄞ', 'ai'),
- ('ㄟ', 'ei'),
- ('ㄠ', 'au'),
- ('ㄡ', 'ou'),
- ('ㄧㄢ', 'yeNN'),
- ('ㄢ', 'aNN'),
- ('ㄧㄣ', 'iNN'),
- ('ㄣ', 'əNN'),
- ('ㄤ', 'aNg'),
- ('ㄧㄥ', 'iNg'),
- ('ㄨㄥ', 'uNg'),
- ('ㄩㄥ', 'yuNg'),
- ('ㄥ', 'əNg'),
- ('ㄦ', 'əɻ'),
- ('ㄧ', 'i'),
- ('ㄨ', 'u'),
- ('ㄩ', 'ɥ'),
- ('ˉ', '→'),
- ('ˊ', '↑'),
- ('ˇ', '↓↑'),
- ('ˋ', '↓'),
- ('˙', ''),
- (',', ','),
- ('。', '.'),
- ('!', '!'),
- ('?', '?'),
- ('—', '-')
-]]
-
-# List of (romaji, ipa) pairs:
-_romaji_to_ipa = [(re.compile('%s' % x[0], re.IGNORECASE), x[1]) for x in [
- ('ʃy', 'ʃ'),
- ('ʧʰy', 'ʧʰ'),
- ('ʧ⁼y', 'ʧ⁼'),
- ('NN', 'n'),
- ('Ng', 'ŋ'),
- ('y', 'j'),
- ('h', 'x')
-]]
-
-# List of (bopomofo, ipa) pairs:
-_bopomofo_to_ipa = [(re.compile('%s' % x[0]), x[1]) for x in [
- ('ㄅㄛ', 'p⁼wo'),
- ('ㄆㄛ', 'pʰwo'),
- ('ㄇㄛ', 'mwo'),
- ('ㄈㄛ', 'fwo'),
- ('ㄅ', 'p⁼'),
- ('ㄆ', 'pʰ'),
- ('ㄇ', 'm'),
- ('ㄈ', 'f'),
- ('ㄉ', 't⁼'),
- ('ㄊ', 'tʰ'),
- ('ㄋ', 'n'),
- ('ㄌ', 'l'),
- ('ㄍ', 'k⁼'),
- ('ㄎ', 'kʰ'),
- ('ㄏ', 'x'),
- ('ㄐ', 'tʃ⁼'),
- ('ㄑ', 'tʃʰ'),
- ('ㄒ', 'ʃ'),
- ('ㄓ', 'ts`⁼'),
- ('ㄔ', 'ts`ʰ'),
- ('ㄕ', 's`'),
- ('ㄖ', 'ɹ`'),
- ('ㄗ', 'ts⁼'),
- ('ㄘ', 'tsʰ'),
- ('ㄙ', 's'),
- ('ㄚ', 'a'),
- ('ㄛ', 'o'),
- ('ㄜ', 'ə'),
- ('ㄝ', 'ɛ'),
- ('ㄞ', 'aɪ'),
- ('ㄟ', 'eɪ'),
- ('ㄠ', 'ɑʊ'),
- ('ㄡ', 'oʊ'),
- ('ㄧㄢ', 'jɛn'),
- ('ㄩㄢ', 'ɥæn'),
- ('ㄢ', 'an'),
- ('ㄧㄣ', 'in'),
- ('ㄩㄣ', 'ɥn'),
- ('ㄣ', 'ən'),
- ('ㄤ', 'ɑŋ'),
- ('ㄧㄥ', 'iŋ'),
- ('ㄨㄥ', 'ʊŋ'),
- ('ㄩㄥ', 'jʊŋ'),
- ('ㄥ', 'əŋ'),
- ('ㄦ', 'əɻ'),
- ('ㄧ', 'i'),
- ('ㄨ', 'u'),
- ('ㄩ', 'ɥ'),
- ('ˉ', '→'),
- ('ˊ', '↑'),
- ('ˇ', '↓↑'),
- ('ˋ', '↓'),
- ('˙', ''),
- (',', ','),
- ('。', '.'),
- ('!', '!'),
- ('?', '?'),
- ('—', '-')
-]]
-
-# List of (bopomofo, ipa2) pairs:
-_bopomofo_to_ipa2 = [(re.compile('%s' % x[0]), x[1]) for x in [
- ('ㄅㄛ', 'pwo'),
- ('ㄆㄛ', 'pʰwo'),
- ('ㄇㄛ', 'mwo'),
- ('ㄈㄛ', 'fwo'),
- ('ㄅ', 'p'),
- ('ㄆ', 'pʰ'),
- ('ㄇ', 'm'),
- ('ㄈ', 'f'),
- ('ㄉ', 't'),
- ('ㄊ', 'tʰ'),
- ('ㄋ', 'n'),
- ('ㄌ', 'l'),
- ('ㄍ', 'k'),
- ('ㄎ', 'kʰ'),
- ('ㄏ', 'h'),
- ('ㄐ', 'tɕ'),
- ('ㄑ', 'tɕʰ'),
- ('ㄒ', 'ɕ'),
- ('ㄓ', 'tʂ'),
- ('ㄔ', 'tʂʰ'),
- ('ㄕ', 'ʂ'),
- ('ㄖ', 'ɻ'),
- ('ㄗ', 'ts'),
- ('ㄘ', 'tsʰ'),
- ('ㄙ', 's'),
- ('ㄚ', 'a'),
- ('ㄛ', 'o'),
- ('ㄜ', 'ɤ'),
- ('ㄝ', 'ɛ'),
- ('ㄞ', 'aɪ'),
- ('ㄟ', 'eɪ'),
- ('ㄠ', 'ɑʊ'),
- ('ㄡ', 'oʊ'),
- ('ㄧㄢ', 'jɛn'),
- ('ㄩㄢ', 'yæn'),
- ('ㄢ', 'an'),
- ('ㄧㄣ', 'in'),
- ('ㄩㄣ', 'yn'),
- ('ㄣ', 'ən'),
- ('ㄤ', 'ɑŋ'),
- ('ㄧㄥ', 'iŋ'),
- ('ㄨㄥ', 'ʊŋ'),
- ('ㄩㄥ', 'jʊŋ'),
- ('ㄥ', 'ɤŋ'),
- ('ㄦ', 'əɻ'),
- ('ㄧ', 'i'),
- ('ㄨ', 'u'),
- ('ㄩ', 'y'),
- ('ˉ', '˥'),
- ('ˊ', '˧˥'),
- ('ˇ', '˨˩˦'),
- ('ˋ', '˥˩'),
- ('˙', ''),
- (',', ','),
- ('。', '.'),
- ('!', '!'),
- ('?', '?'),
- ('—', '-')
-]]
-
-
-def number_to_chinese(text):
- numbers = re.findall(r'\d+(?:\.?\d+)?', text)
- for number in numbers:
- text = text.replace(number, cn2an.an2cn(number), 1)
- return text
-
-
-def chinese_to_bopomofo(text, taiwanese=False):
- text = text.replace('、', ',').replace(';', ',').replace(':', ',')
- words = jieba.lcut(text, cut_all=False)
- text = ''
- for word in words:
- bopomofos = lazy_pinyin(word, BOPOMOFO)
- if not re.search('[\u4e00-\u9fff]', word):
- text += word
- continue
- for i in range(len(bopomofos)):
- bopomofos[i] = re.sub(r'([\u3105-\u3129])$', r'\1ˉ', bopomofos[i])
- if text != '':
- text += ' '
- if taiwanese:
- text += '#'+'#'.join(bopomofos)
- else:
- text += ''.join(bopomofos)
- return text
-
-
-def latin_to_bopomofo(text):
- for regex, replacement in _latin_to_bopomofo:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def bopomofo_to_romaji(text):
- for regex, replacement in _bopomofo_to_romaji:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def bopomofo_to_ipa(text):
- for regex, replacement in _bopomofo_to_ipa:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def bopomofo_to_ipa2(text):
- for regex, replacement in _bopomofo_to_ipa2:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def chinese_to_romaji(text):
- text = number_to_chinese(text)
- text = chinese_to_bopomofo(text)
- text = latin_to_bopomofo(text)
- text = bopomofo_to_romaji(text)
- text = re.sub('i([aoe])', r'y\1', text)
- text = re.sub('u([aoəe])', r'w\1', text)
- text = re.sub('([ʦsɹ]`[⁼ʰ]?)([→↓↑ ]+|$)',
- r'\1ɹ`\2', text).replace('ɻ', 'ɹ`')
- text = re.sub('([ʦs][⁼ʰ]?)([→↓↑ ]+|$)', r'\1ɹ\2', text)
- return text
-
-
-def chinese_to_lazy_ipa(text):
- text = chinese_to_romaji(text)
- for regex, replacement in _romaji_to_ipa:
- text = re.sub(regex, replacement, text)
- return text
-
-
-def chinese_to_ipa(text):
- text = number_to_chinese(text)
- text = chinese_to_bopomofo(text)
- text = latin_to_bopomofo(text)
- text = bopomofo_to_ipa(text)
- text = re.sub('i([aoe])', r'j\1', text)
- text = re.sub('u([aoəe])', r'w\1', text)
- text = re.sub('([sɹ]`[⁼ʰ]?)([→↓↑ ]+|$)',
- r'\1ɹ`\2', text).replace('ɻ', 'ɹ`')
- text = re.sub('([s][⁼ʰ]?)([→↓↑ ]+|$)', r'\1ɹ\2', text)
- return text
-
-
-def chinese_to_ipa2(text, taiwanese=False):
- text = number_to_chinese(text)
- text = chinese_to_bopomofo(text, taiwanese)
- text = latin_to_bopomofo(text)
- text = bopomofo_to_ipa2(text)
- text = re.sub(r'i([aoe])', r'j\1', text)
- text = re.sub(r'u([aoəe])', r'w\1', text)
- text = re.sub(r'([ʂɹ]ʰ?)([˩˨˧˦˥ ]+|$)', r'\1ʅ\2', text)
- text = re.sub(r'(sʰ?)([˩˨˧˦˥ ]+|$)', r'\1ɿ\2', text)
- return text
diff --git a/spaces/chendl/compositional_test/transformers/docs/source/de/_config.py b/spaces/chendl/compositional_test/transformers/docs/source/de/_config.py
deleted file mode 100644
index a6d75853f572193e4c04bb931d9254c23fbd838b..0000000000000000000000000000000000000000
--- a/spaces/chendl/compositional_test/transformers/docs/source/de/_config.py
+++ /dev/null
@@ -1,14 +0,0 @@
-# docstyle-ignore
-INSTALL_CONTENT = """
-# Transformers installation
-! pip install transformers datasets
-# To install from source instead of the last release, comment the command above and uncomment the following one.
-# ! pip install git+https://github.com/huggingface/transformers.git
-"""
-
-notebook_first_cells = [{"type": "code", "content": INSTALL_CONTENT}]
-black_avoid_patterns = {
- "{processor_class}": "FakeProcessorClass",
- "{model_class}": "FakeModelClass",
- "{object_class}": "FakeObjectClass",
-}
diff --git a/spaces/chendl/compositional_test/transformers/examples/pytorch/image-classification/run_image_classification_no_trainer.py b/spaces/chendl/compositional_test/transformers/examples/pytorch/image-classification/run_image_classification_no_trainer.py
deleted file mode 100644
index bf1646c33c4ffd6ddf967207d4bec7903641a5c6..0000000000000000000000000000000000000000
--- a/spaces/chendl/compositional_test/transformers/examples/pytorch/image-classification/run_image_classification_no_trainer.py
+++ /dev/null
@@ -1,575 +0,0 @@
-# coding=utf-8
-# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-""" Finetuning any 🤗 Transformers model for image classification leveraging 🤗 Accelerate."""
-import argparse
-import json
-import logging
-import math
-import os
-from pathlib import Path
-
-import datasets
-import evaluate
-import torch
-from accelerate import Accelerator
-from accelerate.logging import get_logger
-from accelerate.utils import set_seed
-from datasets import load_dataset
-from huggingface_hub import Repository, create_repo
-from torch.utils.data import DataLoader
-from torchvision.transforms import (
- CenterCrop,
- Compose,
- Normalize,
- RandomHorizontalFlip,
- RandomResizedCrop,
- Resize,
- ToTensor,
-)
-from tqdm.auto import tqdm
-
-import transformers
-from transformers import AutoConfig, AutoImageProcessor, AutoModelForImageClassification, SchedulerType, get_scheduler
-from transformers.utils import check_min_version, get_full_repo_name, send_example_telemetry
-from transformers.utils.versions import require_version
-
-
-# Will error if the minimal version of Transformers is not installed. Remove at your own risks.
-check_min_version("4.28.0")
-
-logger = get_logger(__name__)
-
-require_version("datasets>=2.0.0", "To fix: pip install -r examples/pytorch/image-classification/requirements.txt")
-
-
-def parse_args():
- parser = argparse.ArgumentParser(description="Fine-tune a Transformers model on an image classification dataset")
- parser.add_argument(
- "--dataset_name",
- type=str,
- default="cifar10",
- help=(
- "The name of the Dataset (from the HuggingFace hub) to train on (could be your own, possibly private,"
- " dataset)."
- ),
- )
- parser.add_argument("--train_dir", type=str, default=None, help="A folder containing the training data.")
- parser.add_argument("--validation_dir", type=str, default=None, help="A folder containing the validation data.")
- parser.add_argument(
- "--max_train_samples",
- type=int,
- default=None,
- help=(
- "For debugging purposes or quicker training, truncate the number of training examples to this "
- "value if set."
- ),
- )
- parser.add_argument(
- "--max_eval_samples",
- type=int,
- default=None,
- help=(
- "For debugging purposes or quicker training, truncate the number of evaluation examples to this "
- "value if set."
- ),
- )
- parser.add_argument(
- "--train_val_split",
- type=float,
- default=0.15,
- help="Percent to split off of train for validation",
- )
- parser.add_argument(
- "--model_name_or_path",
- type=str,
- help="Path to pretrained model or model identifier from huggingface.co/models.",
- default="google/vit-base-patch16-224-in21k",
- )
- parser.add_argument(
- "--per_device_train_batch_size",
- type=int,
- default=8,
- help="Batch size (per device) for the training dataloader.",
- )
- parser.add_argument(
- "--per_device_eval_batch_size",
- type=int,
- default=8,
- help="Batch size (per device) for the evaluation dataloader.",
- )
- parser.add_argument(
- "--learning_rate",
- type=float,
- default=5e-5,
- help="Initial learning rate (after the potential warmup period) to use.",
- )
- parser.add_argument("--weight_decay", type=float, default=0.0, help="Weight decay to use.")
- parser.add_argument("--num_train_epochs", type=int, default=3, help="Total number of training epochs to perform.")
- parser.add_argument(
- "--max_train_steps",
- type=int,
- default=None,
- help="Total number of training steps to perform. If provided, overrides num_train_epochs.",
- )
- parser.add_argument(
- "--gradient_accumulation_steps",
- type=int,
- default=1,
- help="Number of updates steps to accumulate before performing a backward/update pass.",
- )
- parser.add_argument(
- "--lr_scheduler_type",
- type=SchedulerType,
- default="linear",
- help="The scheduler type to use.",
- choices=["linear", "cosine", "cosine_with_restarts", "polynomial", "constant", "constant_with_warmup"],
- )
- parser.add_argument(
- "--num_warmup_steps", type=int, default=0, help="Number of steps for the warmup in the lr scheduler."
- )
- parser.add_argument("--output_dir", type=str, default=None, help="Where to store the final model.")
- parser.add_argument("--seed", type=int, default=None, help="A seed for reproducible training.")
- parser.add_argument("--push_to_hub", action="store_true", help="Whether or not to push the model to the Hub.")
- parser.add_argument(
- "--hub_model_id", type=str, help="The name of the repository to keep in sync with the local `output_dir`."
- )
- parser.add_argument("--hub_token", type=str, help="The token to use to push to the Model Hub.")
- parser.add_argument(
- "--checkpointing_steps",
- type=str,
- default=None,
- help="Whether the various states should be saved at the end of every n steps, or 'epoch' for each epoch.",
- )
- parser.add_argument(
- "--resume_from_checkpoint",
- type=str,
- default=None,
- help="If the training should continue from a checkpoint folder.",
- )
- parser.add_argument(
- "--with_tracking",
- action="store_true",
- help="Whether to enable experiment trackers for logging.",
- )
- parser.add_argument(
- "--report_to",
- type=str,
- default="all",
- help=(
- 'The integration to report the results and logs to. Supported platforms are `"tensorboard"`,'
- ' `"wandb"`, `"comet_ml"` and `"clearml"`. Use `"all"` (default) to report to all integrations.'
- "Only applicable when `--with_tracking` is passed."
- ),
- )
- parser.add_argument(
- "--ignore_mismatched_sizes",
- action="store_true",
- help="Whether or not to enable to load a pretrained model whose head dimensions are different.",
- )
- args = parser.parse_args()
-
- # Sanity checks
- if args.dataset_name is None and args.train_dir is None and args.validation_dir is None:
- raise ValueError("Need either a dataset name or a training/validation folder.")
-
- if args.push_to_hub or args.with_tracking:
- if args.output_dir is None:
- raise ValueError(
- "Need an `output_dir` to create a repo when `--push_to_hub` or `with_tracking` is specified."
- )
-
- if args.output_dir is not None:
- os.makedirs(args.output_dir, exist_ok=True)
-
- return args
-
-
-def main():
- args = parse_args()
-
- # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The
- # information sent is the one passed as arguments along with your Python/PyTorch versions.
- send_example_telemetry("run_image_classification_no_trainer", args)
-
- # Initialize the accelerator. We will let the accelerator handle device placement for us in this example.
- # If we're using tracking, we also need to initialize it here and it will by default pick up all supported trackers
- # in the environment
- accelerator_log_kwargs = {}
-
- if args.with_tracking:
- accelerator_log_kwargs["log_with"] = args.report_to
- accelerator_log_kwargs["logging_dir"] = args.output_dir
-
- accelerator = Accelerator(gradient_accumulation_steps=args.gradient_accumulation_steps, **accelerator_log_kwargs)
-
- logger.info(accelerator.state)
- # Make one log on every process with the configuration for debugging.
- logging.basicConfig(
- format="%(asctime)s - %(levelname)s - %(name)s - %(message)s",
- datefmt="%m/%d/%Y %H:%M:%S",
- level=logging.INFO,
- )
- logger.info(accelerator.state, main_process_only=False)
- if accelerator.is_local_main_process:
- datasets.utils.logging.set_verbosity_warning()
- transformers.utils.logging.set_verbosity_info()
- else:
- datasets.utils.logging.set_verbosity_error()
- transformers.utils.logging.set_verbosity_error()
-
- # If passed along, set the training seed now.
- if args.seed is not None:
- set_seed(args.seed)
-
- # Handle the repository creation
- if accelerator.is_main_process:
- if args.push_to_hub:
- if args.hub_model_id is None:
- repo_name = get_full_repo_name(Path(args.output_dir).name, token=args.hub_token)
- else:
- repo_name = args.hub_model_id
- create_repo(repo_name, exist_ok=True, token=args.hub_token)
- repo = Repository(args.output_dir, clone_from=repo_name, token=args.hub_token)
-
- with open(os.path.join(args.output_dir, ".gitignore"), "w+") as gitignore:
- if "step_*" not in gitignore:
- gitignore.write("step_*\n")
- if "epoch_*" not in gitignore:
- gitignore.write("epoch_*\n")
- elif args.output_dir is not None:
- os.makedirs(args.output_dir, exist_ok=True)
- accelerator.wait_for_everyone()
-
- # Get the datasets: you can either provide your own training and evaluation files (see below)
- # or specify a Dataset from the hub (the dataset will be downloaded automatically from the datasets Hub).
-
- # In distributed training, the load_dataset function guarantees that only one local process can concurrently
- # download the dataset.
- if args.dataset_name is not None:
- # Downloading and loading a dataset from the hub.
- dataset = load_dataset(args.dataset_name, task="image-classification")
- else:
- data_files = {}
- if args.train_dir is not None:
- data_files["train"] = os.path.join(args.train_dir, "**")
- if args.validation_dir is not None:
- data_files["validation"] = os.path.join(args.validation_dir, "**")
- dataset = load_dataset(
- "imagefolder",
- data_files=data_files,
- cache_dir=args.cache_dir,
- task="image-classification",
- )
- # See more about loading custom images at
- # https://huggingface.co/docs/datasets/v2.0.0/en/image_process#imagefolder.
-
- # If we don't have a validation split, split off a percentage of train as validation.
- args.train_val_split = None if "validation" in dataset.keys() else args.train_val_split
- if isinstance(args.train_val_split, float) and args.train_val_split > 0.0:
- split = dataset["train"].train_test_split(args.train_val_split)
- dataset["train"] = split["train"]
- dataset["validation"] = split["test"]
-
- # Prepare label mappings.
- # We'll include these in the model's config to get human readable labels in the Inference API.
- labels = dataset["train"].features["labels"].names
- label2id = {label: str(i) for i, label in enumerate(labels)}
- id2label = {str(i): label for i, label in enumerate(labels)}
-
- # Load pretrained model and image processor
- #
- # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently
- # download model & vocab.
- config = AutoConfig.from_pretrained(
- args.model_name_or_path,
- num_labels=len(labels),
- i2label=id2label,
- label2id=label2id,
- finetuning_task="image-classification",
- )
- image_processor = AutoImageProcessor.from_pretrained(args.model_name_or_path)
- model = AutoModelForImageClassification.from_pretrained(
- args.model_name_or_path,
- from_tf=bool(".ckpt" in args.model_name_or_path),
- config=config,
- ignore_mismatched_sizes=args.ignore_mismatched_sizes,
- )
-
- # Preprocessing the datasets
-
- # Define torchvision transforms to be applied to each image.
- if "shortest_edge" in image_processor.size:
- size = image_processor.size["shortest_edge"]
- else:
- size = (image_processor.size["height"], image_processor.size["width"])
- normalize = Normalize(mean=image_processor.image_mean, std=image_processor.image_std)
- train_transforms = Compose(
- [
- RandomResizedCrop(size),
- RandomHorizontalFlip(),
- ToTensor(),
- normalize,
- ]
- )
- val_transforms = Compose(
- [
- Resize(size),
- CenterCrop(size),
- ToTensor(),
- normalize,
- ]
- )
-
- def preprocess_train(example_batch):
- """Apply _train_transforms across a batch."""
- example_batch["pixel_values"] = [train_transforms(image.convert("RGB")) for image in example_batch["image"]]
- return example_batch
-
- def preprocess_val(example_batch):
- """Apply _val_transforms across a batch."""
- example_batch["pixel_values"] = [val_transforms(image.convert("RGB")) for image in example_batch["image"]]
- return example_batch
-
- with accelerator.main_process_first():
- if args.max_train_samples is not None:
- dataset["train"] = dataset["train"].shuffle(seed=args.seed).select(range(args.max_train_samples))
- # Set the training transforms
- train_dataset = dataset["train"].with_transform(preprocess_train)
- if args.max_eval_samples is not None:
- dataset["validation"] = dataset["validation"].shuffle(seed=args.seed).select(range(args.max_eval_samples))
- # Set the validation transforms
- eval_dataset = dataset["validation"].with_transform(preprocess_val)
-
- # DataLoaders creation:
- def collate_fn(examples):
- pixel_values = torch.stack([example["pixel_values"] for example in examples])
- labels = torch.tensor([example["labels"] for example in examples])
- return {"pixel_values": pixel_values, "labels": labels}
-
- train_dataloader = DataLoader(
- train_dataset, shuffle=True, collate_fn=collate_fn, batch_size=args.per_device_train_batch_size
- )
- eval_dataloader = DataLoader(eval_dataset, collate_fn=collate_fn, batch_size=args.per_device_eval_batch_size)
-
- # Optimizer
- # Split weights in two groups, one with weight decay and the other not.
- no_decay = ["bias", "LayerNorm.weight"]
- optimizer_grouped_parameters = [
- {
- "params": [p for n, p in model.named_parameters() if not any(nd in n for nd in no_decay)],
- "weight_decay": args.weight_decay,
- },
- {
- "params": [p for n, p in model.named_parameters() if any(nd in n for nd in no_decay)],
- "weight_decay": 0.0,
- },
- ]
- optimizer = torch.optim.AdamW(optimizer_grouped_parameters, lr=args.learning_rate)
-
- # Scheduler and math around the number of training steps.
- overrode_max_train_steps = False
- num_update_steps_per_epoch = math.ceil(len(train_dataloader) / args.gradient_accumulation_steps)
- if args.max_train_steps is None:
- args.max_train_steps = args.num_train_epochs * num_update_steps_per_epoch
- overrode_max_train_steps = True
-
- lr_scheduler = get_scheduler(
- name=args.lr_scheduler_type,
- optimizer=optimizer,
- num_warmup_steps=args.num_warmup_steps * args.gradient_accumulation_steps,
- num_training_steps=args.max_train_steps * args.gradient_accumulation_steps,
- )
-
- # Prepare everything with our `accelerator`.
- model, optimizer, train_dataloader, eval_dataloader, lr_scheduler = accelerator.prepare(
- model, optimizer, train_dataloader, eval_dataloader, lr_scheduler
- )
-
- # We need to recalculate our total training steps as the size of the training dataloader may have changed.
- num_update_steps_per_epoch = math.ceil(len(train_dataloader) / args.gradient_accumulation_steps)
- if overrode_max_train_steps:
- args.max_train_steps = args.num_train_epochs * num_update_steps_per_epoch
- # Afterwards we recalculate our number of training epochs
- args.num_train_epochs = math.ceil(args.max_train_steps / num_update_steps_per_epoch)
-
- # Figure out how many steps we should save the Accelerator states
- checkpointing_steps = args.checkpointing_steps
- if checkpointing_steps is not None and checkpointing_steps.isdigit():
- checkpointing_steps = int(checkpointing_steps)
-
- # We need to initialize the trackers we use, and also store our configuration.
- # The trackers initializes automatically on the main process.
- if args.with_tracking:
- experiment_config = vars(args)
- # TensorBoard cannot log Enums, need the raw value
- experiment_config["lr_scheduler_type"] = experiment_config["lr_scheduler_type"].value
- accelerator.init_trackers("image_classification_no_trainer", experiment_config)
-
- # Get the metric function
- metric = evaluate.load("accuracy")
-
- # Train!
- total_batch_size = args.per_device_train_batch_size * accelerator.num_processes * args.gradient_accumulation_steps
-
- logger.info("***** Running training *****")
- logger.info(f" Num examples = {len(train_dataset)}")
- logger.info(f" Num Epochs = {args.num_train_epochs}")
- logger.info(f" Instantaneous batch size per device = {args.per_device_train_batch_size}")
- logger.info(f" Total train batch size (w. parallel, distributed & accumulation) = {total_batch_size}")
- logger.info(f" Gradient Accumulation steps = {args.gradient_accumulation_steps}")
- logger.info(f" Total optimization steps = {args.max_train_steps}")
- # Only show the progress bar once on each machine.
- progress_bar = tqdm(range(args.max_train_steps), disable=not accelerator.is_local_main_process)
- completed_steps = 0
- starting_epoch = 0
- # Potentially load in the weights and states from a previous save
- if args.resume_from_checkpoint:
- if args.resume_from_checkpoint is not None or args.resume_from_checkpoint != "":
- accelerator.print(f"Resumed from checkpoint: {args.resume_from_checkpoint}")
- accelerator.load_state(args.resume_from_checkpoint)
- path = os.path.basename(args.resume_from_checkpoint)
- else:
- # Get the most recent checkpoint
- dirs = [f.name for f in os.scandir(os.getcwd()) if f.is_dir()]
- dirs.sort(key=os.path.getctime)
- path = dirs[-1] # Sorts folders by date modified, most recent checkpoint is the last
- # Extract `epoch_{i}` or `step_{i}`
- training_difference = os.path.splitext(path)[0]
-
- if "epoch" in training_difference:
- starting_epoch = int(training_difference.replace("epoch_", "")) + 1
- resume_step = None
- else:
- resume_step = int(training_difference.replace("step_", ""))
- starting_epoch = resume_step // len(train_dataloader)
- resume_step -= starting_epoch * len(train_dataloader)
-
- for epoch in range(starting_epoch, args.num_train_epochs):
- model.train()
- if args.with_tracking:
- total_loss = 0
- for step, batch in enumerate(train_dataloader):
- # We need to skip steps until we reach the resumed step
- if args.resume_from_checkpoint and epoch == starting_epoch:
- if resume_step is not None and step < resume_step:
- completed_steps += 1
- continue
-
- with accelerator.accumulate(model):
- outputs = model(**batch)
- loss = outputs.loss
- # We keep track of the loss at each epoch
- if args.with_tracking:
- total_loss += loss.detach().float()
- accelerator.backward(loss)
- optimizer.step()
- lr_scheduler.step()
- optimizer.zero_grad()
-
- # Checks if the accelerator has performed an optimization step behind the scenes
- if accelerator.sync_gradients:
- progress_bar.update(1)
- completed_steps += 1
-
- if isinstance(checkpointing_steps, int):
- if completed_steps % checkpointing_steps == 0:
- output_dir = f"step_{completed_steps }"
- if args.output_dir is not None:
- output_dir = os.path.join(args.output_dir, output_dir)
- accelerator.save_state(output_dir)
-
- if args.push_to_hub and epoch < args.num_train_epochs - 1:
- accelerator.wait_for_everyone()
- unwrapped_model = accelerator.unwrap_model(model)
- unwrapped_model.save_pretrained(
- args.output_dir,
- is_main_process=accelerator.is_main_process,
- save_function=accelerator.save,
- )
- if accelerator.is_main_process:
- image_processor.save_pretrained(args.output_dir)
- repo.push_to_hub(
- commit_message=f"Training in progress {completed_steps} steps",
- blocking=False,
- auto_lfs_prune=True,
- )
-
- if completed_steps >= args.max_train_steps:
- break
-
- model.eval()
- for step, batch in enumerate(eval_dataloader):
- with torch.no_grad():
- outputs = model(**batch)
- predictions = outputs.logits.argmax(dim=-1)
- predictions, references = accelerator.gather_for_metrics((predictions, batch["labels"]))
- metric.add_batch(
- predictions=predictions,
- references=references,
- )
-
- eval_metric = metric.compute()
- logger.info(f"epoch {epoch}: {eval_metric}")
-
- if args.with_tracking:
- accelerator.log(
- {
- "accuracy": eval_metric,
- "train_loss": total_loss.item() / len(train_dataloader),
- "epoch": epoch,
- "step": completed_steps,
- },
- step=completed_steps,
- )
-
- if args.push_to_hub and epoch < args.num_train_epochs - 1:
- accelerator.wait_for_everyone()
- unwrapped_model = accelerator.unwrap_model(model)
- unwrapped_model.save_pretrained(
- args.output_dir, is_main_process=accelerator.is_main_process, save_function=accelerator.save
- )
- if accelerator.is_main_process:
- image_processor.save_pretrained(args.output_dir)
- repo.push_to_hub(
- commit_message=f"Training in progress epoch {epoch}", blocking=False, auto_lfs_prune=True
- )
-
- if args.checkpointing_steps == "epoch":
- output_dir = f"epoch_{epoch}"
- if args.output_dir is not None:
- output_dir = os.path.join(args.output_dir, output_dir)
- accelerator.save_state(output_dir)
-
- if args.with_tracking:
- accelerator.end_training()
-
- if args.output_dir is not None:
- accelerator.wait_for_everyone()
- unwrapped_model = accelerator.unwrap_model(model)
- unwrapped_model.save_pretrained(
- args.output_dir, is_main_process=accelerator.is_main_process, save_function=accelerator.save
- )
- if accelerator.is_main_process:
- image_processor.save_pretrained(args.output_dir)
- if args.push_to_hub:
- repo.push_to_hub(commit_message="End of training", auto_lfs_prune=True)
-
- all_results = {f"eval_{k}": v for k, v in eval_metric.items()}
- with open(os.path.join(args.output_dir, "all_results.json"), "w") as f:
- json.dump(all_results, f)
-
-
-if __name__ == "__main__":
- main()
diff --git a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/PIL/ImageDraw.py b/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/PIL/ImageDraw.py
deleted file mode 100644
index 7d1790faa93e98dbdf32b1c3d1dd4b49b65e7cc0..0000000000000000000000000000000000000000
--- a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/PIL/ImageDraw.py
+++ /dev/null
@@ -1,1038 +0,0 @@
-#
-# The Python Imaging Library
-# $Id$
-#
-# drawing interface operations
-#
-# History:
-# 1996-04-13 fl Created (experimental)
-# 1996-08-07 fl Filled polygons, ellipses.
-# 1996-08-13 fl Added text support
-# 1998-06-28 fl Handle I and F images
-# 1998-12-29 fl Added arc; use arc primitive to draw ellipses
-# 1999-01-10 fl Added shape stuff (experimental)
-# 1999-02-06 fl Added bitmap support
-# 1999-02-11 fl Changed all primitives to take options
-# 1999-02-20 fl Fixed backwards compatibility
-# 2000-10-12 fl Copy on write, when necessary
-# 2001-02-18 fl Use default ink for bitmap/text also in fill mode
-# 2002-10-24 fl Added support for CSS-style color strings
-# 2002-12-10 fl Added experimental support for RGBA-on-RGB drawing
-# 2002-12-11 fl Refactored low-level drawing API (work in progress)
-# 2004-08-26 fl Made Draw() a factory function, added getdraw() support
-# 2004-09-04 fl Added width support to line primitive
-# 2004-09-10 fl Added font mode handling
-# 2006-06-19 fl Added font bearing support (getmask2)
-#
-# Copyright (c) 1997-2006 by Secret Labs AB
-# Copyright (c) 1996-2006 by Fredrik Lundh
-#
-# See the README file for information on usage and redistribution.
-#
-
-import math
-import numbers
-
-from . import Image, ImageColor
-
-"""
-A simple 2D drawing interface for PIL images.
-
-Application code should use the Draw factory, instead of
-directly.
-"""
-
-
-class ImageDraw:
- font = None
-
- def __init__(self, im, mode=None):
- """
- Create a drawing instance.
-
- :param im: The image to draw in.
- :param mode: Optional mode to use for color values. For RGB
- images, this argument can be RGB or RGBA (to blend the
- drawing into the image). For all other modes, this argument
- must be the same as the image mode. If omitted, the mode
- defaults to the mode of the image.
- """
- im.load()
- if im.readonly:
- im._copy() # make it writeable
- blend = 0
- if mode is None:
- mode = im.mode
- if mode != im.mode:
- if mode == "RGBA" and im.mode == "RGB":
- blend = 1
- else:
- msg = "mode mismatch"
- raise ValueError(msg)
- if mode == "P":
- self.palette = im.palette
- else:
- self.palette = None
- self._image = im
- self.im = im.im
- self.draw = Image.core.draw(self.im, blend)
- self.mode = mode
- if mode in ("I", "F"):
- self.ink = self.draw.draw_ink(1)
- else:
- self.ink = self.draw.draw_ink(-1)
- if mode in ("1", "P", "I", "F"):
- # FIXME: fix Fill2 to properly support matte for I+F images
- self.fontmode = "1"
- else:
- self.fontmode = "L" # aliasing is okay for other modes
- self.fill = False
-
- def getfont(self):
- """
- Get the current default font.
-
- To set the default font for this ImageDraw instance::
-
- from PIL import ImageDraw, ImageFont
- draw.font = ImageFont.truetype("Tests/fonts/FreeMono.ttf")
-
- To set the default font for all future ImageDraw instances::
-
- from PIL import ImageDraw, ImageFont
- ImageDraw.ImageDraw.font = ImageFont.truetype("Tests/fonts/FreeMono.ttf")
-
- If the current default font is ``None``,
- it is initialized with ``ImageFont.load_default()``.
-
- :returns: An image font."""
- if not self.font:
- # FIXME: should add a font repository
- from . import ImageFont
-
- self.font = ImageFont.load_default()
- return self.font
-
- def _getink(self, ink, fill=None):
- if ink is None and fill is None:
- if self.fill:
- fill = self.ink
- else:
- ink = self.ink
- else:
- if ink is not None:
- if isinstance(ink, str):
- ink = ImageColor.getcolor(ink, self.mode)
- if self.palette and not isinstance(ink, numbers.Number):
- ink = self.palette.getcolor(ink, self._image)
- ink = self.draw.draw_ink(ink)
- if fill is not None:
- if isinstance(fill, str):
- fill = ImageColor.getcolor(fill, self.mode)
- if self.palette and not isinstance(fill, numbers.Number):
- fill = self.palette.getcolor(fill, self._image)
- fill = self.draw.draw_ink(fill)
- return ink, fill
-
- def arc(self, xy, start, end, fill=None, width=1):
- """Draw an arc."""
- ink, fill = self._getink(fill)
- if ink is not None:
- self.draw.draw_arc(xy, start, end, ink, width)
-
- def bitmap(self, xy, bitmap, fill=None):
- """Draw a bitmap."""
- bitmap.load()
- ink, fill = self._getink(fill)
- if ink is None:
- ink = fill
- if ink is not None:
- self.draw.draw_bitmap(xy, bitmap.im, ink)
-
- def chord(self, xy, start, end, fill=None, outline=None, width=1):
- """Draw a chord."""
- ink, fill = self._getink(outline, fill)
- if fill is not None:
- self.draw.draw_chord(xy, start, end, fill, 1)
- if ink is not None and ink != fill and width != 0:
- self.draw.draw_chord(xy, start, end, ink, 0, width)
-
- def ellipse(self, xy, fill=None, outline=None, width=1):
- """Draw an ellipse."""
- ink, fill = self._getink(outline, fill)
- if fill is not None:
- self.draw.draw_ellipse(xy, fill, 1)
- if ink is not None and ink != fill and width != 0:
- self.draw.draw_ellipse(xy, ink, 0, width)
-
- def line(self, xy, fill=None, width=0, joint=None):
- """Draw a line, or a connected sequence of line segments."""
- ink = self._getink(fill)[0]
- if ink is not None:
- self.draw.draw_lines(xy, ink, width)
- if joint == "curve" and width > 4:
- if not isinstance(xy[0], (list, tuple)):
- xy = [tuple(xy[i : i + 2]) for i in range(0, len(xy), 2)]
- for i in range(1, len(xy) - 1):
- point = xy[i]
- angles = [
- math.degrees(math.atan2(end[0] - start[0], start[1] - end[1]))
- % 360
- for start, end in ((xy[i - 1], point), (point, xy[i + 1]))
- ]
- if angles[0] == angles[1]:
- # This is a straight line, so no joint is required
- continue
-
- def coord_at_angle(coord, angle):
- x, y = coord
- angle -= 90
- distance = width / 2 - 1
- return tuple(
- p + (math.floor(p_d) if p_d > 0 else math.ceil(p_d))
- for p, p_d in (
- (x, distance * math.cos(math.radians(angle))),
- (y, distance * math.sin(math.radians(angle))),
- )
- )
-
- flipped = (
- angles[1] > angles[0] and angles[1] - 180 > angles[0]
- ) or (angles[1] < angles[0] and angles[1] + 180 > angles[0])
- coords = [
- (point[0] - width / 2 + 1, point[1] - width / 2 + 1),
- (point[0] + width / 2 - 1, point[1] + width / 2 - 1),
- ]
- if flipped:
- start, end = (angles[1] + 90, angles[0] + 90)
- else:
- start, end = (angles[0] - 90, angles[1] - 90)
- self.pieslice(coords, start - 90, end - 90, fill)
-
- if width > 8:
- # Cover potential gaps between the line and the joint
- if flipped:
- gap_coords = [
- coord_at_angle(point, angles[0] + 90),
- point,
- coord_at_angle(point, angles[1] + 90),
- ]
- else:
- gap_coords = [
- coord_at_angle(point, angles[0] - 90),
- point,
- coord_at_angle(point, angles[1] - 90),
- ]
- self.line(gap_coords, fill, width=3)
-
- def shape(self, shape, fill=None, outline=None):
- """(Experimental) Draw a shape."""
- shape.close()
- ink, fill = self._getink(outline, fill)
- if fill is not None:
- self.draw.draw_outline(shape, fill, 1)
- if ink is not None and ink != fill:
- self.draw.draw_outline(shape, ink, 0)
-
- def pieslice(self, xy, start, end, fill=None, outline=None, width=1):
- """Draw a pieslice."""
- ink, fill = self._getink(outline, fill)
- if fill is not None:
- self.draw.draw_pieslice(xy, start, end, fill, 1)
- if ink is not None and ink != fill and width != 0:
- self.draw.draw_pieslice(xy, start, end, ink, 0, width)
-
- def point(self, xy, fill=None):
- """Draw one or more individual pixels."""
- ink, fill = self._getink(fill)
- if ink is not None:
- self.draw.draw_points(xy, ink)
-
- def polygon(self, xy, fill=None, outline=None, width=1):
- """Draw a polygon."""
- ink, fill = self._getink(outline, fill)
- if fill is not None:
- self.draw.draw_polygon(xy, fill, 1)
- if ink is not None and ink != fill and width != 0:
- if width == 1:
- self.draw.draw_polygon(xy, ink, 0, width)
- else:
- # To avoid expanding the polygon outwards,
- # use the fill as a mask
- mask = Image.new("1", self.im.size)
- mask_ink = self._getink(1)[0]
-
- fill_im = mask.copy()
- draw = Draw(fill_im)
- draw.draw.draw_polygon(xy, mask_ink, 1)
-
- ink_im = mask.copy()
- draw = Draw(ink_im)
- width = width * 2 - 1
- draw.draw.draw_polygon(xy, mask_ink, 0, width)
-
- mask.paste(ink_im, mask=fill_im)
-
- im = Image.new(self.mode, self.im.size)
- draw = Draw(im)
- draw.draw.draw_polygon(xy, ink, 0, width)
- self.im.paste(im.im, (0, 0) + im.size, mask.im)
-
- def regular_polygon(
- self, bounding_circle, n_sides, rotation=0, fill=None, outline=None, width=1
- ):
- """Draw a regular polygon."""
- xy = _compute_regular_polygon_vertices(bounding_circle, n_sides, rotation)
- self.polygon(xy, fill, outline, width)
-
- def rectangle(self, xy, fill=None, outline=None, width=1):
- """Draw a rectangle."""
- ink, fill = self._getink(outline, fill)
- if fill is not None:
- self.draw.draw_rectangle(xy, fill, 1)
- if ink is not None and ink != fill and width != 0:
- self.draw.draw_rectangle(xy, ink, 0, width)
-
- def rounded_rectangle(
- self, xy, radius=0, fill=None, outline=None, width=1, *, corners=None
- ):
- """Draw a rounded rectangle."""
- if isinstance(xy[0], (list, tuple)):
- (x0, y0), (x1, y1) = xy
- else:
- x0, y0, x1, y1 = xy
- if x1 < x0:
- msg = "x1 must be greater than or equal to x0"
- raise ValueError(msg)
- if y1 < y0:
- msg = "y1 must be greater than or equal to y0"
- raise ValueError(msg)
- if corners is None:
- corners = (True, True, True, True)
-
- d = radius * 2
-
- full_x, full_y = False, False
- if all(corners):
- full_x = d >= x1 - x0 - 1
- if full_x:
- # The two left and two right corners are joined
- d = x1 - x0
- full_y = d >= y1 - y0 - 1
- if full_y:
- # The two top and two bottom corners are joined
- d = y1 - y0
- if full_x and full_y:
- # If all corners are joined, that is a circle
- return self.ellipse(xy, fill, outline, width)
-
- if d == 0 or not any(corners):
- # If the corners have no curve,
- # or there are no corners,
- # that is a rectangle
- return self.rectangle(xy, fill, outline, width)
-
- r = d // 2
- ink, fill = self._getink(outline, fill)
-
- def draw_corners(pieslice):
- if full_x:
- # Draw top and bottom halves
- parts = (
- ((x0, y0, x0 + d, y0 + d), 180, 360),
- ((x0, y1 - d, x0 + d, y1), 0, 180),
- )
- elif full_y:
- # Draw left and right halves
- parts = (
- ((x0, y0, x0 + d, y0 + d), 90, 270),
- ((x1 - d, y0, x1, y0 + d), 270, 90),
- )
- else:
- # Draw four separate corners
- parts = []
- for i, part in enumerate(
- (
- ((x0, y0, x0 + d, y0 + d), 180, 270),
- ((x1 - d, y0, x1, y0 + d), 270, 360),
- ((x1 - d, y1 - d, x1, y1), 0, 90),
- ((x0, y1 - d, x0 + d, y1), 90, 180),
- )
- ):
- if corners[i]:
- parts.append(part)
- for part in parts:
- if pieslice:
- self.draw.draw_pieslice(*(part + (fill, 1)))
- else:
- self.draw.draw_arc(*(part + (ink, width)))
-
- if fill is not None:
- draw_corners(True)
-
- if full_x:
- self.draw.draw_rectangle((x0, y0 + r + 1, x1, y1 - r - 1), fill, 1)
- else:
- self.draw.draw_rectangle((x0 + r + 1, y0, x1 - r - 1, y1), fill, 1)
- if not full_x and not full_y:
- left = [x0, y0, x0 + r, y1]
- if corners[0]:
- left[1] += r + 1
- if corners[3]:
- left[3] -= r + 1
- self.draw.draw_rectangle(left, fill, 1)
-
- right = [x1 - r, y0, x1, y1]
- if corners[1]:
- right[1] += r + 1
- if corners[2]:
- right[3] -= r + 1
- self.draw.draw_rectangle(right, fill, 1)
- if ink is not None and ink != fill and width != 0:
- draw_corners(False)
-
- if not full_x:
- top = [x0, y0, x1, y0 + width - 1]
- if corners[0]:
- top[0] += r + 1
- if corners[1]:
- top[2] -= r + 1
- self.draw.draw_rectangle(top, ink, 1)
-
- bottom = [x0, y1 - width + 1, x1, y1]
- if corners[3]:
- bottom[0] += r + 1
- if corners[2]:
- bottom[2] -= r + 1
- self.draw.draw_rectangle(bottom, ink, 1)
- if not full_y:
- left = [x0, y0, x0 + width - 1, y1]
- if corners[0]:
- left[1] += r + 1
- if corners[3]:
- left[3] -= r + 1
- self.draw.draw_rectangle(left, ink, 1)
-
- right = [x1 - width + 1, y0, x1, y1]
- if corners[1]:
- right[1] += r + 1
- if corners[2]:
- right[3] -= r + 1
- self.draw.draw_rectangle(right, ink, 1)
-
- def _multiline_check(self, text):
- split_character = "\n" if isinstance(text, str) else b"\n"
-
- return split_character in text
-
- def _multiline_split(self, text):
- split_character = "\n" if isinstance(text, str) else b"\n"
-
- return text.split(split_character)
-
- def _multiline_spacing(self, font, spacing, stroke_width):
- return (
- self.textbbox((0, 0), "A", font, stroke_width=stroke_width)[3]
- + stroke_width
- + spacing
- )
-
- def text(
- self,
- xy,
- text,
- fill=None,
- font=None,
- anchor=None,
- spacing=4,
- align="left",
- direction=None,
- features=None,
- language=None,
- stroke_width=0,
- stroke_fill=None,
- embedded_color=False,
- *args,
- **kwargs,
- ):
- """Draw text."""
- if self._multiline_check(text):
- return self.multiline_text(
- xy,
- text,
- fill,
- font,
- anchor,
- spacing,
- align,
- direction,
- features,
- language,
- stroke_width,
- stroke_fill,
- embedded_color,
- )
-
- if embedded_color and self.mode not in ("RGB", "RGBA"):
- msg = "Embedded color supported only in RGB and RGBA modes"
- raise ValueError(msg)
-
- if font is None:
- font = self.getfont()
-
- def getink(fill):
- ink, fill = self._getink(fill)
- if ink is None:
- return fill
- return ink
-
- def draw_text(ink, stroke_width=0, stroke_offset=None):
- mode = self.fontmode
- if stroke_width == 0 and embedded_color:
- mode = "RGBA"
- coord = []
- start = []
- for i in range(2):
- coord.append(int(xy[i]))
- start.append(math.modf(xy[i])[0])
- try:
- mask, offset = font.getmask2(
- text,
- mode,
- direction=direction,
- features=features,
- language=language,
- stroke_width=stroke_width,
- anchor=anchor,
- ink=ink,
- start=start,
- *args,
- **kwargs,
- )
- coord = coord[0] + offset[0], coord[1] + offset[1]
- except AttributeError:
- try:
- mask = font.getmask(
- text,
- mode,
- direction,
- features,
- language,
- stroke_width,
- anchor,
- ink,
- start=start,
- *args,
- **kwargs,
- )
- except TypeError:
- mask = font.getmask(text)
- if stroke_offset:
- coord = coord[0] + stroke_offset[0], coord[1] + stroke_offset[1]
- if mode == "RGBA":
- # font.getmask2(mode="RGBA") returns color in RGB bands and mask in A
- # extract mask and set text alpha
- color, mask = mask, mask.getband(3)
- color.fillband(3, (ink >> 24) & 0xFF)
- x, y = coord
- self.im.paste(color, (x, y, x + mask.size[0], y + mask.size[1]), mask)
- else:
- self.draw.draw_bitmap(coord, mask, ink)
-
- ink = getink(fill)
- if ink is not None:
- stroke_ink = None
- if stroke_width:
- stroke_ink = getink(stroke_fill) if stroke_fill is not None else ink
-
- if stroke_ink is not None:
- # Draw stroked text
- draw_text(stroke_ink, stroke_width)
-
- # Draw normal text
- draw_text(ink, 0)
- else:
- # Only draw normal text
- draw_text(ink)
-
- def multiline_text(
- self,
- xy,
- text,
- fill=None,
- font=None,
- anchor=None,
- spacing=4,
- align="left",
- direction=None,
- features=None,
- language=None,
- stroke_width=0,
- stroke_fill=None,
- embedded_color=False,
- ):
- if direction == "ttb":
- msg = "ttb direction is unsupported for multiline text"
- raise ValueError(msg)
-
- if anchor is None:
- anchor = "la"
- elif len(anchor) != 2:
- msg = "anchor must be a 2 character string"
- raise ValueError(msg)
- elif anchor[1] in "tb":
- msg = "anchor not supported for multiline text"
- raise ValueError(msg)
-
- widths = []
- max_width = 0
- lines = self._multiline_split(text)
- line_spacing = self._multiline_spacing(font, spacing, stroke_width)
- for line in lines:
- line_width = self.textlength(
- line, font, direction=direction, features=features, language=language
- )
- widths.append(line_width)
- max_width = max(max_width, line_width)
-
- top = xy[1]
- if anchor[1] == "m":
- top -= (len(lines) - 1) * line_spacing / 2.0
- elif anchor[1] == "d":
- top -= (len(lines) - 1) * line_spacing
-
- for idx, line in enumerate(lines):
- left = xy[0]
- width_difference = max_width - widths[idx]
-
- # first align left by anchor
- if anchor[0] == "m":
- left -= width_difference / 2.0
- elif anchor[0] == "r":
- left -= width_difference
-
- # then align by align parameter
- if align == "left":
- pass
- elif align == "center":
- left += width_difference / 2.0
- elif align == "right":
- left += width_difference
- else:
- msg = 'align must be "left", "center" or "right"'
- raise ValueError(msg)
-
- self.text(
- (left, top),
- line,
- fill,
- font,
- anchor,
- direction=direction,
- features=features,
- language=language,
- stroke_width=stroke_width,
- stroke_fill=stroke_fill,
- embedded_color=embedded_color,
- )
- top += line_spacing
-
- def textlength(
- self,
- text,
- font=None,
- direction=None,
- features=None,
- language=None,
- embedded_color=False,
- ):
- """Get the length of a given string, in pixels with 1/64 precision."""
- if self._multiline_check(text):
- msg = "can't measure length of multiline text"
- raise ValueError(msg)
- if embedded_color and self.mode not in ("RGB", "RGBA"):
- msg = "Embedded color supported only in RGB and RGBA modes"
- raise ValueError(msg)
-
- if font is None:
- font = self.getfont()
- mode = "RGBA" if embedded_color else self.fontmode
- return font.getlength(text, mode, direction, features, language)
-
- def textbbox(
- self,
- xy,
- text,
- font=None,
- anchor=None,
- spacing=4,
- align="left",
- direction=None,
- features=None,
- language=None,
- stroke_width=0,
- embedded_color=False,
- ):
- """Get the bounding box of a given string, in pixels."""
- if embedded_color and self.mode not in ("RGB", "RGBA"):
- msg = "Embedded color supported only in RGB and RGBA modes"
- raise ValueError(msg)
-
- if self._multiline_check(text):
- return self.multiline_textbbox(
- xy,
- text,
- font,
- anchor,
- spacing,
- align,
- direction,
- features,
- language,
- stroke_width,
- embedded_color,
- )
-
- if font is None:
- font = self.getfont()
- mode = "RGBA" if embedded_color else self.fontmode
- bbox = font.getbbox(
- text, mode, direction, features, language, stroke_width, anchor
- )
- return bbox[0] + xy[0], bbox[1] + xy[1], bbox[2] + xy[0], bbox[3] + xy[1]
-
- def multiline_textbbox(
- self,
- xy,
- text,
- font=None,
- anchor=None,
- spacing=4,
- align="left",
- direction=None,
- features=None,
- language=None,
- stroke_width=0,
- embedded_color=False,
- ):
- if direction == "ttb":
- msg = "ttb direction is unsupported for multiline text"
- raise ValueError(msg)
-
- if anchor is None:
- anchor = "la"
- elif len(anchor) != 2:
- msg = "anchor must be a 2 character string"
- raise ValueError(msg)
- elif anchor[1] in "tb":
- msg = "anchor not supported for multiline text"
- raise ValueError(msg)
-
- widths = []
- max_width = 0
- lines = self._multiline_split(text)
- line_spacing = self._multiline_spacing(font, spacing, stroke_width)
- for line in lines:
- line_width = self.textlength(
- line,
- font,
- direction=direction,
- features=features,
- language=language,
- embedded_color=embedded_color,
- )
- widths.append(line_width)
- max_width = max(max_width, line_width)
-
- top = xy[1]
- if anchor[1] == "m":
- top -= (len(lines) - 1) * line_spacing / 2.0
- elif anchor[1] == "d":
- top -= (len(lines) - 1) * line_spacing
-
- bbox = None
-
- for idx, line in enumerate(lines):
- left = xy[0]
- width_difference = max_width - widths[idx]
-
- # first align left by anchor
- if anchor[0] == "m":
- left -= width_difference / 2.0
- elif anchor[0] == "r":
- left -= width_difference
-
- # then align by align parameter
- if align == "left":
- pass
- elif align == "center":
- left += width_difference / 2.0
- elif align == "right":
- left += width_difference
- else:
- msg = 'align must be "left", "center" or "right"'
- raise ValueError(msg)
-
- bbox_line = self.textbbox(
- (left, top),
- line,
- font,
- anchor,
- direction=direction,
- features=features,
- language=language,
- stroke_width=stroke_width,
- embedded_color=embedded_color,
- )
- if bbox is None:
- bbox = bbox_line
- else:
- bbox = (
- min(bbox[0], bbox_line[0]),
- min(bbox[1], bbox_line[1]),
- max(bbox[2], bbox_line[2]),
- max(bbox[3], bbox_line[3]),
- )
-
- top += line_spacing
-
- if bbox is None:
- return xy[0], xy[1], xy[0], xy[1]
- return bbox
-
-
-def Draw(im, mode=None):
- """
- A simple 2D drawing interface for PIL images.
-
- :param im: The image to draw in.
- :param mode: Optional mode to use for color values. For RGB
- images, this argument can be RGB or RGBA (to blend the
- drawing into the image). For all other modes, this argument
- must be the same as the image mode. If omitted, the mode
- defaults to the mode of the image.
- """
- try:
- return im.getdraw(mode)
- except AttributeError:
- return ImageDraw(im, mode)
-
-
-# experimental access to the outline API
-try:
- Outline = Image.core.outline
-except AttributeError:
- Outline = None
-
-
-def getdraw(im=None, hints=None):
- """
- (Experimental) A more advanced 2D drawing interface for PIL images,
- based on the WCK interface.
-
- :param im: The image to draw in.
- :param hints: An optional list of hints.
- :returns: A (drawing context, drawing resource factory) tuple.
- """
- # FIXME: this needs more work!
- # FIXME: come up with a better 'hints' scheme.
- handler = None
- if not hints or "nicest" in hints:
- try:
- from . import _imagingagg as handler
- except ImportError:
- pass
- if handler is None:
- from . import ImageDraw2 as handler
- if im:
- im = handler.Draw(im)
- return im, handler
-
-
-def floodfill(image, xy, value, border=None, thresh=0):
- """
- (experimental) Fills a bounded region with a given color.
-
- :param image: Target image.
- :param xy: Seed position (a 2-item coordinate tuple). See
- :ref:`coordinate-system`.
- :param value: Fill color.
- :param border: Optional border value. If given, the region consists of
- pixels with a color different from the border color. If not given,
- the region consists of pixels having the same color as the seed
- pixel.
- :param thresh: Optional threshold value which specifies a maximum
- tolerable difference of a pixel value from the 'background' in
- order for it to be replaced. Useful for filling regions of
- non-homogeneous, but similar, colors.
- """
- # based on an implementation by Eric S. Raymond
- # amended by yo1995 @20180806
- pixel = image.load()
- x, y = xy
- try:
- background = pixel[x, y]
- if _color_diff(value, background) <= thresh:
- return # seed point already has fill color
- pixel[x, y] = value
- except (ValueError, IndexError):
- return # seed point outside image
- edge = {(x, y)}
- # use a set to keep record of current and previous edge pixels
- # to reduce memory consumption
- full_edge = set()
- while edge:
- new_edge = set()
- for x, y in edge: # 4 adjacent method
- for s, t in ((x + 1, y), (x - 1, y), (x, y + 1), (x, y - 1)):
- # If already processed, or if a coordinate is negative, skip
- if (s, t) in full_edge or s < 0 or t < 0:
- continue
- try:
- p = pixel[s, t]
- except (ValueError, IndexError):
- pass
- else:
- full_edge.add((s, t))
- if border is None:
- fill = _color_diff(p, background) <= thresh
- else:
- fill = p != value and p != border
- if fill:
- pixel[s, t] = value
- new_edge.add((s, t))
- full_edge = edge # discard pixels processed
- edge = new_edge
-
-
-def _compute_regular_polygon_vertices(bounding_circle, n_sides, rotation):
- """
- Generate a list of vertices for a 2D regular polygon.
-
- :param bounding_circle: The bounding circle is a tuple defined
- by a point and radius. The polygon is inscribed in this circle.
- (e.g. ``bounding_circle=(x, y, r)`` or ``((x, y), r)``)
- :param n_sides: Number of sides
- (e.g. ``n_sides=3`` for a triangle, ``6`` for a hexagon)
- :param rotation: Apply an arbitrary rotation to the polygon
- (e.g. ``rotation=90``, applies a 90 degree rotation)
- :return: List of regular polygon vertices
- (e.g. ``[(25, 50), (50, 50), (50, 25), (25, 25)]``)
-
- How are the vertices computed?
- 1. Compute the following variables
- - theta: Angle between the apothem & the nearest polygon vertex
- - side_length: Length of each polygon edge
- - centroid: Center of bounding circle (1st, 2nd elements of bounding_circle)
- - polygon_radius: Polygon radius (last element of bounding_circle)
- - angles: Location of each polygon vertex in polar grid
- (e.g. A square with 0 degree rotation => [225.0, 315.0, 45.0, 135.0])
-
- 2. For each angle in angles, get the polygon vertex at that angle
- The vertex is computed using the equation below.
- X= xcos(φ) + ysin(φ)
- Y= −xsin(φ) + ycos(φ)
-
- Note:
- φ = angle in degrees
- x = 0
- y = polygon_radius
-
- The formula above assumes rotation around the origin.
- In our case, we are rotating around the centroid.
- To account for this, we use the formula below
- X = xcos(φ) + ysin(φ) + centroid_x
- Y = −xsin(φ) + ycos(φ) + centroid_y
- """
- # 1. Error Handling
- # 1.1 Check `n_sides` has an appropriate value
- if not isinstance(n_sides, int):
- msg = "n_sides should be an int"
- raise TypeError(msg)
- if n_sides < 3:
- msg = "n_sides should be an int > 2"
- raise ValueError(msg)
-
- # 1.2 Check `bounding_circle` has an appropriate value
- if not isinstance(bounding_circle, (list, tuple)):
- msg = "bounding_circle should be a tuple"
- raise TypeError(msg)
-
- if len(bounding_circle) == 3:
- *centroid, polygon_radius = bounding_circle
- elif len(bounding_circle) == 2:
- centroid, polygon_radius = bounding_circle
- else:
- msg = (
- "bounding_circle should contain 2D coordinates "
- "and a radius (e.g. (x, y, r) or ((x, y), r) )"
- )
- raise ValueError(msg)
-
- if not all(isinstance(i, (int, float)) for i in (*centroid, polygon_radius)):
- msg = "bounding_circle should only contain numeric data"
- raise ValueError(msg)
-
- if not len(centroid) == 2:
- msg = "bounding_circle centre should contain 2D coordinates (e.g. (x, y))"
- raise ValueError(msg)
-
- if polygon_radius <= 0:
- msg = "bounding_circle radius should be > 0"
- raise ValueError(msg)
-
- # 1.3 Check `rotation` has an appropriate value
- if not isinstance(rotation, (int, float)):
- msg = "rotation should be an int or float"
- raise ValueError(msg)
-
- # 2. Define Helper Functions
- def _apply_rotation(point, degrees, centroid):
- return (
- round(
- point[0] * math.cos(math.radians(360 - degrees))
- - point[1] * math.sin(math.radians(360 - degrees))
- + centroid[0],
- 2,
- ),
- round(
- point[1] * math.cos(math.radians(360 - degrees))
- + point[0] * math.sin(math.radians(360 - degrees))
- + centroid[1],
- 2,
- ),
- )
-
- def _compute_polygon_vertex(centroid, polygon_radius, angle):
- start_point = [polygon_radius, 0]
- return _apply_rotation(start_point, angle, centroid)
-
- def _get_angles(n_sides, rotation):
- angles = []
- degrees = 360 / n_sides
- # Start with the bottom left polygon vertex
- current_angle = (270 - 0.5 * degrees) + rotation
- for _ in range(0, n_sides):
- angles.append(current_angle)
- current_angle += degrees
- if current_angle > 360:
- current_angle -= 360
- return angles
-
- # 3. Variable Declarations
- angles = _get_angles(n_sides, rotation)
-
- # 4. Compute Vertices
- return [
- _compute_polygon_vertex(centroid, polygon_radius, angle) for angle in angles
- ]
-
-
-def _color_diff(color1, color2):
- """
- Uses 1-norm distance to calculate difference between two values.
- """
- if isinstance(color2, tuple):
- return sum(abs(color1[i] - color2[i]) for i in range(0, len(color2)))
- else:
- return abs(color1 - color2)
diff --git a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/filelock/_error.py b/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/filelock/_error.py
deleted file mode 100644
index f7ff08c0f508ad7077eb6ed1990898840c952b3a..0000000000000000000000000000000000000000
--- a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/filelock/_error.py
+++ /dev/null
@@ -1,30 +0,0 @@
-from __future__ import annotations
-
-from typing import Any
-
-
-class Timeout(TimeoutError): # noqa: N818
- """Raised when the lock could not be acquired in *timeout* seconds."""
-
- def __init__(self, lock_file: str) -> None:
- super().__init__()
- self._lock_file = lock_file
-
- def __reduce__(self) -> str | tuple[Any, ...]:
- return self.__class__, (self._lock_file,) # Properly pickle the exception
-
- def __str__(self) -> str:
- return f"The file lock '{self._lock_file}' could not be acquired."
-
- def __repr__(self) -> str:
- return f"{self.__class__.__name__}({self.lock_file!r})"
-
- @property
- def lock_file(self) -> str:
- """:return: The path of the file lock."""
- return self._lock_file
-
-
-__all__ = [
- "Timeout",
-]
diff --git a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/gradio/templates/cdn/assets/Model3D-b938dbb2.js b/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/gradio/templates/cdn/assets/Model3D-b938dbb2.js
deleted file mode 100644
index e5609085718777ad6a3e67dc27d3283303e52714..0000000000000000000000000000000000000000
--- a/spaces/chuan-hd/law-assistant-chatbot/.venv/lib/python3.11/site-packages/gradio/templates/cdn/assets/Model3D-b938dbb2.js
+++ /dev/null
@@ -1,2 +0,0 @@
-import{S as o,e as d,s as u,N as _,P as g,K as r,U as i,p as v,M as y,R as m,n as c,A as b}from"./index-f877dfd5.js";function M(a){let e,s;return{c(){e=_("div"),s=g(a[0]),r(e,"class","svelte-1ayixqk"),i(e,"table",a[1]==="table"),i(e,"gallery",a[1]==="gallery"),i(e,"selected",a[2])},m(t,l){v(t,e,l),y(e,s)},p(t,[l]){l&1&&m(s,t[0]),l&2&&i(e,"table",t[1]==="table"),l&2&&i(e,"gallery",t[1]==="gallery"),l&4&&i(e,"selected",t[2])},i:c,o:c,d(t){t&&b(e)}}}function D(a,e,s){let{value:t}=e,{type:l}=e,{selected:f=!1}=e;return a.$$set=n=>{"value"in n&&s(0,t=n.value),"type"in n&&s(1,l=n.type),"selected"in n&&s(2,f=n.selected)},[t,l,f]}class h extends o{constructor(e){super(),d(this,e,D,M,u,{value:0,type:1,selected:2})}}const E=h;export{E};
-//# sourceMappingURL=Model3D-b938dbb2.js.map
diff --git a/spaces/cifkao/context-probing/highlighted_text/build/static/css/main.59eacdd9.chunk.css b/spaces/cifkao/context-probing/highlighted_text/build/static/css/main.59eacdd9.chunk.css
deleted file mode 100644
index bfa18c40d0051e3dd6d924316defaddbf2bec156..0000000000000000000000000000000000000000
--- a/spaces/cifkao/context-probing/highlighted_text/build/static/css/main.59eacdd9.chunk.css
+++ /dev/null
@@ -1,2 +0,0 @@
-body{padding:0;margin:0;font-family:"Source Sans Pro",sans-serif;font-size:1rem;line-height:1.4}.container{border:1px solid #d2d2d2;border-radius:5px;padding:12px 16px}.highlighted-text{color:#000;background-color:#fff;background-color:#fefefe;border:1px solid rgba(49,51,63,.2);border-radius:5px;padding:5px;max-height:350px;overflow-y:auto;cursor:pointer}.highlighted-text .token.prefix~.token:not(.prefix){color:#2563eb}.highlighted-text .token.active{outline:1px solid #444}.status-bar{min-height:1.4em;padding-bottom:12px}.status-bar .token{border:1px solid #aaa;border-radius:2px;background:#f5f5f5;padding:1px;margin:1px 2px 1px 1px}
-/*# sourceMappingURL=main.59eacdd9.chunk.css.map */
\ No newline at end of file
diff --git a/spaces/cihyFjudo/fairness-paper-search/Download Internet Download Manager Keygen Patch The Best Tool for Managing and Accelerating Your Downloads.md b/spaces/cihyFjudo/fairness-paper-search/Download Internet Download Manager Keygen Patch The Best Tool for Managing and Accelerating Your Downloads.md
deleted file mode 100644
index 93eb3eb15d127dcca90628ce6b5aa69b7f79f69f..0000000000000000000000000000000000000000
--- a/spaces/cihyFjudo/fairness-paper-search/Download Internet Download Manager Keygen Patch The Best Tool for Managing and Accelerating Your Downloads.md
+++ /dev/null
@@ -1,24 +0,0 @@
-
-
IDM Crack with Internet Download Manager (IDM) is a tool to increase download speeds, resume, and schedule downloads. Comprehensive error recovery and resume capability will restart broken or interrupted downloads due to lost connections, network problems, computer shutdowns, or unexpected power outages. The simple graphic user interface makes IDM user-friendly and easy to use. Internet Download Manager has a smart download logic accelerator that features intelligent dynamic file segmentation and safe multipart downloading technology to accelerate your downloads. Unlike other download managers and accelerators, Internet Download Manager segments downloaded files dynamically during the download process and reuse available connections without additional connect and login stages to achieve the best acceleration performance.
-Install Internet Download Manager (IDM) to stop waiting for your downloads. You will be truly amazed how fast IDM downloads your files. IDM will also repair broken and resume interrupted downloads due to lost connections, network problems, computer shutdowns,or unexpected power outages.
-Download Internet Download Manager Keygen Patch
Download Zip ✦✦✦ https://tinurli.com/2uwjMy
-Our powerful download engine uses unique algorithms to receive Internet data in a fastest possible way. IDM will accelerate downloads all times because of its innovative dynamic file segmentation technology.Unlike other download managers and accelerators, IDM segments downloaded files dynamically during download process, and it reuses available connections without additional connect and login stages to achieve the best possible acceleration performance.Our engineers have a lot of experience in download acceleration, and we constantly improve this download engine since 1999.
-After installing "IDM integration module" browser extension, just continue surfing the Internet, and you will be amazed how easy is to download everything you want from your favorite web sites with IDM video download panel.
-Internet Download Manager can connect to the Internet at a set time, download the files you want, disconnect, or shut down your computer when it's done. You can also synchronize changes using periodic synchronization of files.It's possible to create and schedule several download queues for downloading or synchronization.
-You may choose the order, buttons and columns to appear on the main IDM window. There are several different skins for the toolbar with different button styles. All skins can be downloaded from IDM home site. Also users can design their own skins.Also you can choose the light or dark IDM theme.
Choose and set up new IDM toolbar
-Internet Download Manager (also called IDM) is a shareware download manager software application owned by American company Tonec, Inc. It is only available for the Microsoft Windows operating system.
-Internet Download Manager (IDM) is a tool that manages and schedules downloads. It can use full bandwidth. It has recovery and resume capabilities to restore the interrupted downloads due to lost connection, network issues, and power outages.
-IDM supports a wide range of proxy servers such as firewall, FTP, and HTTP protocols, cookies, MP3 audio and MPEG video processing. It can work with Google Chrome, Internet Explorer, Mozilla Firefox, Opera and other popular browsers to manage the download.[3]
-Internet Download Manager is a tool for increasing download speeds by up to 5 times, and for resuming, scheduling, and organizing downloads. The program will resume unfinished downloads due to network problems, or unexpected power outages.
-
-The program features a full-fledged site grabber that downloads files that are specified with filters, for example, all pictures from a Web site, different parts of Web sites, or complete Web sites for offline browsing. The program supports HTTP, HTTPS, FTP and MMS protocols, and has an adaptive download accelerator for MP3 audio, FLV and MPEG video files. The program also features Download Video and Audio Panels for Internet Exporer, Chrome, Opera, Safari, Firefox and other Mozilla based browsers that appears on top of a web-player and can be used to download flash videos from sites like YouTube, MySpaceTV, Google Videos.
-Internet Download Manager IDM full version free download has a simple graphical user interface, making it user-friendly and easy to use. Internet Download Manager IDM Crack serial key free download has smart download logic accelerator, intelligent dynamic file segmentation and safe multi-part download technology, speeding up download speed. Unlike other download managers and accelerators, the Internet Download Manager full version latest 2023 dynamically downloads files. It reuses available connections during the download process, without requiring additional connection and login phases to get the best acceleration performance.
-IDM Crack with Internet Download Manager download-adds Windows 10 compatibility and adds an IDM download panel to the web player. It also has full Windows 11, Windows 8.1 (Windows 8, Windows 7 and Vista) support, page scraping tools, redeveloped planning procedures, and MMS protocol support. The new version of IDM Patch 2023 also adds improved integration based on IE 11 and IE, redesigned and enhanced download engine, unique advanced integration with all latest browsers, improved toolbar, and many other improvements new features. You can download IDM crack for Windows 7, Windows 8, Windows 8.1, Windows 10 and Windows 11.
-IDM download free full version with serial key downloads all necessary files from a website specified by filters, such as all images on a website, a subset of a website, or an entire website for offline browsing. Multiple scrape items can be scheduled to run once at a specific time, stop at a specific time, or run periodically to synchronize changes. IDM Full Cracked can add links to all downloads on the current page. Multiple files can be easily downloaded with this feature. It can be used to automatically organize downloads using defined download categories.
-The IDM Cracked free download version of Internet Download Manager serial number includes multi-language support, zip preview, download categories, plan professionals, voices of different events, HTTPS support, queue handlers, HTML help and tutorials. After the download completed, virus protection is enhanced. The following methods Gradually download quota (useful for using certain types of connections) fair access policy or FAP, such as Direcway, Direct PC, Hughes, built-in download accelerator, etc. You can also download IDM Crack 2023 Torrent.
-Internet Download Manager provides users with speed and time management is the Speed Limiter. This option limits download speed to some extent. To check if this option has been disabled from the IDM application, go to Downloads>Speed Limiter and click Turn off.
-Please follow the instruction here
Open Program Files (x86) and find internet download manager IDM folder.
Find the file named IDMGCExt.crx in the folder.
Drag and drop the file into the Google Chrome extension tab. The IDM extension will be installed in the Chrome browser.
-Identify and terminate files detected as PUA.Win32.KeyGen.ITZ [ Learn More ][ back ]
Windows Task Manager may not display all running processes. In this case, please use a third-party process viewer, preferably Process Explorer, to terminate the malware/grayware/spyware file. You may download the said tool here. If the detected file is displayed in either Windows Task Manager or Process Explorer but you cannot delete it, restart your computer in safe mode. To do this, refer to this link for the complete steps. If the detected file is not displayed in either Windows Task Manager or Process Explorer, continue doing the next steps. To terminate the malware/grayware/spyware process:
-idm Serial key is the most fabulous software. Most of the people use this software to download videos, software, games and documents file.you can speedy and easy downloading. IDM is highly and well-performed software able us to solve the problems. According to the IDM users, Internet Download Manager is the fastest tool to download your favorite software and games. You can learn more and more about this.
aaccfb2cb3
-
-
\ No newline at end of file
diff --git a/spaces/cihyFjudo/fairness-paper-search/South Koreans Can Buy Cell Phones AnyWhere from Today Even at SuperMarts How It Works and What It Means for You.md b/spaces/cihyFjudo/fairness-paper-search/South Koreans Can Buy Cell Phones AnyWhere from Today Even at SuperMarts How It Works and What It Means for You.md
deleted file mode 100644
index 0ae40e0a2ddf243fe9f9a21cb6baee3413684656..0000000000000000000000000000000000000000
--- a/spaces/cihyFjudo/fairness-paper-search/South Koreans Can Buy Cell Phones AnyWhere from Today Even at SuperMarts How It Works and What It Means for You.md
+++ /dev/null
@@ -1,9 +0,0 @@
-
-One of the first critical steps is to break the mind-set that equates manufacturing with job growth. The notion that gains in manufacturing will bring equivalent gains in jobs is deeply rooted in the minds of many South Koreas because of the important role manufacturing played in the early development of their economy. Over the long run, however, gains in manufacturing bring ever higher levels of automation and thus come at the expense of jobs. Indeed, between 1995 and 2002, nearly 22 million manufacturing jobs disappeared from the global economy despite numerous policy efforts to promote employment in that sector. No mature economy today, not even Germany or Japan, generates net job growth in manufacturing. South Korea is not immune to this trend, having lost nearly 740,000 manufacturing jobs from 1995 to 2008. Today, manufacturing accounts for 4.1 million jobs out of 23.6 million total.
-South Koreans Can Buy Cell Phones AnyWhere from Today Even at SuperMarts
Download ••• https://tinurli.com/2uwiLQ
-What we are seeing today is only the beginning. Soon it will be hard even to define e-commerce, let alone measure it. Is it an e-commerce sale if the customer goes to a store, finds that the product is out of stock, and uses an in-store terminal to have another location ship it to her home? What if the customer is shopping in one store, uses his smartphone to find a lower price at another, and then orders it electronically for in-store pickup? How about gifts that are ordered from a website but exchanged at a local store? Experts estimate that digital information already influences about 50% of store sales, and that number is growing rapidly.
-Oxxo is the largest chain in the country, with more than 15,000 stores around the country. Other convenience stores, such as Tiendas Extra, 7-Eleven, SuperCity, ampm, and Circle K, are also found in Mexico. The first convenience store in the country, Super 7 (now a 7-Eleven), was opened in 1976 in Monterrey, Nuevo León.[citation needed] There are also some regional chains, like Amigo Express and CB Mas, that operate in Comarca Lagunera, Super Q and El Matador in Queretaro, Coyote in central Mexico, and JV in northeastern Mexico. Stores sell fast food like coffee, hot dogs, nachos and prepaid cellphones between MXN$20 and MXN$500, mainly Telcel and Movistar, newspapers, magazines, and Panini products and other novelties.
-Misceláneas (literally meaning "place where miscellaneous items are sold" and otherwise called tiendas de abarrotes (grocery store) in some parts of the country) are smaller, family-run convenience stores often found in central and southern Mexico. They operate in many locations, from rural communities to suburban residential neighborhoods, usually located in front of or below the family's residence. They often fulfill the role of neighborhood meeting points and places to disseminate community news. While offering a more limited, and sometimes varied, assortment of items than corporate chains, they fill a void in areas where corporations do not operate. Usually they sell homemade snacks such as tortas and sandwiches, made by the owners. They also provide items in smaller quantities than would be offered for sale in larger stores and markets; for example, selling single cigarettes along with full packs.[29]
- aaccfb2cb3
-
-
\ No newline at end of file
diff --git a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/indeo4data.h b/spaces/colakin/video-generater/public/ffmpeg/libavcodec/indeo4data.h
deleted file mode 100644
index cc497c23912cb231398eac0ad2eeee61178692f4..0000000000000000000000000000000000000000
--- a/spaces/colakin/video-generater/public/ffmpeg/libavcodec/indeo4data.h
+++ /dev/null
@@ -1,350 +0,0 @@
-/*
- * Indeo Video Interactive 4 compatible decoder
- * Copyright (c) 2009-2010 Maxim Poliakovski
- *
- * This file is part of FFmpeg.
- *
- * FFmpeg is free software; you can redistribute it and/or
- * modify it under the terms of the GNU Lesser General Public
- * License as published by the Free Software Foundation; either
- * version 2.1 of the License, or (at your option) any later version.
- *
- * FFmpeg is distributed in the hope that it will be useful,
- * but WITHOUT ANY WARRANTY; without even the implied warranty of
- * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
- * Lesser General Public License for more details.
- *
- * You should have received a copy of the GNU Lesser General Public
- * License along with FFmpeg; if not, write to the Free Software
- * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
- */
-
-/**
- * @file
- * This file contains data needed for the Indeo 4 decoder.
- */
-
-#ifndef AVCODEC_INDEO4DATA_H
-#define AVCODEC_INDEO4DATA_H
-
-#include
-
-#include "ivi.h"
-
-/**
- * standard picture dimensions
- */
-static const uint16_t ivi4_common_pic_sizes[14] = {
- 640, 480, 320, 240, 160, 120, 704, 480, 352, 240, 352, 288, 176, 144
-};
-
-/**
- * Indeo 4 8x8 scan (zigzag) patterns
- */
-static const uint8_t ivi4_alternate_scan_8x8[64] = {
- 0, 8, 1, 9, 16, 24, 2, 3, 17, 25, 10, 11, 32, 40, 48, 56,
- 4, 5, 6, 7, 33, 41, 49, 57, 18, 19, 26, 27, 12, 13, 14, 15,
- 34, 35, 43, 42, 50, 51, 59, 58, 20, 21, 22, 23, 31, 30, 29, 28,
- 36, 37, 38, 39, 47, 46, 45, 44, 52, 53, 54, 55, 63, 62, 61, 60
-};
-
-static const uint8_t ivi4_alternate_scan_4x4[16] = {
- 0, 1, 4, 5, 8, 12, 2, 3, 9, 13, 6, 7, 10, 11, 14, 15
-};
-
-static const uint8_t ivi4_vertical_scan_4x4[16] = {
- 0, 4, 8, 12, 1, 5, 9, 13, 2, 6, 10, 14, 3, 7, 11, 15
-};
-
-static const uint8_t ivi4_horizontal_scan_4x4[16] = {
- 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15
-};
-
-static const uint8_t * const scan_index_to_tab[15] = {
- // for 8x8 transforms
- ff_zigzag_direct,
- ivi4_alternate_scan_8x8,
- ff_ivi_horizontal_scan_8x8,
- ff_ivi_vertical_scan_8x8,
- ff_zigzag_direct,
-
- // for 4x4 transforms
- ff_ivi_direct_scan_4x4,
- ivi4_alternate_scan_4x4,
- ivi4_vertical_scan_4x4,
- ivi4_horizontal_scan_4x4,
- ff_ivi_direct_scan_4x4,
-
- // TODO: check if those are needed
- ff_ivi_horizontal_scan_8x8,
- ff_ivi_horizontal_scan_8x8,
- ff_ivi_horizontal_scan_8x8,
- ff_ivi_horizontal_scan_8x8,
- ff_ivi_horizontal_scan_8x8
-};
-
-/**
- * Indeo 4 dequant tables
- */
-static const uint16_t ivi4_quant_8x8_intra[9][64] = {
- {
- 43, 342, 385, 470, 555, 555, 598, 726,
- 342, 342, 470, 513, 555, 598, 726, 769,
- 385, 470, 555, 555, 598, 726, 726, 811,
- 470, 470, 555, 555, 598, 726, 769, 854,
- 470, 555, 555, 598, 683, 726, 854, 1025,
- 555, 555, 598, 683, 726, 854, 1025, 1153,
- 555, 555, 598, 726, 811, 982, 1195, 1451,
- 555, 598, 726, 811, 982, 1195, 1451, 1793
- },
- {
- 86, 1195, 2390, 2390, 4865, 4865, 4865, 4865,
- 1195, 1195, 2390, 2390, 4865, 4865, 4865, 4865,
- 2390, 2390, 4865, 4865, 6827, 6827, 6827, 6827,
- 2390, 2390, 4865, 4865, 6827, 6827, 6827, 6827,
- 4865, 4865, 6827, 6827, 6827, 6827, 6827, 6827,
- 4865, 4865, 6827, 6827, 6827, 6827, 6827, 6827,
- 4865, 4865, 6827, 6827, 6827, 6827, 6827, 6827,
- 4865, 4865, 6827, 6827, 6827, 6827, 6827, 6827
- },
- {
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835,
- 235, 1067, 1195, 1323, 1451, 1579, 1707, 1835
- },
- {
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414,
- 1707, 1707, 3414, 3414, 3414, 3414, 3414, 3414
- },
- {
- 897, 897, 897, 897, 897, 897, 897, 897,
- 1067, 1067, 1067, 1067, 1067, 1067, 1067, 1067,
- 1238, 1238, 1238, 1238, 1238, 1238, 1238, 1238,
- 1409, 1409, 1409, 1409, 1409, 1409, 1409, 1409,
- 1579, 1579, 1579, 1579, 1579, 1579, 1579, 1579,
- 1750, 1750, 1750, 1750, 1750, 1750, 1750, 1750,
- 1921, 1921, 1921, 1921, 1921, 1921, 1921, 1921,
- 2091, 2091, 2091, 2091, 2091, 2091, 2091, 2091
- },
- {
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414
- },
- {
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390,
- 2390, 2390, 2390, 2390, 2390, 2390, 2390, 2390
- },
- {
- 22, 171, 214, 257, 257, 299, 299, 342,
- 171, 171, 257, 257, 299, 299, 342, 385,
- 214, 257, 257, 299, 299, 342, 342, 385,
- 257, 257, 257, 299, 299, 342, 385, 427,
- 257, 257, 299, 299, 342, 385, 427, 513,
- 257, 299, 299, 342, 385, 427, 513, 598,
- 299, 299, 299, 385, 385, 470, 598, 726,
- 299, 299, 385, 385, 470, 598, 726, 897
- },
- {
- 86, 598, 1195, 1195, 2390, 2390, 2390, 2390,
- 598, 598, 1195, 1195, 2390, 2390, 2390, 2390,
- 1195, 1195, 2390, 2390, 3414, 3414, 3414, 3414,
- 1195, 1195, 2390, 2390, 3414, 3414, 3414, 3414,
- 2390, 2390, 3414, 3414, 3414, 3414, 3414, 3414,
- 2390, 2390, 3414, 3414, 3414, 3414, 3414, 3414,
- 2390, 2390, 3414, 3414, 3414, 3414, 3414, 3414,
- 2390, 2390, 3414, 3414, 3414, 3414, 3414, 3414
- }
-};
-
-static const uint16_t ivi4_quant_8x8_inter[9][64] = {
- {
- 427, 427, 470, 427, 427, 427, 470, 470,
- 427, 427, 470, 427, 427, 427, 470, 470,
- 470, 470, 470, 470, 470, 470, 470, 470,
- 427, 427, 470, 470, 427, 427, 470, 470,
- 427, 427, 470, 427, 427, 427, 470, 470,
- 427, 427, 470, 427, 427, 427, 470, 470,
- 470, 470, 470, 470, 470, 470, 470, 470,
- 470, 470, 470, 470, 470, 470, 470, 470
- },
- {
- 1707, 1707, 2433, 2433, 3414, 3414, 3414, 3414,
- 1707, 1707, 2433, 2433, 3414, 3414, 3414, 3414,
- 2433, 2433, 3414, 3414, 4822, 4822, 4822, 4822,
- 2433, 2433, 3414, 3414, 4822, 4822, 4822, 4822,
- 3414, 3414, 4822, 4822, 3414, 3414, 3414, 3414,
- 3414, 3414, 4822, 4822, 3414, 3414, 3414, 3414,
- 3414, 3414, 4822, 4822, 3414, 3414, 3414, 3414,
- 3414, 3414, 4822, 4822, 3414, 3414, 3414, 3414
- },
- {
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281,
- 1195, 1195, 1281, 1238, 1195, 1195, 1281, 1281
- },
- {
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433,
- 2433, 2433, 3414, 3414, 2433, 2433, 2433, 2433
- },
- {
- 1195, 1195, 1195, 1195, 1195, 1195, 1195, 1195,
- 1195, 1195, 1195, 1195, 1195, 1195, 1195, 1195,
- 1281, 1281, 1281, 1281, 1281, 1281, 1281, 1281,
- 1238, 1238, 1238, 1238, 1238, 1238, 1238, 1238,
- 1195, 1195, 1195, 1195, 1195, 1195, 1195, 1195,
- 1195, 1195, 1195, 1195, 1195, 1195, 1195, 1195,
- 1281, 1281, 1281, 1281, 1281, 1281, 1281, 1281,
- 1281, 1281, 1281, 1281, 1281, 1281, 1281, 1281
- },
- {
- 2433, 2433, 2433, 2433, 2433, 2433, 2433, 2433,
- 2433, 2433, 2433, 2433, 2433, 2433, 2433, 2433,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 3414, 3414, 3414, 3414, 3414, 3414, 3414, 3414,
- 2433, 2433, 2433, 2433, 2433, 2433, 2433, 2433,
- 2433, 2433, 2433, 2433, 2433, 2433, 2433, 2433,
- 2433, 2433, 2433, 2433, 2433, 2433, 2433, 2433,
- 2433, 2433, 2433, 2433, 2433, 2433, 2433, 2433
- },
- {
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707,
- 1707, 1707, 1707, 1707, 1707, 1707, 1707, 1707
- },
- {
- 86, 171, 171, 214, 214, 214, 214, 257,
- 171, 171, 214, 214, 214, 214, 257, 257,
- 171, 214, 214, 214, 214, 257, 257, 257,
- 214, 214, 214, 214, 257, 257, 257, 299,
- 214, 214, 214, 257, 257, 257, 299, 299,
- 214, 214, 257, 257, 257, 299, 299, 299,
- 214, 257, 257, 257, 299, 299, 299, 342,
- 257, 257, 257, 299, 299, 299, 342, 342
- },
- {
- 854, 854, 1195, 1195, 1707, 1707, 1707, 1707,
- 854, 854, 1195, 1195, 1707, 1707, 1707, 1707,
- 1195, 1195, 1707, 1707, 2390, 2390, 2390, 2390,
- 1195, 1195, 1707, 1707, 2390, 2390, 2390, 2390,
- 1707, 1707, 2390, 2390, 1707, 1707, 1707, 1707,
- 1707, 1707, 2390, 2390, 1707, 1707, 1707, 1707,
- 1707, 1707, 2390, 2390, 1707, 1707, 1707, 1707,
- 1707, 1707, 2390, 2390, 1707, 1707, 1707, 1707
- }
-};
-
-static const uint16_t ivi4_quant_4x4_intra[5][16] = {
- {
- 22, 214, 257, 299,
- 214, 257, 299, 342,
- 257, 299, 342, 427,
- 299, 342, 427, 513
- },
- {
- 129, 1025, 1451, 1451,
- 1025, 1025, 1451, 1451,
- 1451, 1451, 2049, 2049,
- 1451, 1451, 2049, 2049
- },
- {
- 43, 171, 171, 171,
- 43, 171, 171, 171,
- 43, 171, 171, 171,
- 43, 171, 171, 171
- },
- {
- 43, 43, 43, 43,
- 171, 171, 171, 171,
- 171, 171, 171, 171,
- 171, 171, 171, 171
- },
- {
- 43, 43, 43, 43,
- 43, 43, 43, 43,
- 43, 43, 43, 43,
- 43, 43, 43, 43
- }
-};
-
-static const uint16_t ivi4_quant_4x4_inter[5][16] = {
- {
- 107, 214, 257, 299,
- 214, 257, 299, 299,
- 257, 299, 299, 342,
- 299, 299, 342, 342
- },
- {
- 513, 1025, 1238, 1238,
- 1025, 1025, 1238, 1238,
- 1238, 1238, 1451, 1451,
- 1238, 1238, 1451, 1451
- },
- {
- 43, 171, 171, 171,
- 43, 171, 171, 171,
- 43, 171, 171, 171,
- 43, 171, 171, 171
- },
- {
- 43, 43, 43, 43,
- 171, 171, 171, 171,
- 171, 171, 171, 171,
- 171, 171, 171, 171
- },
- {
- 43, 43, 43, 43,
- 43, 43, 43, 43,
- 43, 43, 43, 43,
- 43, 43, 43, 43
- }
-};
-
-/**
- * Table for mapping quant matrix index from the bitstream
- * into internal quant table number.
- */
-static const uint8_t quant_index_to_tab[22] = {
- 0, 1, 0, 2, 1, 3, 0, 4, 1, 5, 0, 1, 6, 7, 8, // for 8x8 quant matrixes
- 0, 1, 2, 2, 3, 3, 4 // for 4x4 quant matrixes
-};
-
-#endif /* AVCODEC_INDEO4DATA_H */
diff --git a/spaces/congsaPfin/Manga-OCR/logs/Download Dummy Video The Ultimate Guide to Free 4K Stock Footage.md b/spaces/congsaPfin/Manga-OCR/logs/Download Dummy Video The Ultimate Guide to Free 4K Stock Footage.md
deleted file mode 100644
index 77e22a5ec7183d82d1bc15932ec4b03671bc5235..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/Download Dummy Video The Ultimate Guide to Free 4K Stock Footage.md
+++ /dev/null
@@ -1,194 +0,0 @@
-
-How to Download Dummy Videos for Demo Use
-Are you looking for videos of different resolution and sizes to test while designing or developing a mobile app, website, or presentation? Do you want to use some sample videos for demo purposes without violating any copyright laws or spending too much time and money? If so, you might want to consider using dummy videos.
-download dummy video
Download ❤ https://urlca.com/2uO9o5
-Dummy videos are video files that are created for testing and demonstration purposes. They can be used to fill in the gaps, show the functionality, or simulate the appearance of a real video. In this article, we will explain what dummy videos are, why you might want to use them, how to download them from the web, and how to use them for demo purposes.
- What are dummy videos and why use them?
-Definition and examples of dummy videos
-Dummy videos are video files that are not meant to be watched by the end users, but rather to serve as placeholders, samples, or mockups for testing and demonstration purposes. They can have different formats, such as MP4, AVI, MOV, WMV, FLV, 3GP, etc., and different quality options, such as 1080p, 720p, 480p, or 360p. They can also have different content, such as color bars, test patterns, countdowns, animations, logos, text, etc.
-Some examples of dummy videos are:
-
-- Sample Videos: A website that offers free stock video footage and sample HD video clips in various categories and resolutions.
-- File Examples: A website that provides sample / dummy video files for testing and demo purposes in different formats and sizes.
-- Learning Container: A website that provides sample video files for testing purposes in MP4 format with different resolutions and durations.
-
- Benefits and drawbacks of using dummy videos
-Using dummy videos can have some benefits and drawbacks depending on your needs and goals. Some of the benefits are:
-download sample 4k video footage
-download free dummy video clips
-download dummy video files for testing
-download sample video files in different formats
-download dummy video for HTML5 display
-download sample video with audio and subtitles
-download dummy video for presentation purpose
-download sample video from Pexels website
-download dummy video for upload testing
-download sample video in MOV format
-download dummy video in AVI format
-download sample video in MP4 format
-download dummy video in MKV format
-download sample video in WEBM format
-download dummy video in FLV format
-download sample video in WMV format
-download dummy video in MPEG format
-download sample video in OGG format
-download dummy video in 3GP format
-download sample video in M4V format
-download dummy video with different resolutions
-download sample video with different frame rates
-download dummy video with different bitrates
-download sample video with different aspect ratios
-download dummy video with different codecs
-download sample video with different durations
-download dummy video with different sizes
-download sample video with different colors
-download dummy video with different effects
-download sample video with different transitions
-download dummy video with different animations
-download sample video with different text overlays
-download dummy video with different audio tracks
-download sample video with different sound effects
-download dummy video with different music genres
-download sample video with different voice overs
-download dummy video with different languages
-download sample video with different subtitles formats
-download dummy video with different captions styles
-download sample video with different metadata tags
-download dummy video from File Examples website
-download sample videos for demo or test purpose
-download dummy videos for content creation or editing purpose
-download sample videos for learning or teaching purpose
-download dummy videos for research or analysis purpose
-download sample videos for entertainment or fun purpose
-
-- They can save you time and money by avoiding the need to create or purchase real videos.
-- They can help you test the functionality, compatibility, performance, or design of your app, website, or presentation.
-- They can help you showcase your work or idea without revealing any sensitive or confidential information.
-- They can help you avoid any legal issues or ethical concerns by respecting the intellectual property rights of the original video creators.
-
-Some of the drawbacks are:
-
-- They can be boring, unrealistic, or unappealing to the end users or potential customers.
-- They can create confusion or misunderstanding if they are not clearly labeled or explained as dummy videos.
-- They can limit your creativity or flexibility by restricting you to the available options or formats.
-- They can pose some technical challenges or risks by requiring additional software, tools, or skills to download or use them.
-
- How to download dummy videos from the web
-Using online video downloader tools
-Steps to use online video downloader tools
-One of the easiest ways to download dummy videos from the web is to use online video downloader tools. These are specially made websites or web apps that look for and extract videos from webpages. All you need is a web browser and an internet connection. Here are the steps:
- Go to the website or webpage that contains the dummy video you want to download.
-- Copy the URL of the webpage from the address bar of your browser.
-- Go to an online video downloader tool, such as Video Downloader, SaveFrom.net, or Y2mate.
-- Paste the URL of the webpage into the input box of the online video downloader tool.
-- Click on the download button or icon and wait for the tool to analyze the webpage and find the video.
-- Choose the format, quality, and size of the video you want to download from the available options.
-- Click on the download link or button and save the video file to your device or computer.
- Pros and cons of online video downloader tools
-Using online video downloader tools can have some pros and cons depending on your preferences and requirements. Some of the pros are:
-
-- They are free and easy to use, without requiring any installation or registration.
-- They can download videos from various websites and sources, such as YouTube, Vimeo, Facebook, Instagram, etc.
-- They can download videos in different formats and quality options, such as MP4, MP3, HD, 4K, etc.
-- They can download videos quickly and efficiently, without affecting the original quality or content of the videos.
-
-Some of the cons are:
-
-- They may not work for some websites or videos that have encryption or protection mechanisms.
-- They may not support some formats or quality options that you need or want.
-- They may have some limitations or restrictions on the number, size, or duration of the videos you can download.
-- They may have some ads, pop-ups, or malware that can affect your browsing experience or device security.
-
- Using video downloader software or apps
-Steps to use video downloader software or apps
-Another way to download dummy videos from the web is to use video downloader software or apps. These are specially made programs or applications that you can install on your device or computer and use to download videos from various websites. You will need some storage space and compatible device or computer. Here are the steps:
- Download and install a video downloader software or app, such as 4K Video Downloader, Freemake Video Downloader, or VidMate.
-- Launch the video downloader software or app and go to the settings or preferences menu to adjust the format, quality, and location of the videos you want to download.
-- Go to the website or webpage that contains the dummy video you want to download.
-- Copy the URL of the webpage from the address bar of your browser.
-- Paste the URL of the webpage into the input box of the video downloader software or app and click on the download button or icon.
-- Wait for the software or app to analyze the webpage and find the video.
-- Choose the format, quality, and size of the video you want to download from the available options.
-- Click on the download link or button and save the video file to your device or computer.
- Pros and cons of video downloader software or apps
-Using video downloader software or apps can have some pros and cons depending on your preferences and requirements. Some of the pros are:
-
-- They can download videos from various websites and sources, including some that are not supported by online video downloader tools.
-- They can download videos in different formats and quality options, including some that are not available by online video downloader tools.
-- They can download videos in bulk or batch mode, allowing you to download multiple videos at once.
-- They can download videos in the background, allowing you to continue browsing or working while downloading.
-
-Some of the cons are:
-
-- They may require some installation space and compatible device or computer to run properly.
-- They may require some updates or upgrades to keep up with the changes or improvements of the websites or videos.
-- They may have some costs or fees associated with them, either for downloading, installing, or using them.
-- They may have some ads, pop-ups, or malware that can affect your device security or performance.
-
- How to use dummy videos for demo purposes
-Tips and best practices for using dummy videos
-Once you have downloaded some dummy videos from the web, you can use them for demo purposes in various ways. However, there are some tips and best practices that you should follow to make sure that you use them effectively and appropriately. Here are some of them:
-Choose the dummy videos that match your needs and goals, such as the format, quality, size, duration, and content of the videos.
-Label or mark the dummy videos clearly as such, so that the end users or potential customers can distinguish them from the real videos.
-Explain or demonstrate the purpose or functionality of the dummy videos, such as how they will be replaced, modified, or enhanced in the final product or service.
-Use the dummy videos creatively and strategically, such as by adding some effects, transitions, captions, or voiceovers to make them more engaging or appealing.
-Test and review the dummy videos before using them for demo purposes, such as by checking their quality, compatibility, performance, or design.
-
- Examples of using dummy videos for different scenarios
-To give you some ideas of how to use dummy videos for demo purposes, here are some examples of using dummy videos for different scenarios:
-
-
-Scenario |
-Example |
-
-
-You are designing a mobile app that allows users to watch and share videos. |
-You can use some dummy videos of different resolutions and sizes to test how your app works on different devices and networks. You can also use some dummy videos of different categories and genres to show how your app organizes and displays the videos. You can also use some dummy videos of different content and quality to show how your app supports various video formats and options. |
-
-
-You are developing a website that offers online courses and tutorials. |
-You can use some dummy videos of different durations and topics to test how your website handles and streams the videos. You can also use some dummy videos of different instructors and styles to show how your website features and promotes the courses and tutorials. You can also use some dummy videos of different feedback and ratings to show how your website collects and displays the reviews and testimonials. |
-
-
-You are creating a presentation that showcases your work or idea. |
-You can use some dummy videos of different animations and transitions to test how your presentation looks and sounds. You can also use some dummy videos of different logos and text to show how your presentation introduces and summarizes your work or idea. You can also use some dummy videos of different examples and scenarios to show how your presentation illustrates and explains your work or idea. |
-
-
- Conclusion
-Dummy videos are video files that are created for testing and demonstration purposes. They can be used to fill in the gaps, show the functionality, or simulate the appearance of a real video. In this article, we have explained what dummy videos are, why you might want to use them, how to download them from the web, and how to use them for demo purposes. We hope that this article has helped you understand and appreciate the benefits and drawbacks of using dummy videos. We also hope that this article has given you some tips and best practices for using dummy videos effectively and appropriately. We also hope that this article has provided you with some examples of using dummy videos for different scenarios.
- FAQs
-What are some sources of dummy videos?
-Some sources of dummy videos are:
-Sample Videos: A website that offers free stock video footage and sample HD video clips in various categories and resolutions.
-File Examples: A website that provides sample / dummy video files for testing and demo purposes in different formats and sizes.
-Learning Container: A website that provides sample video files for testing purposes in MP4 format with different resolutions and durations.
-Dummy Video Generator: A website that allows you to create and download custom dummy videos with different options and parameters.
-YouTube: A website that hosts millions of videos on various topics and categories that you can download using online video downloader tools or software.
-
- How to convert dummy videos to different formats?
-If you need to convert dummy videos to different formats, such as from MP4 to AVI or from MOV to WMV, you can use some online or offline video converter tools or software. Some examples are:
-
-- Online Video Converter: An online tool that allows you to convert videos to different formats for free and without registration.
-- Any Video Converter: An offline software that allows you to convert videos to different formats with high quality and fast speed.
-- HandBrake: An offline software that allows you to convert videos to different formats with advanced settings and options.
-
- How to edit dummy videos for demo purposes?
-If you need to edit dummy videos for demo purposes, such as by trimming, cropping, rotating, adding effects, transitions, captions, or voiceovers, you can use some online or offline video editor tools or software. Some examples are:
-Clipchamp: An online tool that allows you to edit videos for free and without watermark.
-Filmora: An offline software that allows you to edit videos with professional features and effects.
-iMovie: An offline software that allows you to edit videos with simple and intuitive interface and tools.
-
- How to compress dummy videos for demo purposes?
-If you need to compress dummy videos for demo purposes, such as by reducing the file size, resolution, or bitrate of the videos, you can use some online or offline video compressor tools or software. Some examples are:
-
-- Online Video Compressor: An online tool that allows you to compress videos for free and without quality loss.
-- Video Compressor: An offline software that allows you to compress videos with fast speed and high quality.
-- VLC Media Player: An offline software that allows you to compress videos with various settings and options.
-
- How to embed dummy videos for demo purposes?
-If you need to embed dummy videos for demo purposes, such as by inserting the videos into your app, website, or presentation, you can use some HTML code or embed code that is provided by the source of the videos. Here are the steps:
-- Go to the website or webpage that contains the dummy video you want to embed.
-- Right-click on the video and select the option to copy the HTML code or embed code of the video.
-- Go to your app, website, or presentation and paste the HTML code or embed code where you want the video to appear.
-- Save and preview your app, website, or presentation to see how the video looks and works.
- 197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/Flight Simulator Mobile Tips and Tricks for a Smooth and Fun Flight.md b/spaces/congsaPfin/Manga-OCR/logs/Flight Simulator Mobile Tips and Tricks for a Smooth and Fun Flight.md
deleted file mode 100644
index a66565582e6f2da5f265b535a657f207c25f22bf..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/Flight Simulator Mobile Tips and Tricks for a Smooth and Fun Flight.md
+++ /dev/null
@@ -1,134 +0,0 @@
-
-Flight Simulator Mobile Games: How to Fly Anywhere in the World from Your Phone
- Have you ever dreamed of flying a plane, exploring different places, or experiencing the thrill of aviation? If so, you might be interested in trying out some flight simulator mobile games. These are apps that let you simulate various aspects of flying, such as taking off, landing, navigating, and performing maneuvers. You can choose from different aircraft, regions, weather conditions, and scenarios, and enjoy a realistic and immersive flight simulation experience on your phone.
- What are flight simulator mobile games?
- Flight simulator mobile games are games that use global satellite images, real-world navigation data, and advanced physics models to create a realistic representation of flying on your mobile device. They are different from arcade-style flying games that focus more on action, combat, or racing. Flight simulator mobile games aim to simulate the actual physics, dynamics, and challenges of flying as accurately as possible.
-flight simulator mobile
DOWNLOAD >> https://urlca.com/2uO72s
- The benefits of playing flight simulator mobile games
- Playing flight simulator mobile games can have many benefits for both beginners and experienced pilots. Some of them are:
-
-- You can learn the basics of flying, such as how to use the instruments, controls, and procedures of a plane.
-- You can practice your skills and improve your confidence in a safe and controlled environment.
-- You can explore different regions and airports around the world, and see them in high definition scenery.
-- You can experience different weather conditions and scenarios, such as day and night, rain and snow, wind and turbulence, and emergencies.
-- You can have fun and challenge yourself with various missions, challenges, and tutorials.
-- You can join a community of other pilots and air traffic controllers, and fly together in a multiplayer mode.
-
- The challenges of playing flight simulator mobile games
- Playing flight simulator mobile games can also have some challenges that you need to be aware of. Some of them are:
-
-- You need a stable internet connection to access the full features and scenery of the game.
-- You need a compatible device that can run the game smoothly and handle the graphics quality.
-- You need to invest some time and effort to learn how to play the game properly and master the controls.
-- You need to be prepared for some frustration and mistakes as you learn from your errors and improve your skills.
-
- How to choose the best flight simulator mobile game for you
- There are many flight simulator mobile games available on Android and iOS devices, but not all of them are created equal. Some are more realistic, detailed, and comprehensive than others. Some are more user-friendly, accessible, and affordable than others. How do you choose the best one for you? Here are some factors to consider:
- The features to look for in a flight simulator mobile game
- When choosing a flight simulator mobile game, you should look for these features:
-
- The best flight simulator mobile games available on Android and iOS
- Based on these features and other factors such as ratings, reviews, popularity, etc., here are some of the best flight simulator mobile games available on Android and iOS devices:
- Infinite Flight Simulator
- Infinite Flight Simulator is one of the most popular and comprehensive flight simulator mobile games on the market. It offers over 80 aircraft to choose from, including commercial jets, military fighters, cargo planes, helicopters, etc. It also offers over 20 regions and 500 airports to explore, with high-resolution satellite imagery and 3D buildings. It has a realistic flight model that simulates the physics and performance of each aircraft. It has a realistic weather system that simulates the conditions and effects of each region. It has a realistic navigation system that allows you to use the instruments and procedures of each aircraft. It has a realistic sound system that reproduces the engine noise and air traffic control of each aircraft. It has a realistic damage system that shows the effects of crashes and failures on each aircraft. It has a variety of modes to play, such as free flight,
career, missions, challenges, tutorials, etc. It has a multiplayer mode that allows you to fly with or against other players online. It has a user-friendly interface that allows you to customize the settings, options, views, controls, etc. of the game. It has a regular update system that adds new features, content, improvements, bug fixes, etc. to the game. Infinite Flight Simulator is available on Android and iOS devices for $4.99, with some in-app purchases for additional content.
-best flight simulator mobile app
-realistic flight simulator mobile game
-flight simulator mobile with global scenery
-flight simulator mobile with live weather
-flight simulator mobile with multiplayer mode
-flight simulator mobile with 3D cockpits
-flight simulator mobile with real-world navigation
-flight simulator mobile with aircraft liveries
-flight simulator mobile with emergency scenarios
-flight simulator mobile with combat system
-x-plane mobile flight simulator
-infinite flight mobile flight simulator
-geo-fs mobile flight simulator
-rfs - real flight simulator mobile
-aerofly fs 2021 mobile flight simulator
-microsoft flight simulator mobile version
-pro flight simulator mobile edition
-airbus a320 flight simulator mobile
-boeing 737 flight simulator mobile
-cessna 172 flight simulator mobile
-f-22 raptor flight simulator mobile
-a-10 thunderbolt ii flight simulator mobile
-helicopter flight simulator mobile free
-glider flight simulator mobile online
-drone flight simulator mobile download
-space shuttle flight simulator mobile apk
-fighter jet flight simulator mobile mod
-passenger plane flight simulator mobile ios
-cargo plane flight simulator mobile android
-air traffic control flight simulator mobile pc
-airport city flight simulator mobile cheats
-airline commander flight simulator mobile hack
-air navy fighters flight simulator mobile review
-air combat pilot: ww2 pacific flight simulator mobile gameplay
-air force one 3d: president plane simulators 2021 - free flying games for android & ios (mobile)
-air tycoon online 2 - realistic airline management simulation game for iphone, ipad and ipod touch (mobile)
-airplane pilot car transport sim 2020 - new airplane games for android devices (mobile)
-airplane pilot sim - realistic flying simulation game for android and ios devices (mobile)
-airplane! - fun and easy to play casual game for all ages (mobile)
-avion flight simulator ™ 2015 - the most sophisticated flight simulator game for android phones and tablets (mobile)
-extreme landings - test your piloting skills and manage the most critical flight conditions of history (mobile)
-f18 carrier landing ii - the most advanced mobile aircraft landing system ever (mobile)
-falcon bms - a community mod for the combat flightsim f4 allied force (pc)
-flywings 2018 - the ultimate simulation for your android device (mobile)
-fsx steam edition - the gold standard of modern flightsims (pc)
-il2 sturmovik: battle of stalingrad - the next generation of the legendary il2 sturmovik series of wwii combat flightsims (pc)
-just planes - the world's largest aviation video website (web)
-kerbal space program - build spacecraft, fly them, and try to help the kerbals fulfill their ultimate mission of conquering space (pc, console, web, and soon on mobile)
-lockheed martin prepar3d® v5 - the latest release in the prepar3d franchise, offering unprecedented realism and immersion in a dynamic and living world (pc)
-rc plane 3 - the third chapter in the most realistic rc plane simulator ever created (mobile)
- X-Plane Mobile
- X-Plane Mobile is another popular and comprehensive flight simulator mobile game on the market. It offers over 50 aircraft to choose from, including commercial jets, military fighters, general aviation, helicopters, etc. It also offers over 10 regions and 250 airports to explore, with high-resolution satellite imagery and 3D buildings. It has a realistic flight model that simulates the physics and performance of each aircraft. It has a realistic weather system that simulates the conditions and effects of each region. It has a realistic navigation system that allows you to use the instruments and procedures of each aircraft. It has a realistic sound system that reproduces the engine noise and air traffic control of each aircraft. It has a realistic damage system that shows the effects of crashes and failures on each aircraft. It has a variety of modes to play, such as free flight, career, missions, challenges, tutorials, etc. It has a multiplayer mode that allows you to fly with or against other players online. It has a user-friendly interface that allows you to customize the settings, options, views, controls, etc. of the game. It has a regular update system that adds new features, content, improvements, bug fixes, etc. to the game. X-Plane Mobile is available on Android and iOS devices for free, with some in-app purchases for additional content.
- Microsoft Flight Simulator
- Microsoft Flight Simulator is one of the most anticipated and advanced flight simulator mobile games on the market. It is based on the popular PC game of the same name, which is known for its stunning graphics and realism. It offers over 30 aircraft to choose from, including commercial jets, general aviation, military fighters, etc. It also offers over 40 regions and 2 million airports to explore, with high-resolution satellite imagery and 3D buildings. It has a realistic flight model that simulates the physics and performance of each aircraft. It has a realistic weather system that simulates the conditions and effects of each region. It has a realistic navigation system that allows you to use the instruments and procedures of each aircraft. It has a realistic sound system that reproduces the engine noise and air traffic control of each aircraft. It has a realistic damage system that shows the effects of crashes and failures on each aircraft. It has a variety of modes to play,
such as free flight, career, missions, challenges, tutorials, etc. It has a multiplayer mode that allows you to fly with or against other players online. It has a user-friendly interface that allows you to customize the settings, options, views, controls, etc. of the game. It has a regular update system that adds new features, content, improvements, bug fixes, etc. to the game. Microsoft Flight Simulator is available on Android and iOS devices for free, but it requires a subscription to access the full features and content of the game.
- GeoFS - Free Online Flight Simulator
- GeoFS - Free Online Flight Simulator is one of the most accessible and affordable flight simulator mobile games on the market. It offers over 20 aircraft to choose from, including commercial jets, general aviation, military fighters, helicopters, etc. It also offers over 30 regions and 30,000 airports to explore, with high-resolution satellite imagery and 3D buildings. It has a realistic flight model that simulates the physics and performance of each aircraft. It has a realistic weather system that simulates the conditions and effects of each region. It has a realistic navigation system that allows you to use the instruments and procedures of each aircraft. It has a realistic sound system that reproduces the engine noise and air traffic control of each aircraft. It has a realistic damage system that shows the effects of crashes and failures on each aircraft. It has a variety of modes to play, such as free flight, missions, challenges, tutorials, etc. It has a multiplayer mode that allows you to fly with or against other players online. It has a user-friendly interface that allows you to customize the settings, options, views, controls, etc. of the game. GeoFS - Free Online Flight Simulator is available on Android and iOS devices for free, with some in-app purchases for additional content.
- How to play flight simulator mobile games
- Now that you have chosen your preferred flight simulator mobile game, you might be wondering how to play it. Here are some general steps and tips to help you get started:
- The basic controls and commands of a flight simulator mobile game
- The basic controls and commands of a flight simulator mobile game vary depending on the game and the device you are using, but they usually involve some combination of touch screen gestures, buttons, sliders, joysticks, tilt sensors, etc. Here are some common controls and commands that you might encounter:
-
-- To control the throttle (the power of the engine), you can use a slider or a button on the screen.
-- To control the pitch (the angle of the nose up or down), you can use a joystick or tilt your device forward or backward.
-- To control the roll (the angle of the wings left or right), you can use a joystick or tilt your device left or right.
-- To control the yaw (the angle of the tail left or right), you can use a rudder pedal or a button on the screen.
-- To control the flaps (the extensions of the wings that increase lift and drag), you can use a lever or a button on the screen.
-- To control the landing gear (the wheels that retract or extend), you can use a lever or a button on the screen.
-- To control the brakes (the devices that slow down or stop the plane), you can use a pedal or a button on the screen.
-- To control the autopilot (the system that automatically flies the plane according to a preset course), you can use a switch or a button on the screen.
-- To communicate with the air traffic control (the service that provides guidance and instructions to pilots), you can use a radio or a button on the screen.
-
- The tips and tricks to improve your flying skills and enjoy the game more
- Here are some tips and tricks to improve your flying skills and enjoy the game more:
-
-- Start with easy settings and scenarios, such as clear weather, low traffic, simple aircraft, etc., and gradually increase the difficulty and complexity as you gain more experience and confidence.
-- Follow the tutorials and instructions provided by the game, as they will teach you how to use
frustrating, difficult, or stressful. They can also require a lot of time, effort, and patience to play and master. Therefore, flight simulator mobile games may not suit everyone's taste, preference, or mood. You should try them out and see for yourself if you find them fun or not.
- - Are flight simulator mobile games educational?
-Flight simulator mobile games are educational in many ways, as they can help you learn and improve various skills and knowledge related to flying and aviation. Some of them are:
-
-- You can learn the terminology, concepts, principles, and facts of flying and aviation.
-- You can learn the instruments, controls, procedures, and rules of flying and aviation.
-- You can learn the physics, aerodynamics, performance, and behavior of different aircraft.
-- You can learn the geography, climate, culture, and history of different regions and airports.
-- You can learn the communication, teamwork, and leadership skills of flying and aviation.
-- You can learn the problem-solving, decision-making, and critical-thinking skills of flying and aviation.
-
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/How to Host a Fun and Easy Download Music Quiz Round for Your Friends.md b/spaces/congsaPfin/Manga-OCR/logs/How to Host a Fun and Easy Download Music Quiz Round for Your Friends.md
deleted file mode 100644
index 81ffb2004d31d60f525c9db0b3a1612dbf443934..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/How to Host a Fun and Easy Download Music Quiz Round for Your Friends.md
+++ /dev/null
@@ -1,69 +0,0 @@
-
-How to Download Music Quiz Rounds for Your Next Party
-If you are looking for a fun and easy way to entertain your guests at your next party or gathering, why not try a music quiz round? A music quiz round is a set of questions or challenges
- Quizshark: A website that offers music rounds in various categories, such as animals, body parts, flying, girls names, etc. You can download 10 clips of songs for each round and an answer sheet in a ZIP file. The website also offers other types of quiz rounds, such as movies, TV, sports, etc.
-- Storyblocks: A website that offers royalty-free music and sound effects for various purposes, including quiz shows. You can download music clips that match your quiz theme, such as weather, golf, travel, etc. The website also offers video clips and images that you can use for your quiz.
-- AhaSlides: A website that offers interactive quizzes and presentations that you can play online with your friends. You can choose from a library of pop music quiz questions with images and audio clips, or create your own quiz using their templates. The website also allows you to customize the design and settings of your quiz.
-
- Conclusion
-Downloading music quiz rounds is a great way to spice up your next party or gathering. You can find a wide range of music quiz rounds online that will suit your taste and challenge your guests. Downloading music quiz rounds is easy and convenient, as long as you use a trusted and quality website. So what are you waiting for? Download some music quiz rounds today and have a blast with your friends!
-download music quiz round
Download File ✒ https://urlca.com/2uOcrM
- FAQs
-How much does it cost to download music quiz rounds?
-It depends on the website you use. Some websites offer free music quiz rounds, while others charge a fee or require a subscription. You should always check the terms and conditions before downloading any content from a website.
- How long does it take to download music quiz rounds?
-It depends on the size of the file and your internet speed. Generally, it should not take more than a few minutes to download a music quiz round. However, if you are downloading multiple files or large files, it may take longer.
- How many people can play a music quiz round?
-It depends on how you organize your quiz. You can play a music quiz round individually or in teams. You can also decide how many points to award for each correct answer and how to break ties. The more people you have, the more fun it will be.
-Download quiz royalty-free music and sound effects
-Download music rounds audio in MP3 format
-Download pop music quiz questions with answers
-Download Disney music quiz image round
-Download music quiz themes from TV shows
-Download skiffle music quiz clips
-Download summer songs music quiz round
-Download songs with colours in the title music quiz
-Download songs with numbers in the title music quiz
-Download songs with lady in the title or artist music quiz
-Download musicals songs music quiz round
-Download love songs music quiz round
-Download water cannon songs music quiz round
-Download flying songs music quiz round
-Download body parts songs music quiz round
-Download animals songs music quiz round
-Download US TV themes music quiz round
-Download la la land songs music quiz round
-Download golf terms songs music quiz round
-Download Christmas number two's music quiz round
-Download snippets of famous songs music quiz round
-Download girls names songs music quiz round
-Download bodies of water songs music quiz round
-Download weather songs music quiz round
-Download one to ten songs music quiz round
-Download heart songs music quiz round
-Download trivia questions for music quiz rounds
-Download game show timer for music quiz rounds
-Download feedback sounds for music quiz rounds
-Download mind game or quiz tension background for music quiz rounds
-Download success roll up achieve sound for music quiz rounds
-Download sudden death sound for music quiz rounds
-Download doing it right funk rock sports sound for music quiz rounds
-Download feedback correct right answer sound for music quiz rounds
-Download feedback correct 3 sound for music quiz rounds
-Download Hawaii 50 theme for music quiz rounds
-Download game of thrones theme for music quiz rounds
-Download jungle book theme for music quiz rounds
-Download frozen theme for music quiz rounds
-Download fiddler on the roof theme for music quiz rounds
- How do I play a music quiz round?
-It depends on the format of the quiz round. If you have downloaded audio clips and answer sheets, you can play them on your computer or speaker system and have your guests write down their answers on paper or on their phones. If you have downloaded an interactive quiz from AhaSlides, you can play it online using their platform and have your guests join using their phones.
- What are some tips for creating my own music quiz round?
-Some tips for creating your own music quiz round are:
-
-- Choose a theme or genre that you and your guests enjoy and are familiar with
-- Use songs that are popular, catchy, and recognizable
-- Vary the difficulty level and style of the questions to keep it interesting and fair
-- Use high-quality audio clips and images to make it more engaging and appealing
-
- : https://quizshark.co.uk/music-rounds/ : https://www.storyblocks.com/audio/search/quiz : https://ahaslides.com/pop-music-quiz/ 197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/RPGMaker - The Easiest Way to Create Your Own RPG - Free Trial and Demo.md b/spaces/congsaPfin/Manga-OCR/logs/RPGMaker - The Easiest Way to Create Your Own RPG - Free Trial and Demo.md
deleted file mode 100644
index 3b1b31a131ad48e80acd0c1a98a4909c30a29fd1..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/RPGMaker - The Easiest Way to Create Your Own RPG - Free Trial and Demo.md
+++ /dev/null
@@ -1,127 +0,0 @@
-
-RPG Maker Free Download: How to Get Started with Your Own RPG
-If you are a fan of role-playing games (RPGs) and want to create your own, you may have heard of RPG Maker. RPG Maker is a game engine that allows you to make 2D RPGs with ease. You can use pre-made assets, drag and drop events, and customize your game with a simple scripting language. RPG Maker has been around since 1992 and has many versions and editions for different platforms.
-However, RPG Maker is not a free software. You have to pay a license fee to use it legally. The latest version, RPG Maker MZ, costs $79.99 on Steam. That may be too expensive for some hobbyists or beginners who just want to try it out. So, how can you download RPG Maker for free? Is it even possible? And what are the alternatives if you can't or don't want to use RPG Maker?
-rpgmaker free download
DOWNLOAD 🗹 https://urlca.com/2uO99K
-In this article, we will explore three options for downloading RPG Maker for free and compare their pros and cons. We will also answer some frequently asked questions about RPG Maker and its alternatives. By the end of this article, you should have a better idea of how to get started with your own RPG project.
- Option 1: Download RPG Maker from the official website with a free trial
-The first option is to download RPG Maker from the official website https://www.rpgmakerweb.com/downloads. Here, you can find all the versions of RPG Maker available for Windows and Mac. You can also find additional downloads such as sample games, tools, plugins, and materials.
-The good news is that you can download any version of RPG Maker for free as a trial. The trial allows you to use all the features of the software for 30 days. You can create and test your own games during this period. However, you cannot distribute or sell your games unless you buy a license.
-The pros of this option are:
-
-- You can access the latest version of RPG Maker with all the updates and bug fixes.
-- You can get official support and resources from the website and the community.
-- You can test the software before buying it to see if it suits your needs and preferences.
-
-The cons of this option are:
-
-- You have to register an account on the website to download the trial.
-- You have a limited time of 30 days to use the software.
-- You may not be able to open or edit older projects made with previous versions of RPG Maker.
-
- Option 2: Download RPG Maker from a third-party website or torrent
-The second option is to download RPG Maker from a third-party website or torrent. There are many websites that offer free downloads of RPG Maker or cracked versions that bypass the license verification. You can also find torrents that let you download RPG Maker from other users who have already downloaded it.
-The pros of this option are:
-
-- You don't have to pay anything or register an account to download the software.
-- You don't have a time limit to use the software.
-- You may have access to older versions of RPG licenses.
-- You can have more features and flexibility to create different types of games or customize your game engine.
-- You may have better performance and compatibility with different platforms and devices.
-
-The cons of this option are:
-rpgmaker xp free download
-rpgmaker vx ace free download
-rpgmaker mv free download
-rpgmaker mz free download
-rpgmaker 2000 free download
-rpgmaker 2003 free download
-rpgmaker vx free download
-rpgmaker trial free download
-rpgmaker resources free download
-rpgmaker plugins free download
-rpgmaker games free download
-rpgmaker music free download
-rpgmaker sprites free download
-rpgmaker tilesets free download
-rpgmaker characters free download
-rpgmaker scripts free download
-rpgmaker icons free download
-rpgmaker facesets free download
-rpgmaker battlers free download
-rpgmaker animations free download
-rpgmaker backgrounds free download
-rpgmaker sound effects free download
-rpgmaker fonts free download
-rpgmaker windowskins free download
-rpgmaker portraits free download
-open rpg maker free download
-easyrpg maker free download
-simrpg maker free download
-action game maker free download
-indie game maker free download
-pixel game maker free download
-visual novel maker free download
-renpy game maker free download
-godot game maker free download
-unity game maker free download
-unreal engine game maker free download
-gamemaker studio 2 free download
-construct 3 game maker free download
-gdevelop game maker free download
-stencyl game maker free download
-fusion 2.5 game maker free download
-clickteam game maker free download
-adventure game studio free download
-quest game maker free download
-twine game maker free download
-inform 7 game maker free download
-tads game maker free download
-inklewriter game maker free download
-choicescript game maker free download
-
-- You may have a steeper learning curve to master these game engines, as they may have different interfaces, workflows, and languages.
-- You may find less user-friendly or intuitive tools and options, as these game engines may not be designed specifically for 2D RPGs.
-- You may need to have some coding skills or additional tools to create or edit your game assets, events, or logic.
-
- Conclusion: How to choose the best option for your RPG project
-As you can see, there are three options for downloading RPG Maker for free, each with its own advantages and disadvantages. The best option for you depends on your needs and goals. Here are some tips on how to choose the best option for your RPG project:
-
-- If you want to use the official and latest version of RPG Maker with full support and resources, you should download it from the official website with a free trial. However, you should be aware of the time limit and the compatibility issues.
-- If you want to use RPG Maker without paying or registering, you can download it from a third-party website or torrent. However, you should be careful of the risks and the legal issues involved.
-- If you want to use a free and open source game engine that is similar to RPG Maker, you can download one of the alternatives we mentioned. However, you should be prepared to learn and adapt to a different game engine.
-
-Whatever option you choose, we hope that this article has helped you get started with your own RPG project. We wish you good luck and happy gaming!
- FAQs: Frequently asked questions about RPG Maker and its alternatives
-What is the difference between RPG Maker MZ and RPG Maker MV?
-RPG Maker MZ is the latest version of RPG Maker, released in 2020. It is an improved version of RPG Maker MV, which was released in 2015. Some of the differences between RPG Maker MZ and RPG Maker MV are:
-
-- RPG Maker MZ has a faster and more stable engine than RPG Maker MV.
-- RPG Maker MZ has more features and options than RPG Maker MV, such as layer mapping, character generator, event list, plugin manager, and more.
-- RPG Maker MZ has more assets and resources than RPG Maker MV, such as graphics, sounds, music, animations, and more.
-
- Can I use RPG Maker on my phone or tablet?
-No, you cannot use RPG Maker on your phone or tablet. RPG Maker is a desktop software that requires a Windows or Mac computer to run. However, you can play games made with RPG Maker on your phone or tablet if they are exported as HTML5 or Android/iOS apps. You can also use remote desktop apps to access your computer from your phone or tablet and run RPG Maker.
- Can I sell games made with RPG Maker?
-Yes, you can sell games made with RPG Maker if you have a valid license for the software. You can sell your games on any platform or store that allows it, such as Steam, itch.io, Google Play, App Store, etc. However, you have to follow some rules and guidelines when selling your games, such as:
-
-- You have to credit RPG Maker and any other tools or resources you used in your game.
-- You have to own or have permission to use any assets or plugins that are not included in RPG Maker.
-- You have to respect the intellectual property rights of any other games or works that may inspire or influence your game.
-
- What are some examples of games made with RPG Maker?
-There are many games made with RPG Maker that have been successful or popular among gamers. Some examples are:
-
-- To The Moon: A story-driven game that follows two doctors who try to fulfill the dying wish of a man by altering his memories.
-- Lisa: A dark comedy game that explores the themes of survival, sacrifice, and morality in a post-apocalyptic world.
-- Corpse Party: A horror game that follows a group of students who are trapped in a haunted school after performing a ritual.
-
- What are some other game engines that I can use to create 2D games?
-Besides the ones we mentioned in this article, there are many other game engines that you can use to create 2D games. Some examples are:
-
-- Unity: A powerful and popular game engine that supports 2D and 3D graphics, scripting, animation, physics, audio, and more. It has a large and active community and a rich asset store.
-- GameMaker Studio: A game engine that is designed for 2D games and allows you to create games with drag and drop or coding. It has a simple and intuitive interface and a built-in editor.
-- Construct: A game engine that lets you create 2D games without coding. It has a visual event system and a web-based editor.
-
401be4b1e0
-
-
\ No newline at end of file
diff --git a/spaces/congsaPfin/Manga-OCR/logs/Summertime Saga 20.8 APK Download - Experience the Hottest Adventure.md b/spaces/congsaPfin/Manga-OCR/logs/Summertime Saga 20.8 APK Download - Experience the Hottest Adventure.md
deleted file mode 100644
index 4fb8f039e1cf9dbcdd030691d75cb3d75233abad..0000000000000000000000000000000000000000
--- a/spaces/congsaPfin/Manga-OCR/logs/Summertime Saga 20.8 APK Download - Experience the Hottest Adventure.md
+++ /dev/null
@@ -1,109 +0,0 @@
-
-How to Download Summertime Saga 20.8 APK for Android
-If you are looking for a fun and spicy visual novel game, you should definitely check out Summertime Saga. This game is not your typical romance story. It is full of humor, drama, mystery, and adventure. You will play as a young student who has to deal with the death of his father, the debt to the mafia, and the challenges of high school life. Along the way, you will meet many interesting characters, explore different locations, and make choices that will affect the outcome of your story.
-download summertime saga 20.8 apk
Download Zip ⇔ https://urlca.com/2uOa1R
-In this article, we will show you how to download Summertime Saga 20.8 APK for Android devices. This is the latest version of the game, which has many new features and improvements. We will also give you some tips and tricks on how to play the game and enjoy it to the fullest.
-Step 1: Enable unknown sources on your device
-Before you can install Summertime Saga 20.8 APK on your Android device, you need to enable unknown sources. This is a security setting that allows you to install apps from sources other than the Google Play Store. To do this, follow these steps:
-
-- Go to your device's settings and tap on security or privacy.
-- Find the option that says unknown sources or install unknown apps and toggle it on.
-- You may see a warning message that says installing apps from unknown sources can harm your device. Tap on OK or allow to proceed.
-
-Now you are ready to download the APK file.
-Step 2: Download the APK file from a trusted source
-The next step is to download the Summertime Saga 20.8 APK file from a trusted source. There are many websites that offer APK files, but not all of them are safe and reliable. Some of them may contain viruses, malware, or unwanted ads that can harm your device or compromise your privacy.
-One of the best sources to download Summertime Saga 20.8 APK is Uptodown. This is a popular platform that provides free and secure downloads of apps and games for Android devices. To download the APK file from Uptodown, follow these steps:
-
-- Open your browser and go to https://summertime-saga.en.uptodown.com/android.
-- Tap on the green download button that says latest version.
-- You may see a pop-up window that asks you to choose a mirror site. Tap on any of them to start the download.
-- Wait for the download to finish. You can check the progress in your notification bar or in your downloads folder.
-
-Once you have downloaded the APK file, you can proceed to install it.
-Step 3: Install the APK file and launch the game
-The final step is to install the Summertime Saga 20.8 APK file and launch the game. To do this, follow these steps:
-
-- Locate the APK file in your downloads folder or in your notification bar.
-- Tap on the file to open it.
-- You may see a pop-up window that asks you for permission to install the app. Tap on install or next to continue the installation.
-- Wait for the installation to complete. You may see a pop-up window that says app installed. Tap on open or done to launch the game or exit the installer.
-
-Congratulations! You have successfully installed Summertime Saga 20.8 APK on your Android device. You can now enjoy the game and its features.
-Features of Summertime Saga 20.8 APK
-Summertime Saga 20.8 APK is the latest version of the game, which was released on June 1, 2023. This version has many new features and improvements that make the game more fun and exciting. Here are some of the features of Summertime Saga 20.8 APK:
-How to download summertime saga 20.8 apk for free
-Summertime saga 20.8 apk mod download with unlimited money
-Download summertime saga 20.8 apk latest version for android
-Summertime saga 20.8 apk walkthrough and guide
-Download summertime saga 20.8 apk offline installer
-Summertime saga 20.8 apk cheats and tips
-Download summertime saga 20.8 apk from uptodown[^1^]
-Summertime saga 20.8 apk review and rating
-Download summertime saga 20.8 apk full game
-Summertime saga 20.8 apk new features and updates
-Download summertime saga 20.8 apk for PC
-Summertime saga 20.8 apk best characters and routes
-Download summertime saga 20.8 apk from apkpure
-Summertime saga 20.8 apk system requirements and compatibility
-Download summertime saga 20.8 apk from mediafire
-Summertime saga 20.8 apk secrets and easter eggs
-Download summertime saga 20.8 apk for iOS
-Summertime saga 20.8 apk fan art and wallpapers
-Download summertime saga 20.8 apk from google drive
-Summertime saga 20.8 apk bugs and fixes
-Download summertime saga 20.8 apk for mac
-Summertime saga 20.8 apk alternatives and similar games
-Download summertime saga 20.8 apk from mega.nz
-Summertime saga 20.8 apk trivia and facts
-Download summertime saga 20.8 apk from softonic
-Summertime saga 20.8 apk developer and publisher
-Download summertime saga 20.8 apk from zippyshare
-Summertime saga 20.8 apk news and updates
-Download summertime saga 20.8 apk from dropbox
-Summertime saga 20.8 apk forum and community
-Improved graphics and animations
-The game has improved its graphics and animations, making it more realistic and immersive. The characters, backgrounds, and scenes are more detailed and colorful. The animations are smoother and more expressive. The game also supports higher resolutions and aspect ratios, making it compatible with different devices and screens.
-New characters and locations
-The game has added new characters and locations to the story, giving you more options and possibilities. You can meet new people, make new friends, or find new lovers. You can also explore new places, such as the beach, the park, the mall, or the casino. Each character and location has its own story, secrets, and events that you can discover and experience.
-More choices and outcomes
-The game has increased its choices and outcomes, making it more interactive and dynamic. You can make different decisions that will affect your relationships, your reputation, your money, your stats, and your endings. You can also choose different paths and routes that will lead you to different scenarios and situations. The game has multiple endings, depending on your actions and consequences.
-Bug fixes and optimizations
-The game has fixed some bugs and glitches that were present in the previous versions, making it more stable and smooth. The game has also optimized its performance and compatibility, making it faster and easier to play. The game has reduced its loading times, improved its memory usage, and enhanced its security.
-Tips and Tricks for Playing Summertime Saga
-Summertime Saga is a fun and addictive game, but it can also be challenging and complex. There are many things to do, many people to meet, many secrets to uncover, and many endings to achieve. To help you enjoy the game and get the most out of it, here are some tips and tricks for playing Summertime Saga:
-How to save and load your progress
-The game allows you to save and load your progress at any time, so you don't have to worry about losing your data or starting over. To save your progress, tap on the menu icon at the top right corner of the screen and select save. You can choose from 10 different slots to save your game. To load your progress, tap on the menu icon again and select load. You can choose from any of the slots that have a saved game.
-How to earn money and stats
-The game requires you to earn money and stats to advance in the story and unlock different options and features. Money is used to buy items, gifts, services, or information. Stats are used to improve your skills, abilities, or attributes. To earn money and stats, you can do various activities, such as working at different jobs, studying at school or library, training at gym or dojo, playing games at arcade or casino, or completing quests or tasks.
-How to unlock different routes and endings
-The game has different routes and endings for each character that you interact with. Each route has its own story line, events, choices, outcomes and endings. To unlock different routes and endings, you need to follow these steps:
-
-- Choose a character that you want to pursue and focus on them. You can check their profile and status in the menu.
-- Interact with them whenever you see them or when they call you. You can also call them or visit them at their location.
-- Give them gifts or compliments that they like. You can find out their preferences by talking to them or by buying information from Eve or Larry.
-- Make choices that will increase your relationship level with them. You can check your relationship level in the menu.
-- Complete their quests or tasks that they give you or ask you to do. You can check your quests or tasks in the menu.
-- Once you reach a certain relationship level, you will trigger a special event or scene that will advance your route and lead you to an ending.
-
-How to avoid common pitfalls and mistakes
-The game has some pitfalls and mistakes that can hinder your progress or ruin your experience. To avoid these pitfalls and mistakes, you should be aware of these tips:
-
-- Don't skip the dialogue or the cutscenes. They contain important information, clues, hints, or instructions that you need to follow.
-- Don't ignore the phone calls or messages. They can give you updates, reminders, invitations, or opportunities that you don't want to miss.
-- Don't waste your time or money on things that are not relevant or useful. You have limited resources and you need to use them wisely.
-- Don't cheat or use hacks. They can cause errors, glitches, crashes, or bans that can ruin your game.
-
-Conclusion and FAQs
-Summertime Saga is a great game for anyone who loves visual novels, romance, comedy, and adventure. It has a captivating story, engaging characters, stunning graphics, and multiple features. It is easy to download and install on your Android device, as long as you follow the steps we have provided in this article. We hope you have fun playing Summertime Saga and exploring its different routes and endings.
-If you have any questions or doubts about the game, you can check out these FAQs:
-
-Q: How big is the Summertime Saga 20.8 APK file? | A: The Summertime Saga 20.8 APK file is about 1 GB in size. |
-Q: Is Summertime Saga 20.8 APK safe and virus-free? | A: Yes, Summertime Saga 20.8 APK is safe and virus-free, as long as you download it from a trusted source like Uptodown. |
-Q: Is Summertime Saga 20.8 APK free and legal? | A: Yes, Summertime Saga 20.8 APK is free and legal, as it is developed by an independent team of creators who do not charge any money for their work. |
-Q: How often is Summertime Saga updated? | A: Summertime Saga is updated every few months, depending on the progress and feedback of the developers and the community. |
-Q: Where can I find more information or support for Summertime Saga? | A: You can find more information or support for Summertime Saga on their official website, their Patreon page, their Discord server, or their Reddit forum. |
-
- : https://summertime-saga.en.uptodown.com/android : https://summertimesaga.com/ : https://www.patreon.com/summertimesaga : https://discord.gg/summertimesaga : https://www.reddit.com/r/SummertimeSaga/ 197e85843d
-
-
\ No newline at end of file
diff --git a/spaces/contluForse/HuggingGPT/assets/ 52 .md b/spaces/contluForse/HuggingGPT/assets/ 52 .md
deleted file mode 100644
index 614c3d073ce6bccd73e6cbb5ed218e5dce4a3f8a..0000000000000000000000000000000000000000
--- a/spaces/contluForse/HuggingGPT/assets/ 52 .md
+++ /dev/null
@@ -1,6 +0,0 @@
-Краткое Содержание Великолепный Век 52 Серия
Download Zip ✔ https://ssurll.com/2uzxa8
-
- aaccfb2cb3
-
-
-
diff --git a/spaces/coreml-community/ControlNet-v1-1-Annotators-cpu/annotator/mmpkg/mmcv/utils/testing.py b/spaces/coreml-community/ControlNet-v1-1-Annotators-cpu/annotator/mmpkg/mmcv/utils/testing.py
deleted file mode 100644
index a27f936da8ec14bac18562ede0a79d476d82f797..0000000000000000000000000000000000000000
--- a/spaces/coreml-community/ControlNet-v1-1-Annotators-cpu/annotator/mmpkg/mmcv/utils/testing.py
+++ /dev/null
@@ -1,140 +0,0 @@
-# Copyright (c) Open-MMLab.
-import sys
-from collections.abc import Iterable
-from runpy import run_path
-from shlex import split
-from typing import Any, Dict, List
-from unittest.mock import patch
-
-
-def check_python_script(cmd):
- """Run the python cmd script with `__main__`. The difference between
- `os.system` is that, this function exectues code in the current process, so
- that it can be tracked by coverage tools. Currently it supports two forms:
-
- - ./tests/data/scripts/hello.py zz
- - python tests/data/scripts/hello.py zz
- """
- args = split(cmd)
- if args[0] == 'python':
- args = args[1:]
- with patch.object(sys, 'argv', args):
- run_path(args[0], run_name='__main__')
-
-
-def _any(judge_result):
- """Since built-in ``any`` works only when the element of iterable is not
- iterable, implement the function."""
- if not isinstance(judge_result, Iterable):
- return judge_result
-
- try:
- for element in judge_result:
- if _any(element):
- return True
- except TypeError:
- # Maybe encounter the case: torch.tensor(True) | torch.tensor(False)
- if judge_result:
- return True
- return False
-
-
-def assert_dict_contains_subset(dict_obj: Dict[Any, Any],
- expected_subset: Dict[Any, Any]) -> bool:
- """Check if the dict_obj contains the expected_subset.
-
- Args:
- dict_obj (Dict[Any, Any]): Dict object to be checked.
- expected_subset (Dict[Any, Any]): Subset expected to be contained in
- dict_obj.
-
- Returns:
- bool: Whether the dict_obj contains the expected_subset.
- """
-
- for key, value in expected_subset.items():
- if key not in dict_obj.keys() or _any(dict_obj[key] != value):
- return False
- return True
-
-
-def assert_attrs_equal(obj: Any, expected_attrs: Dict[str, Any]) -> bool:
- """Check if attribute of class object is correct.
-
- Args:
- obj (object): Class object to be checked.
- expected_attrs (Dict[str, Any]): Dict of the expected attrs.
-
- Returns:
- bool: Whether the attribute of class object is correct.
- """
- for attr, value in expected_attrs.items():
- if not hasattr(obj, attr) or _any(getattr(obj, attr) != value):
- return False
- return True
-
-
-def assert_dict_has_keys(obj: Dict[str, Any],
- expected_keys: List[str]) -> bool:
- """Check if the obj has all the expected_keys.
-
- Args:
- obj (Dict[str, Any]): Object to be checked.
- expected_keys (List[str]): Keys expected to contained in the keys of
- the obj.
-
- Returns:
- bool: Whether the obj has the expected keys.
- """
- return set(expected_keys).issubset(set(obj.keys()))
-
-
-def assert_keys_equal(result_keys: List[str], target_keys: List[str]) -> bool:
- """Check if target_keys is equal to result_keys.
-
- Args:
- result_keys (List[str]): Result keys to be checked.
- target_keys (List[str]): Target keys to be checked.
-
- Returns:
- bool: Whether target_keys is equal to result_keys.
- """
- return set(result_keys) == set(target_keys)
-
-
-def assert_is_norm_layer(module) -> bool:
- """Check if the module is a norm layer.
-
- Args:
- module (nn.Module): The module to be checked.
-
- Returns:
- bool: Whether the module is a norm layer.
- """
- from .parrots_wrapper import _BatchNorm, _InstanceNorm
- from torch.nn import GroupNorm, LayerNorm
- norm_layer_candidates = (_BatchNorm, _InstanceNorm, GroupNorm, LayerNorm)
- return isinstance(module, norm_layer_candidates)
-
-
-def assert_params_all_zeros(module) -> bool:
- """Check if the parameters of the module is all zeros.
-
- Args:
- module (nn.Module): The module to be checked.
-
- Returns:
- bool: Whether the parameters of the module is all zeros.
- """
- weight_data = module.weight.data
- is_weight_zero = weight_data.allclose(
- weight_data.new_zeros(weight_data.size()))
-
- if hasattr(module, 'bias') and module.bias is not None:
- bias_data = module.bias.data
- is_bias_zero = bias_data.allclose(
- bias_data.new_zeros(bias_data.size()))
- else:
- is_bias_zero = True
-
- return is_weight_zero and is_bias_zero
diff --git a/spaces/dafqi/indo_twitter_sentiment_app/README.md b/spaces/dafqi/indo_twitter_sentiment_app/README.md
deleted file mode 100644
index c536ad4053375d0ef366986e23a24f6324e518ca..0000000000000000000000000000000000000000
--- a/spaces/dafqi/indo_twitter_sentiment_app/README.md
+++ /dev/null
@@ -1,17 +0,0 @@
----
-title: Indo Twitter Sentiment App
-emoji: 👀
-colorFrom: green
-colorTo: yellow
-sdk: streamlit
-sdk_version: 1.15.2
-app_file: app.py
-pinned: false
----
-
-# twitter sentiment app
-
-Aplikasi sederhana untuk melakukan analisis sentimen terhadap tweet yang diinputkan dan mengekstrak topik dari setiap sentimen
-
-link website : https://dafiqrahman-twitter-sentiment-app-app-shcgk3.streamlit.app/
-
diff --git a/spaces/dakaiye/dky_xuexi/crazy_functions/test_project/latex/attention/model_architecture.tex b/spaces/dakaiye/dky_xuexi/crazy_functions/test_project/latex/attention/model_architecture.tex
deleted file mode 100644
index c82be6242cc9d26203360e90d3ac9184ef6ad842..0000000000000000000000000000000000000000
--- a/spaces/dakaiye/dky_xuexi/crazy_functions/test_project/latex/attention/model_architecture.tex
+++ /dev/null
@@ -1,155 +0,0 @@
-
-\begin{figure}
- \centering
- \includegraphics[scale=0.6]{Figures/ModalNet-21}
- \caption{The Transformer - model architecture.}
- \label{fig:model-arch}
-\end{figure}
-
-% Although the primary workhorse of our model is attention,
-%Our model maintains the encoder-decoder structure that is common to many so-called sequence-to-sequence models \citep{bahdanau2014neural,sutskever14}. As in all such architectures, the encoder computes a representation of the input sequence, and the decoder consumes these representations along with the output tokens to autoregressively produce the output sequence. Where, traditionally, the encoder and decoder contain stacks of recurrent or convolutional layers, our encoder and decoder stacks are composed of attention layers and position-wise feed-forward layers (Figure~\ref{fig:model-arch}). The following sections describe the gross architecture and these particular components in detail.
-
-Most competitive neural sequence transduction models have an encoder-decoder structure \citep{cho2014learning,bahdanau2014neural,sutskever14}. Here, the encoder maps an input sequence of symbol representations $(x_1, ..., x_n)$ to a sequence of continuous representations $\mathbf{z} = (z_1, ..., z_n)$. Given $\mathbf{z}$, the decoder then generates an output sequence $(y_1,...,y_m)$ of symbols one element at a time. At each step the model is auto-regressive \citep{graves2013generating}, consuming the previously generated symbols as additional input when generating the next.
-
-The Transformer follows this overall architecture using stacked self-attention and point-wise, fully connected layers for both the encoder and decoder, shown in the left and right halves of Figure~\ref{fig:model-arch}, respectively.
-
-\subsection{Encoder and Decoder Stacks}
-
-\paragraph{Encoder:}The encoder is composed of a stack of $N=6$ identical layers. Each layer has two sub-layers. The first is a multi-head self-attention mechanism, and the second is a simple, position-wise fully connected feed-forward network. We employ a residual connection \citep{he2016deep} around each of the two sub-layers, followed by layer normalization \cite{layernorm2016}. That is, the output of each sub-layer is $\mathrm{LayerNorm}(x + \mathrm{Sublayer}(x))$, where $\mathrm{Sublayer}(x)$ is the function implemented by the sub-layer itself. To facilitate these residual connections, all sub-layers in the model, as well as the embedding layers, produce outputs of dimension $\dmodel=512$.
-
-\paragraph{Decoder:}The decoder is also composed of a stack of $N=6$ identical layers. In addition to the two sub-layers in each encoder layer, the decoder inserts a third sub-layer, which performs multi-head attention over the output of the encoder stack. Similar to the encoder, we employ residual connections around each of the sub-layers, followed by layer normalization. We also modify the self-attention sub-layer in the decoder stack to prevent positions from attending to subsequent positions. This masking, combined with fact that the output embeddings are offset by one position, ensures that the predictions for position $i$ can depend only on the known outputs at positions less than $i$.
-
-% In our model (Figure~\ref{fig:model-arch}), the encoder and decoder are composed of stacks of alternating self-attention layers (for cross-positional communication) and position-wise feed-forward layers (for in-place computation). In addition, the decoder stack contains encoder-decoder attention layers. Since attention is agnostic to the distances between words, our model requires a "positional encoding" to be added to the encoder and decoder input. The following sections describe all of these components in detail.
-
-\subsection{Attention} \label{sec:attention}
-An attention function can be described as mapping a query and a set of key-value pairs to an output, where the query, keys, values, and output are all vectors. The output is computed as a weighted sum of the values, where the weight assigned to each value is computed by a compatibility function of the query with the corresponding key.
-
-\subsubsection{Scaled Dot-Product Attention} \label{sec:scaled-dot-prod}
-
-% \begin{figure}
-% \centering
-% \includegraphics[scale=0.6]{Figures/ModalNet-19}
-% \caption{Scaled Dot-Product Attention.}
-% \label{fig:multi-head-att}
-% \end{figure}
-
-We call our particular attention "Scaled Dot-Product Attention" (Figure~\ref{fig:multi-head-att}). The input consists of queries and keys of dimension $d_k$, and values of dimension $d_v$. We compute the dot products of the query with all keys, divide each by $\sqrt{d_k}$, and apply a softmax function to obtain the weights on the values.
-
-In practice, we compute the attention function on a set of queries simultaneously, packed together into a matrix $Q$. The keys and values are also packed together into matrices $K$ and $V$. We compute the matrix of outputs as:
-
-\begin{equation}
- \mathrm{Attention}(Q, K, V) = \mathrm{softmax}(\frac{QK^T}{\sqrt{d_k}})V
-\end{equation}
-
-The two most commonly used attention functions are additive attention \citep{bahdanau2014neural}, and dot-product (multiplicative) attention. Dot-product attention is identical to our algorithm, except for the scaling factor of $\frac{1}{\sqrt{d_k}}$. Additive attention computes the compatibility function using a feed-forward network with a single hidden layer. While the two are similar in theoretical complexity, dot-product attention is much faster and more space-efficient in practice, since it can be implemented using highly optimized matrix multiplication code.
-
-%We scale the dot products by $1/\sqrt{d_k}$ to limit the magnitude of the dot products, which works well in practice. Otherwise, we found applying the softmax to often result in weights very close to 0 or 1, and hence minuscule gradients.
-
-% Already described in the subsequent section
-%When used as part of decoder self-attention, an optional mask function is applied just before the softmax to prevent positions from attending to subsequent positions. This mask simply sets the logits corresponding to all illegal connections (those outside of the lower triangle) to $-\infty$.
-
-%\paragraph{Comparison to Additive Attention: } We choose dot product attention over additive attention \citep{bahdanau2014neural} since it can be computed using highly optimized matrix multiplication code. This optimization is particularly important to us, as we employ many attention layers in our model.
-
-While for small values of $d_k$ the two mechanisms perform similarly, additive attention outperforms dot product attention without scaling for larger values of $d_k$ \citep{DBLP:journals/corr/BritzGLL17}. We suspect that for large values of $d_k$, the dot products grow large in magnitude, pushing the softmax function into regions where it has extremely small gradients \footnote{To illustrate why the dot products get large, assume that the components of $q$ and $k$ are independent random variables with mean $0$ and variance $1$. Then their dot product, $q \cdot k = \sum_{i=1}^{d_k} q_ik_i$, has mean $0$ and variance $d_k$.}. To counteract this effect, we scale the dot products by $\frac{1}{\sqrt{d_k}}$.
-
-
-%We suspect this to be caused by the dot products growing too large in magnitude to result in useful gradients after applying the softmax function. To counteract this, we scale the dot product by $1/\sqrt{d_k}$.
-
-
-\subsubsection{Multi-Head Attention} \label{sec:multihead}
-
-\begin{figure}
-\begin{minipage}[t]{0.5\textwidth}
- \centering
- Scaled Dot-Product Attention \\
- \vspace{0.5cm}
- \includegraphics[scale=0.6]{Figures/ModalNet-19}
-\end{minipage}
-\begin{minipage}[t]{0.5\textwidth}
- \centering
- Multi-Head Attention \\
- \vspace{0.1cm}
- \includegraphics[scale=0.6]{Figures/ModalNet-20}
-\end{minipage}
-
-
- % \centering
-
- \caption{(left) Scaled Dot-Product Attention. (right) Multi-Head Attention consists of several attention layers running in parallel.}
- \label{fig:multi-head-att}
-\end{figure}
-
-Instead of performing a single attention function with $\dmodel$-dimensional keys, values and queries, we found it beneficial to linearly project the queries, keys and values $h$ times with different, learned linear projections to $d_k$, $d_k$ and $d_v$ dimensions, respectively.
-On each of these projected versions of queries, keys and values we then perform the attention function in parallel, yielding $d_v$-dimensional output values. These are concatenated and once again projected, resulting in the final values, as depicted in Figure~\ref{fig:multi-head-att}.
-
-Multi-head attention allows the model to jointly attend to information from different representation subspaces at different positions. With a single attention head, averaging inhibits this.
-
-\begin{align*}
- \mathrm{MultiHead}(Q, K, V) &= \mathrm{Concat}(\mathrm{head_1}, ..., \mathrm{head_h})W^O\\
-% \mathrm{where} \mathrm{head_i} &= \mathrm{Attention}(QW_Q_i^{\dmodel \times d_q}, KW_K_i^{\dmodel \times d_k}, VW^V_i^{\dmodel \times d_v})\\
- \text{where}~\mathrm{head_i} &= \mathrm{Attention}(QW^Q_i, KW^K_i, VW^V_i)\\
-\end{align*}
-
-Where the projections are parameter matrices $W^Q_i \in \mathbb{R}^{\dmodel \times d_k}$, $W^K_i \in \mathbb{R}^{\dmodel \times d_k}$, $W^V_i \in \mathbb{R}^{\dmodel \times d_v}$ and $W^O \in \mathbb{R}^{hd_v \times \dmodel}$.
-
-
-%find it better (and no more expensive) to have multiple parallel attention layers (each over the full set of positions) with proportionally lower-dimensional keys, values and queries. We call this "Multi-Head Attention" (Figure~\ref{fig:multi-head-att}). The keys, values, and queries for each of these parallel attention layers are computed by learned linear transformations of the inputs to the multi-head attention. We use different linear transformations across different parallel attention layers. The output of the parallel attention layers are concatenated, and then passed through a final learned linear transformation.
-
-In this work we employ $h=8$ parallel attention layers, or heads. For each of these we use $d_k=d_v=\dmodel/h=64$.
-Due to the reduced dimension of each head, the total computational cost is similar to that of single-head attention with full dimensionality.
-
-\subsubsection{Applications of Attention in our Model}
-
-The Transformer uses multi-head attention in three different ways:
-\begin{itemize}
- \item In "encoder-decoder attention" layers, the queries come from the previous decoder layer, and the memory keys and values come from the output of the encoder. This allows every position in the decoder to attend over all positions in the input sequence. This mimics the typical encoder-decoder attention mechanisms in sequence-to-sequence models such as \citep{wu2016google, bahdanau2014neural,JonasFaceNet2017}.
-
- \item The encoder contains self-attention layers. In a self-attention layer all of the keys, values and queries come from the same place, in this case, the output of the previous layer in the encoder. Each position in the encoder can attend to all positions in the previous layer of the encoder.
-
- \item Similarly, self-attention layers in the decoder allow each position in the decoder to attend to all positions in the decoder up to and including that position. We need to prevent leftward information flow in the decoder to preserve the auto-regressive property. We implement this inside of scaled dot-product attention by masking out (setting to $-\infty$) all values in the input of the softmax which correspond to illegal connections. See Figure~\ref{fig:multi-head-att}.
-
-\end{itemize}
-
-\subsection{Position-wise Feed-Forward Networks}\label{sec:ffn}
-
-In addition to attention sub-layers, each of the layers in our encoder and decoder contains a fully connected feed-forward network, which is applied to each position separately and identically. This consists of two linear transformations with a ReLU activation in between.
-
-\begin{equation}
- \mathrm{FFN}(x)=\max(0, xW_1 + b_1) W_2 + b_2
-\end{equation}
-
-While the linear transformations are the same across different positions, they use different parameters from layer to layer. Another way of describing this is as two convolutions with kernel size 1. The dimensionality of input and output is $\dmodel=512$, and the inner-layer has dimensionality $d_{ff}=2048$.
-
-
-
-%In the appendix, we describe how the position-wise feed-forward network can also be seen as a form of attention.
-
-%from Jakob: The number of operations required for the model to relate signals from two arbitrary input or output positions grows in the distance between positions in input or output, linearly for ConvS2S and logarithmically for ByteNet, making it harder to learn dependencies between these positions \citep{hochreiter2001gradient}. In the transformer this is reduced to a constant number of operations, albeit at the cost of effective resolution caused by averaging attention-weighted positions, an effect we aim to counteract with multi-headed attention.
-
-
-%Figure~\ref{fig:simple-att} presents a simple attention function, $A$, with a single head, that forms the basis of our multi-head attention. $A$ takes a query key vector $\kq$, matrices of memory keys $\km$ and memory values $\vm$ ,and produces a query value vector $\vq$ as
-%\begin{equation*} \label{eq:attention}
-% A(\kq, \km, \vm) = {\vm}^T (Softmax(\km \kq).
-%\end{equation*}
-%We linearly transform $\kq,\,\km$, and $\vm$ with learned matrices ${\Wkq \text{,} \, \Wkm}$, and ${\Wvm}$ before calling the attention function, and transform the output query with $\Wvq$ before handing it to the feed forward layer. Each attention layer has it's own set of transformation matrices, which are shared across all query positions. $A$ is applied in parallel for each query position, and is implemented very efficiently as a batch of matrix multiplies. The self-attention and encoder-decoder attention layers use $A$, but with different arguments. For example, in encdoder self-attention, queries in encoder layer $i$ attention to memories in encoder layer $i-1$. To ensure that decoder self-attention layers do not look at future words, we add $- \inf$ to the softmax logits in positions $j+1$ to query length for query position $l$.
-
-%In simple attention, the query value is a weighted combination of the memory values where the attention weights sum to one. Although this function performs well in practice, the constraint on attention weights can restrict the amount of information that flows from memories to queries because the query cannot focus on multiple memory positions at once, which might be desirable when translating long sequences. \marginpar{@usz, could you think of an example of this ?} We remedy this by maintaining multiple attention heads at each query position that attend to all memory positions in parallel, with a different set of parameters per attention head $h$.
-%\marginpar{}
-
-\subsection{Embeddings and Softmax}
-Similarly to other sequence transduction models, we use learned embeddings to convert the input tokens and output tokens to vectors of dimension $\dmodel$. We also use the usual learned linear transformation and softmax function to convert the decoder output to predicted next-token probabilities. In our model, we share the same weight matrix between the two embedding layers and the pre-softmax linear transformation, similar to \citep{press2016using}. In the embedding layers, we multiply those weights by $\sqrt{\dmodel}$.
-
-
-\subsection{Positional Encoding}
-Since our model contains no recurrence and no convolution, in order for the model to make use of the order of the sequence, we must inject some information about the relative or absolute position of the tokens in the sequence. To this end, we add "positional encodings" to the input embeddings at the bottoms of the encoder and decoder stacks. The positional encodings have the same dimension $\dmodel$ as the embeddings, so that the two can be summed. There are many choices of positional encodings, learned and fixed \citep{JonasFaceNet2017}.
-
-In this work, we use sine and cosine functions of different frequencies:
-
-\begin{align*}
- PE_{(pos,2i)} = sin(pos / 10000^{2i/\dmodel}) \\
- PE_{(pos,2i+1)} = cos(pos / 10000^{2i/\dmodel})
-\end{align*}
-
-where $pos$ is the position and $i$ is the dimension. That is, each dimension of the positional encoding corresponds to a sinusoid. The wavelengths form a geometric progression from $2\pi$ to $10000 \cdot 2\pi$. We chose this function because we hypothesized it would allow the model to easily learn to attend by relative positions, since for any fixed offset $k$, $PE_{pos+k}$ can be represented as a linear function of $PE_{pos}$.
-
-We also experimented with using learned positional embeddings \citep{JonasFaceNet2017} instead, and found that the two versions produced nearly identical results (see Table~\ref{tab:variations} row (E)). We chose the sinusoidal version because it may allow the model to extrapolate to sequence lengths longer than the ones encountered during training.
diff --git a/spaces/datagpt/pdf2gpt/README.md b/spaces/datagpt/pdf2gpt/README.md
deleted file mode 100644
index 96a4148db542d0e6b92f3f08ba34f555efc9662b..0000000000000000000000000000000000000000
--- a/spaces/datagpt/pdf2gpt/README.md
+++ /dev/null
@@ -1,13 +0,0 @@
----
-title: Pdf2gpt
-emoji: 👁
-colorFrom: yellow
-colorTo: pink
-sdk: gradio
-sdk_version: 3.27.0
-app_file: app.py
-pinned: false
-license: mit
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/daveward/smaragd-hentaidiffusion/README.md b/spaces/daveward/smaragd-hentaidiffusion/README.md
deleted file mode 100644
index 1a7f55c84074fa4dc04dad484c852ca90b1dadd6..0000000000000000000000000000000000000000
--- a/spaces/daveward/smaragd-hentaidiffusion/README.md
+++ /dev/null
@@ -1,12 +0,0 @@
----
-title: Smaragd Hentaidiffusion
-emoji: 🐠
-colorFrom: red
-colorTo: red
-sdk: gradio
-sdk_version: 3.12.0
-app_file: app.py
-pinned: false
----
-
-Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
diff --git a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/altair/vegalite/v5/schema/channels.py b/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/altair/vegalite/v5/schema/channels.py
deleted file mode 100644
index 07f9f43e8e1387a374e60ae99ee9a92e1549d1e1..0000000000000000000000000000000000000000
--- a/spaces/dcarpintero/nlp-summarizer-pegasus/.venv/lib/python3.9/site-packages/altair/vegalite/v5/schema/channels.py
+++ /dev/null
@@ -1,17317 +0,0 @@
-# The contents of this file are automatically written by
-# tools/generate_schema_wrapper.py. Do not modify directly.
-
-import sys
-from . import core
-import pandas as pd
-from altair.utils.schemapi import Undefined, with_property_setters
-from altair.utils import parse_shorthand
-from typing import overload, List
-
-if sys.version_info >= (3, 8):
- from typing import Literal
-else:
- from typing_extensions import Literal
-
-
-class FieldChannelMixin:
- def to_dict(self, validate=True, ignore=(), context=None):
- context = context or {}
- shorthand = self._get('shorthand')
- field = self._get('field')
-
- if shorthand is not Undefined and field is not Undefined:
- raise ValueError("{} specifies both shorthand={} and field={}. "
- "".format(self.__class__.__name__, shorthand, field))
-
- if isinstance(shorthand, (tuple, list)):
- # If given a list of shorthands, then transform it to a list of classes
- kwds = self._kwds.copy()
- kwds.pop('shorthand')
- return [self.__class__(sh, **kwds).to_dict(validate=validate, ignore=ignore, context=context)
- for sh in shorthand]
-
- if shorthand is Undefined:
- parsed = {}
- elif isinstance(shorthand, str):
- parsed = parse_shorthand(shorthand, data=context.get('data', None))
- type_required = 'type' in self._kwds
- type_in_shorthand = 'type' in parsed
- type_defined_explicitly = self._get('type') is not Undefined
- if not type_required:
- # Secondary field names don't require a type argument in VegaLite 3+.
- # We still parse it out of the shorthand, but drop it here.
- parsed.pop('type', None)
- elif not (type_in_shorthand or type_defined_explicitly):
- if isinstance(context.get('data', None), pd.DataFrame):
- raise ValueError(
- 'Unable to determine data type for the field "{}";'
- " verify that the field name is not misspelled."
- " If you are referencing a field from a transform,"
- " also confirm that the data type is specified correctly.".format(shorthand)
- )
- else:
- raise ValueError("{} encoding field is specified without a type; "
- "the type cannot be automatically inferred because "
- "the data is not specified as a pandas.DataFrame."
- "".format(shorthand))
- else:
- # Shorthand is not a string; we pass the definition to field,
- # and do not do any parsing.
- parsed = {'field': shorthand}
- context["parsed_shorthand"] = parsed
-
- return super(FieldChannelMixin, self).to_dict(
- validate=validate,
- ignore=ignore,
- context=context
- )
-
-
-class ValueChannelMixin:
- def to_dict(self, validate=True, ignore=(), context=None):
- context = context or {}
- condition = self._get('condition', Undefined)
- copy = self # don't copy unless we need to
- if condition is not Undefined:
- if isinstance(condition, core.SchemaBase):
- pass
- elif 'field' in condition and 'type' not in condition:
- kwds = parse_shorthand(condition['field'], context.get('data', None))
- copy = self.copy(deep=['condition'])
- copy['condition'].update(kwds)
- return super(ValueChannelMixin, copy).to_dict(validate=validate,
- ignore=ignore,
- context=context)
-
-
-class DatumChannelMixin:
- def to_dict(self, validate=True, ignore=(), context=None):
- context = context or {}
- datum = self._get('datum', Undefined)
- copy = self # don't copy unless we need to
- if datum is not Undefined:
- if isinstance(datum, core.SchemaBase):
- pass
- return super(DatumChannelMixin, copy).to_dict(validate=validate,
- ignore=ignore,
- context=context)
-
-
-@with_property_setters
-class Angle(FieldChannelMixin, core.FieldOrDatumDefWithConditionMarkPropFieldDefnumber):
- """Angle schema wrapper
-
- Mapping(required=[shorthand])
-
- Parameters
- ----------
-
- shorthand : string
- shorthand for field, aggregate, and type
- aggregate : :class:`Aggregate`
- Aggregation function for the field (e.g., ``"mean"``, ``"sum"``, ``"median"``,
- ``"min"``, ``"max"``, ``"count"`` ).
-
- **Default value:** ``undefined`` (None)
-
- **See also:** `aggregate `__
- documentation.
- bandPosition : float
- Relative position on a band of a stacked, binned, time unit, or band scale. For
- example, the marks will be positioned at the beginning of the band if set to ``0``,
- and at the middle of the band if set to ``0.5``.
- bin : anyOf(boolean, :class:`BinParams`, None)
- A flag for binning a ``quantitative`` field, `an object defining binning parameters
- `__, or indicating
- that the data for ``x`` or ``y`` channel are binned before they are imported into
- Vega-Lite ( ``"binned"`` ).
-
-
- If ``true``, default `binning parameters
- `__ will be applied.
-
- If ``"binned"``, this indicates that the data for the ``x`` (or ``y`` ) channel are
- already binned. You can map the bin-start field to ``x`` (or ``y`` ) and the bin-end
- field to ``x2`` (or ``y2`` ). The scale and axis will be formatted similar to
- binning in Vega-Lite. To adjust the axis ticks based on the bin step, you can also
- set the axis's `tickMinStep
- `__ property.
-
- **Default value:** ``false``
-
- **See also:** `bin `__
- documentation.
- condition : anyOf(:class:`ConditionalValueDefnumberExprRef`, List(:class:`ConditionalValueDefnumberExprRef`))
- One or more value definition(s) with `a parameter or a test predicate
- `__.
-
- **Note:** A field definition's ``condition`` property can only contain `conditional
- value definitions `__
- since Vega-Lite only allows at most one encoded field per encoding channel.
- field : :class:`Field`
- **Required.** A string defining the name of the field from which to pull a data
- value or an object defining iterated values from the `repeat
- `__ operator.
-
- **See also:** `field `__
- documentation.
-
- **Notes:** 1) Dots ( ``.`` ) and brackets ( ``[`` and ``]`` ) can be used to access
- nested objects (e.g., ``"field": "foo.bar"`` and ``"field": "foo['bar']"`` ). If
- field names contain dots or brackets but are not nested, you can use ``\\`` to
- escape dots and brackets (e.g., ``"a\\.b"`` and ``"a\\[0\\]"`` ). See more details
- about escaping in the `field documentation
- `__. 2) ``field`` is not required
- if ``aggregate`` is ``count``.
- legend : anyOf(:class:`Legend`, None)
- An object defining properties of the legend. If ``null``, the legend for the
- encoding channel will be removed.
-
- **Default value:** If undefined, default `legend properties
- `__ are applied.
-
- **See also:** `legend `__
- documentation.
- scale : anyOf(:class:`Scale`, None)
- An object defining properties of the channel's scale, which is the function that
- transforms values in the data domain (numbers, dates, strings, etc) to visual values
- (pixels, colors, sizes) of the encoding channels.
-
- If ``null``, the scale will be `disabled and the data value will be directly encoded
- `__.
-
- **Default value:** If undefined, default `scale properties
- `__ are applied.
-
- **See also:** `scale `__
- documentation.
- sort : :class:`Sort`
- Sort order for the encoded field.
-
- For continuous fields (quantitative or temporal), ``sort`` can be either
- ``"ascending"`` or ``"descending"``.
-
- For discrete fields, ``sort`` can be one of the following:
-
-
- * ``"ascending"`` or ``"descending"`` -- for sorting by the values' natural order in
- JavaScript.
- * `A string indicating an encoding channel name to sort by
- `__ (e.g.,
- ``"x"`` or ``"y"`` ) with an optional minus prefix for descending sort (e.g.,
- ``"-x"`` to sort by x-field, descending). This channel string is short-form of `a
- sort-by-encoding definition
- `__. For
- example, ``"sort": "-x"`` is equivalent to ``"sort": {"encoding": "x", "order":
- "descending"}``.
- * `A sort field definition
- `__ for sorting by
- another field.
- * `An array specifying the field values in preferred order
- `__. In this case, the
- sort order will obey the values in the array, followed by any unspecified values
- in their original order. For discrete time field, values in the sort array can be
- `date-time definition objects
- `__. In addition, for time
- units ``"month"`` and ``"day"``, the values can be the month or day names (case
- insensitive) or their 3-letter initials (e.g., ``"Mon"``, ``"Tue"`` ).
- * ``null`` indicating no sort.
-
- **Default value:** ``"ascending"``
-
- **Note:** ``null`` and sorting by another channel is not supported for ``row`` and
- ``column``.
-
- **See also:** `sort `__
- documentation.
- timeUnit : anyOf(:class:`TimeUnit`, :class:`TimeUnitParams`)
- Time unit (e.g., ``year``, ``yearmonth``, ``month``, ``hours`` ) for a temporal
- field. or `a temporal field that gets casted as ordinal
- `__.
-
- **Default value:** ``undefined`` (None)
-
- **See also:** `timeUnit `__
- documentation.
- title : anyOf(:class:`Text`, None)
- A title for the field. If ``null``, the title will be removed.
-
- **Default value:** derived from the field's name and transformation function (
- ``aggregate``, ``bin`` and ``timeUnit`` ). If the field has an aggregate function,
- the function is displayed as part of the title (e.g., ``"Sum of Profit"`` ). If the
- field is binned or has a time unit applied, the applied function is shown in
- parentheses (e.g., ``"Profit (binned)"``, ``"Transaction Date (year-month)"`` ).
- Otherwise, the title is simply the field name.
-
- **Notes** :
-
- 1) You can customize the default field title format by providing the `fieldTitle
- `__ property in
- the `config `__ or `fieldTitle
- function via the compile function's options
- `__.
-
- 2) If both field definition's ``title`` and axis, header, or legend ``title`` are
- defined, axis/header/legend title will be used.
- type : :class:`StandardType`
- The type of measurement ( ``"quantitative"``, ``"temporal"``, ``"ordinal"``, or
- ``"nominal"`` ) for the encoded field or constant value ( ``datum`` ). It can also
- be a ``"geojson"`` type for encoding `'geoshape'
- `__.
-
- Vega-Lite automatically infers data types in many cases as discussed below. However,
- type is required for a field if: (1) the field is not nominal and the field encoding
- has no specified ``aggregate`` (except ``argmin`` and ``argmax`` ), ``bin``, scale
- type, custom ``sort`` order, nor ``timeUnit`` or (2) if you wish to use an ordinal
- scale for a field with ``bin`` or ``timeUnit``.
-
- **Default value:**
-
- 1) For a data ``field``, ``"nominal"`` is the default data type unless the field
- encoding has ``aggregate``, ``channel``, ``bin``, scale type, ``sort``, or
- ``timeUnit`` that satisfies the following criteria:
-
-
- * ``"quantitative"`` is the default type if (1) the encoded field contains ``bin``
- or ``aggregate`` except ``"argmin"`` and ``"argmax"``, (2) the encoding channel is
- ``latitude`` or ``longitude`` channel or (3) if the specified scale type is `a
- quantitative scale `__.
- * ``"temporal"`` is the default type if (1) the encoded field contains ``timeUnit``
- or (2) the specified scale type is a time or utc scale
- * ``"ordinal"`` is the default type if (1) the encoded field contains a `custom sort
- order
- `__,
- (2) the specified scale type is an ordinal/point/band scale, or (3) the encoding
- channel is ``order``.
-
- 2) For a constant value in data domain ( ``datum`` ):
-
-
- * ``"quantitative"`` if the datum is a number
- * ``"nominal"`` if the datum is a string
- * ``"temporal"`` if the datum is `a date time object
- `__
-
- **Note:**
-
-
- * Data ``type`` describes the semantics of the data rather than the primitive data
- types (number, string, etc.). The same primitive data type can have different
- types of measurement. For example, numeric data can represent quantitative,
- ordinal, or nominal data.
- * Data values for a temporal field can be either a date-time string (e.g.,
- ``"2015-03-07 12:32:17"``, ``"17:01"``, ``"2015-03-16"``. ``"2015"`` ) or a
- timestamp number (e.g., ``1552199579097`` ).
- * When using with `bin `__, the
- ``type`` property can be either ``"quantitative"`` (for using a linear bin scale)
- or `"ordinal" (for using an ordinal bin scale)
- `__.
- * When using with `timeUnit
- `__, the ``type`` property
- can be either ``"temporal"`` (default, for using a temporal scale) or `"ordinal"
- (for using an ordinal scale)
- `__.
- * When using with `aggregate
- `__, the ``type`` property
- refers to the post-aggregation data type. For example, we can calculate count
- ``distinct`` of a categorical field ``"cat"`` using ``{"aggregate": "distinct",
- "field": "cat"}``. The ``"type"`` of the aggregate output is ``"quantitative"``.
- * Secondary channels (e.g., ``x2``, ``y2``, ``xError``, ``yError`` ) do not have
- ``type`` as they must have exactly the same type as their primary channels (e.g.,
- ``x``, ``y`` ).
-
- **See also:** `type `__
- documentation.
- """
- _class_is_valid_at_instantiation = False
- _encoding_name = "angle"
-
- @overload # type: ignore[no-overload-impl]
- def aggregate(self, _: Literal["average", "count", "distinct", "max", "mean", "median", "min", "missing", "product", "q1", "q3", "ci0", "ci1", "stderr", "stdev", "stdevp", "sum", "valid", "values", "variance", "variancep"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def aggregate(self, argmax=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def aggregate(self, argmin=Undefined, **kwds) -> 'Angle':
- ...
-
- def bandPosition(self, _: float, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def bin(self, _: bool, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def bin(self, anchor=Undefined, base=Undefined, binned=Undefined, divide=Undefined, extent=Undefined, maxbins=Undefined, minstep=Undefined, nice=Undefined, step=Undefined, steps=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def bin(self, _: None, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, test=Undefined, value=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, empty=Undefined, param=Undefined, value=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, _: List[core.ConditionalValueDefnumberExprRef], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def field(self, _: str, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def field(self, repeat=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def legend(self, aria=Undefined, clipHeight=Undefined, columnPadding=Undefined, columns=Undefined, cornerRadius=Undefined, description=Undefined, direction=Undefined, fillColor=Undefined, format=Undefined, formatType=Undefined, gradientLength=Undefined, gradientOpacity=Undefined, gradientStrokeColor=Undefined, gradientStrokeWidth=Undefined, gradientThickness=Undefined, gridAlign=Undefined, labelAlign=Undefined, labelBaseline=Undefined, labelColor=Undefined, labelExpr=Undefined, labelFont=Undefined, labelFontSize=Undefined, labelFontStyle=Undefined, labelFontWeight=Undefined, labelLimit=Undefined, labelOffset=Undefined, labelOpacity=Undefined, labelOverlap=Undefined, labelPadding=Undefined, labelSeparation=Undefined, legendX=Undefined, legendY=Undefined, offset=Undefined, orient=Undefined, padding=Undefined, rowPadding=Undefined, strokeColor=Undefined, symbolDash=Undefined, symbolDashOffset=Undefined, symbolFillColor=Undefined, symbolLimit=Undefined, symbolOffset=Undefined, symbolOpacity=Undefined, symbolSize=Undefined, symbolStrokeColor=Undefined, symbolStrokeWidth=Undefined, symbolType=Undefined, tickCount=Undefined, tickMinStep=Undefined, title=Undefined, titleAlign=Undefined, titleAnchor=Undefined, titleBaseline=Undefined, titleColor=Undefined, titleFont=Undefined, titleFontSize=Undefined, titleFontStyle=Undefined, titleFontWeight=Undefined, titleLimit=Undefined, titleLineHeight=Undefined, titleOpacity=Undefined, titleOrient=Undefined, titlePadding=Undefined, type=Undefined, values=Undefined, zindex=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def legend(self, _: None, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def scale(self, align=Undefined, base=Undefined, bins=Undefined, clamp=Undefined, constant=Undefined, domain=Undefined, domainMax=Undefined, domainMid=Undefined, domainMin=Undefined, exponent=Undefined, interpolate=Undefined, nice=Undefined, padding=Undefined, paddingInner=Undefined, paddingOuter=Undefined, range=Undefined, rangeMax=Undefined, rangeMin=Undefined, reverse=Undefined, round=Undefined, scheme=Undefined, type=Undefined, zero=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def scale(self, _: None, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: List[float], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: List[str], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: List[bool], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: List[core.DateTime], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: Literal["ascending", "descending"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: Literal["x", "y", "color", "fill", "stroke", "strokeWidth", "size", "shape", "fillOpacity", "strokeOpacity", "opacity", "text"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: Literal["-x", "-y", "-color", "-fill", "-stroke", "-strokeWidth", "-size", "-shape", "-fillOpacity", "-strokeOpacity", "-opacity", "-text"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, field=Undefined, op=Undefined, order=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, encoding=Undefined, order=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def sort(self, _: None, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def timeUnit(self, _: Literal["year", "quarter", "month", "week", "day", "dayofyear", "date", "hours", "minutes", "seconds", "milliseconds"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def timeUnit(self, _: Literal["utcyear", "utcquarter", "utcmonth", "utcweek", "utcday", "utcdayofyear", "utcdate", "utchours", "utcminutes", "utcseconds", "utcmilliseconds"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def timeUnit(self, _: Literal["yearquarter", "yearquartermonth", "yearmonth", "yearmonthdate", "yearmonthdatehours", "yearmonthdatehoursminutes", "yearmonthdatehoursminutesseconds", "yearweek", "yearweekday", "yearweekdayhours", "yearweekdayhoursminutes", "yearweekdayhoursminutesseconds", "yeardayofyear", "quartermonth", "monthdate", "monthdatehours", "monthdatehoursminutes", "monthdatehoursminutesseconds", "weekday", "weeksdayhours", "weekdayhoursminutes", "weekdayhoursminutesseconds", "dayhours", "dayhoursminutes", "dayhoursminutesseconds", "hoursminutes", "hoursminutesseconds", "minutesseconds", "secondsmilliseconds"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def timeUnit(self, _: Literal["utcyearquarter", "utcyearquartermonth", "utcyearmonth", "utcyearmonthdate", "utcyearmonthdatehours", "utcyearmonthdatehoursminutes", "utcyearmonthdatehoursminutesseconds", "utcyearweek", "utcyearweekday", "utcyearweekdayhours", "utcyearweekdayhoursminutes", "utcyearweekdayhoursminutesseconds", "utcyeardayofyear", "utcquartermonth", "utcmonthdate", "utcmonthdatehours", "utcmonthdatehoursminutes", "utcmonthdatehoursminutesseconds", "utcweekday", "utcweeksdayhours", "utcweekdayhoursminutes", "utcweekdayhoursminutesseconds", "utcdayhours", "utcdayhoursminutes", "utcdayhoursminutesseconds", "utchoursminutes", "utchoursminutesseconds", "utcminutesseconds", "utcsecondsmilliseconds"], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def timeUnit(self, maxbins=Undefined, step=Undefined, unit=Undefined, utc=Undefined, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def title(self, _: str, **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def title(self, _: List[str], **kwds) -> 'Angle':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def title(self, _: None, **kwds) -> 'Angle':
- ...
-
- def type(self, _: Literal["quantitative", "ordinal", "temporal", "nominal"], **kwds) -> 'Angle':
- ...
-
-
- def __init__(self, shorthand=Undefined, aggregate=Undefined, bandPosition=Undefined, bin=Undefined,
- condition=Undefined, field=Undefined, legend=Undefined, scale=Undefined,
- sort=Undefined, timeUnit=Undefined, title=Undefined, type=Undefined, **kwds):
- super(Angle, self).__init__(shorthand=shorthand, aggregate=aggregate, bandPosition=bandPosition,
- bin=bin, condition=condition, field=field, legend=legend,
- scale=scale, sort=sort, timeUnit=timeUnit, title=title, type=type,
- **kwds)
-
-
-@with_property_setters
-class AngleDatum(DatumChannelMixin, core.FieldOrDatumDefWithConditionDatumDefnumber):
- """AngleDatum schema wrapper
-
- Mapping(required=[])
-
- Parameters
- ----------
-
- bandPosition : float
- Relative position on a band of a stacked, binned, time unit, or band scale. For
- example, the marks will be positioned at the beginning of the band if set to ``0``,
- and at the middle of the band if set to ``0.5``.
- condition : anyOf(:class:`ConditionalValueDefnumberExprRef`, List(:class:`ConditionalValueDefnumberExprRef`))
- One or more value definition(s) with `a parameter or a test predicate
- `__.
-
- **Note:** A field definition's ``condition`` property can only contain `conditional
- value definitions `__
- since Vega-Lite only allows at most one encoded field per encoding channel.
- datum : anyOf(:class:`PrimitiveValue`, :class:`DateTime`, :class:`ExprRef`, :class:`RepeatRef`)
- A constant value in data domain.
- title : anyOf(:class:`Text`, None)
- A title for the field. If ``null``, the title will be removed.
-
- **Default value:** derived from the field's name and transformation function (
- ``aggregate``, ``bin`` and ``timeUnit`` ). If the field has an aggregate function,
- the function is displayed as part of the title (e.g., ``"Sum of Profit"`` ). If the
- field is binned or has a time unit applied, the applied function is shown in
- parentheses (e.g., ``"Profit (binned)"``, ``"Transaction Date (year-month)"`` ).
- Otherwise, the title is simply the field name.
-
- **Notes** :
-
- 1) You can customize the default field title format by providing the `fieldTitle
- `__ property in
- the `config `__ or `fieldTitle
- function via the compile function's options
- `__.
-
- 2) If both field definition's ``title`` and axis, header, or legend ``title`` are
- defined, axis/header/legend title will be used.
- type : :class:`Type`
- The type of measurement ( ``"quantitative"``, ``"temporal"``, ``"ordinal"``, or
- ``"nominal"`` ) for the encoded field or constant value ( ``datum`` ). It can also
- be a ``"geojson"`` type for encoding `'geoshape'
- `__.
-
- Vega-Lite automatically infers data types in many cases as discussed below. However,
- type is required for a field if: (1) the field is not nominal and the field encoding
- has no specified ``aggregate`` (except ``argmin`` and ``argmax`` ), ``bin``, scale
- type, custom ``sort`` order, nor ``timeUnit`` or (2) if you wish to use an ordinal
- scale for a field with ``bin`` or ``timeUnit``.
-
- **Default value:**
-
- 1) For a data ``field``, ``"nominal"`` is the default data type unless the field
- encoding has ``aggregate``, ``channel``, ``bin``, scale type, ``sort``, or
- ``timeUnit`` that satisfies the following criteria:
-
-
- * ``"quantitative"`` is the default type if (1) the encoded field contains ``bin``
- or ``aggregate`` except ``"argmin"`` and ``"argmax"``, (2) the encoding channel is
- ``latitude`` or ``longitude`` channel or (3) if the specified scale type is `a
- quantitative scale `__.
- * ``"temporal"`` is the default type if (1) the encoded field contains ``timeUnit``
- or (2) the specified scale type is a time or utc scale
- * ``"ordinal"`` is the default type if (1) the encoded field contains a `custom sort
- order
- `__,
- (2) the specified scale type is an ordinal/point/band scale, or (3) the encoding
- channel is ``order``.
-
- 2) For a constant value in data domain ( ``datum`` ):
-
-
- * ``"quantitative"`` if the datum is a number
- * ``"nominal"`` if the datum is a string
- * ``"temporal"`` if the datum is `a date time object
- `__
-
- **Note:**
-
-
- * Data ``type`` describes the semantics of the data rather than the primitive data
- types (number, string, etc.). The same primitive data type can have different
- types of measurement. For example, numeric data can represent quantitative,
- ordinal, or nominal data.
- * Data values for a temporal field can be either a date-time string (e.g.,
- ``"2015-03-07 12:32:17"``, ``"17:01"``, ``"2015-03-16"``. ``"2015"`` ) or a
- timestamp number (e.g., ``1552199579097`` ).
- * When using with `bin `__, the
- ``type`` property can be either ``"quantitative"`` (for using a linear bin scale)
- or `"ordinal" (for using an ordinal bin scale)
- `__.
- * When using with `timeUnit
- `__, the ``type`` property
- can be either ``"temporal"`` (default, for using a temporal scale) or `"ordinal"
- (for using an ordinal scale)
- `__.
- * When using with `aggregate
- `__, the ``type`` property
- refers to the post-aggregation data type. For example, we can calculate count
- ``distinct`` of a categorical field ``"cat"`` using ``{"aggregate": "distinct",
- "field": "cat"}``. The ``"type"`` of the aggregate output is ``"quantitative"``.
- * Secondary channels (e.g., ``x2``, ``y2``, ``xError``, ``yError`` ) do not have
- ``type`` as they must have exactly the same type as their primary channels (e.g.,
- ``x``, ``y`` ).
-
- **See also:** `type `__
- documentation.
- """
- _class_is_valid_at_instantiation = False
- _encoding_name = "angle"
-
- def bandPosition(self, _: float, **kwds) -> 'AngleDatum':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, test=Undefined, value=Undefined, **kwds) -> 'AngleDatum':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, empty=Undefined, param=Undefined, value=Undefined, **kwds) -> 'AngleDatum':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, _: List[core.ConditionalValueDefnumberExprRef], **kwds) -> 'AngleDatum':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def title(self, _: str, **kwds) -> 'AngleDatum':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def title(self, _: List[str], **kwds) -> 'AngleDatum':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def title(self, _: None, **kwds) -> 'AngleDatum':
- ...
-
- def type(self, _: Literal["quantitative", "ordinal", "temporal", "nominal", "geojson"], **kwds) -> 'AngleDatum':
- ...
-
-
- def __init__(self, datum, bandPosition=Undefined, condition=Undefined, title=Undefined,
- type=Undefined, **kwds):
- super(AngleDatum, self).__init__(datum=datum, bandPosition=bandPosition, condition=condition,
- title=title, type=type, **kwds)
-
-
-@with_property_setters
-class AngleValue(ValueChannelMixin, core.ValueDefWithConditionMarkPropFieldOrDatumDefnumber):
- """AngleValue schema wrapper
-
- Mapping(required=[])
-
- Parameters
- ----------
-
- condition : anyOf(:class:`ConditionalMarkPropFieldOrDatumDef`, :class:`ConditionalValueDefnumberExprRef`, List(:class:`ConditionalValueDefnumberExprRef`))
- A field definition or one or more value definition(s) with a parameter predicate.
- value : anyOf(float, :class:`ExprRef`)
- A constant value in visual domain (e.g., ``"red"`` / ``"#0099ff"`` / `gradient
- definition `__ for color,
- values between ``0`` to ``1`` for opacity).
- """
- _class_is_valid_at_instantiation = False
- _encoding_name = "angle"
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, aggregate=Undefined, bandPosition=Undefined, bin=Undefined, field=Undefined, legend=Undefined, scale=Undefined, sort=Undefined, test=Undefined, timeUnit=Undefined, title=Undefined, type=Undefined, **kwds) -> 'AngleValue':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, bandPosition=Undefined, datum=Undefined, legend=Undefined, scale=Undefined, test=Undefined, title=Undefined, type=Undefined, **kwds) -> 'AngleValue':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, aggregate=Undefined, bandPosition=Undefined, bin=Undefined, empty=Undefined, field=Undefined, legend=Undefined, param=Undefined, scale=Undefined, sort=Undefined, timeUnit=Undefined, title=Undefined, type=Undefined, **kwds) -> 'AngleValue':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, bandPosition=Undefined, datum=Undefined, empty=Undefined, legend=Undefined, param=Undefined, scale=Undefined, title=Undefined, type=Undefined, **kwds) -> 'AngleValue':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, test=Undefined, value=Undefined, **kwds) -> 'AngleValue':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, empty=Undefined, param=Undefined, value=Undefined, **kwds) -> 'AngleValue':
- ...
-
- @overload # type: ignore[no-overload-impl]
- def condition(self, _: List[core.ConditionalValueDefnumberExprRef], **kwds) -> 'AngleValue':
- ...
-
-
- def __init__(self, value, condition=Undefined, **kwds):
- super(AngleValue, self).__init__(value=value, condition=condition, **kwds)
-
-
-@with_property_setters
-class Color(FieldChannelMixin, core.FieldOrDatumDefWithConditionMarkPropFieldDefGradientstringnull):
- """Color schema wrapper
-
- Mapping(required=[shorthand])
-
- Parameters
- ----------
-
- shorthand : string
- shorthand for field, aggregate, and type
- aggregate : :class:`Aggregate`
- Aggregation function for the field (e.g., ``"mean"``, ``"sum"``, ``"median"``,
- ``"min"``, ``"max"``, ``"count"`` ).
-
- **Default value:** ``undefined`` (None)
-
- **See also:** `aggregate `__
- documentation.
- bandPosition : float
- Relative position on a band of a stacked, binned, time unit, or band scale. For
- example, the marks will be positioned at the beginning of the band if set to ``0``,
- and at the middle of the band if set to ``0.5``.
- bin : anyOf(boolean, :class:`BinParams`, None)
- A flag for binning a ``quantitative`` field, `an object defining binning parameters
- `__, or indicating
- that the data for ``x`` or ``y`` channel are binned before they are imported into
- Vega-Lite ( ``"binned"`` ).
-
-
- If ``true``, default `binning parameters
- `__ will be applied.
-
- If ``"binned"``, this indicates that the data for the ``x`` (or ``y`` ) channel are
- already binned. You can map the bin-start field to ``x`` (or ``y`` ) and the bin-end
- field to ``x2`` (or ``y2`` ). The scale and axis will be formatted similar to
- binning in Vega-Lite. To adjust the axis ticks based on the bin step, you can also
- set the axis's `tickMinStep
- `__ property.
-
- **Default value:** ``false``
-
- **See also:** `bin `__
- documentation.
- condition : anyOf(:class:`ConditionalValueDefGradientstringnullExprRef`, List(:class:`ConditionalValueDefGradientstringnullExprRef`))
- One or more value definition(s) with `a parameter or a test predicate
- `__.
-
- **Note:** A field definition's ``condition`` property can only contain `conditional
- value definitions `__
- since Vega-Lite only allows at most one encoded field per encoding channel.
- field : :class:`Field`
- **Required.** A string defining the name of the field from which to pull a data
- value or an object defining iterated values from the `repeat
- `__ operator.
-
- **See also:** `field `__
- documentation.
-
- **Notes:** 1) Dots ( ``.`` ) and brackets ( ``[`` and ``]`` ) can be used to access
- nested objects (e.g., ``"field": "foo.bar"`` and ``"field": "foo['bar']"`` ). If
- field names contain dots or brackets but are not nested, you can use ``\\`` to
- escape dots and brackets (e.g., ``"a\\.b"`` and ``"a\\[0\\]"`` ). See more details
- about escaping in the `field documentation
- `__. 2) ``field`` is not required
- if ``aggregate`` is ``count``.
- legend : anyOf(:class:`Legend`, None)
- An object defining properties of the legend. If ``null``, the legend for the
- encoding channel will be removed.
-
- **Default value:** If undefined, default `legend properties
- `__ are applied.
-
- **See also:** `legend `__
- documentation.
- scale : anyOf(:class:`Scale`, None)
- An object defining properties of the channel's scale, which is the function that
- transforms values in the data domain (numbers, dates, strings, etc) to visual values
- (pixels, colors, sizes) of the encoding channels.
-
- If ``null``, the scale will be `disabled and the data value will be directly encoded
- `__.
-
- **Default value:** If undefined, default `scale properties
- `__ are applied.
-
- **See also:** `scale `__
- documentation.
- sort : :class:`Sort`
- Sort order for the encoded field.
-
- For continuous fields (quantitative or temporal), ``sort`` can be either
- ``"ascending"`` or ``"descending"``.
-
- For discrete fields, ``sort`` can be one of the following:
-
-
- * ``"ascending"`` or ``"descending"`` -- for sorting by the values' natural order in
- JavaScript.
- * `A string indicating an encoding channel name to sort by
- `__ (e.g.,
- ``"x"`` or ``"y"`` ) with an optional minus prefix for descending sort (e.g.,
- ``"-x"`` to sort by x-field, descending). This channel string is short-form of `a
- sort-by-encoding definition
- `__. For
- example, ``"sort": "-x"`` is equivalent to ``"sort": {"encoding": "x", "order":
- "descending"}``.
- * `A sort field definition
- `__ for sorting by
- another field.
- * `An array specifying the field values in preferred order
- `__. In this case, the
- sort order will obey the values in the array, followed by any unspecified values
- in their original order. For discrete time field, values in the sort array can be
- `date-time definition objects
- `__. In addition, for time
- units ``"month"`` and ``"day"``, the values can be the month or day names (case
- insensitive) or their 3-letter initials (e.g., ``"Mon"``, ``"Tue"`` ).
- * ``null`` indicating no sort.
-
- **Default value:** ``"ascending"``
-
- **Note:** ``null`` and sorting by another channel is not supported for ``row`` and
- ``column``.
-
- **See also:** `sort