path
stringlengths
7
265
concatenated_notebook
stringlengths
46
17M
lecture/introduction/intro_python_I.ipynb
###Markdown Introduction V - Introduction to Python - I[Peer Herholz (he/him)](https://peerherholz.github.io/) Habilitation candidate - [Fiebach Lab](http://www.fiebachlab.org/), [Neurocognitive Psychology](https://www.psychologie.uni-frankfurt.de/49868684/Abteilungen) at [Goethe-University Frankfurt](https://www.goethe-university-frankfurt.de/en?locale=en) Research affiliate - [NeuroDataScience lab](https://neurodatascience.github.io/) at [MNI](https://www.mcgill.ca/neuro/)/[McGill](https://www.mcgill.ca/) Member - [BIDS](https://bids-specification.readthedocs.io/en/stable/), [ReproNim](https://www.repronim.org/), [Brainhack](https://brainhack.org/), [Neuromod](https://www.cneuromod.ca/), [OHBM SEA-SIG](https://ohbm-environment.org/), [UNIQUE](https://sites.google.com/view/unique-neuro-ai) &nbsp;&nbsp;@peerherholz Before we get started 1...- most of what you’ll see within this lecture was prepared by Ross Markello, Michael Notter and Peer Herholz and further adapted for this course by Peer Herholz - based on Tal Yarkoni's ["Introduction to Python" lecture at Neurohackademy 2019](https://neurohackademy.org/course/introduction-to-python-2/)- based on [IPython notebooks from J. R. Johansson](http://github.com/jrjohansson/scientific-python-lectures)- based on http://www.stavros.io/tutorials/python/ & http://www.swaroopch.com/notes/python- based on https://github.com/oesteban/biss2016 & https://github.com/jvns/pandas-cookbook Objectives 📍* learn basic and efficient usage of the python programming language * what is python & how to utilize it * building blocks of & operations in python What is Python?* Python is a programming language* Specifically, it's a **widely used/very flexible**, **high-level**, **general-purpose**, **dynamic** programming language* That's a mouthful! Let's explore each of these points in more detail... Widely-used* Python is the fastest-growing major programming language* Top 3 overall (with JavaScript, Java) High-levelPython features a high level of abstraction* Many operations that are explicit in lower-level languages (e.g., C/C++) are implicit in Python* E.g., memory allocation, garbage collection, etc.* Python lets you write code faster File reading in Java```javaimport java.io.BufferedReader;import java.io.FileReader;import java.io.IOException; public class ReadFile { public static void main(String[] args) throws IOException{ String fileContents = readEntireFile("./foo.txt"); } private static String readEntireFile(String filename) throws IOException { FileReader in = new FileReader(filename); StringBuilder contents = new StringBuilder(); char[] buffer = new char[4096]; int read = 0; do { contents.append(buffer, 0, read); read = in.read(buffer); } while (read >= 0); return contents.toString(); }}``` File-reading in Python```pythonopen(filename).read()``` General-purposeYou can do almost everything in Python* Comprehensive standard library* Enormous ecosystem of third-party packages* Widely used in many areas of software development (web, dev-ops, data science, etc.) DynamicCode is interpreted at run-time* No compilation process*; code is read line-by-line when executed* Eliminates delays between development and execution* The downside: poorer performance compared to compiled languages (Try typing `import antigravity` into a new cell and running it!) What we will do in this section of the course is a _short_ introduction to `Python` to help beginners to get familiar with this `programming language`.It is divided into the following chapters:- [Module](Module)- [Help and Descriptions](Help-and-Descriptions)- [Variables and types](Variables-and-types) - [Symbol names](Symbol-names) - [Assignment](Assignment) - [Fundamental types](Fundamental-types)- [Operators and comparisons](Operators-and-comparisons) - [Shortcut math operation and assignment](Shortcut-math-operation-and-assignment)- [Strings, List and dictionaries](Strings,-List-and-dictionaries) - [Strings](Strings) - [List](List) - [Tuples](Tuples) - [Dictionaries](Dictionaries)- [Indentation](Indentation)- [Control Flow](Control-Flow) - [Conditional statements: `if`, `elif`, `else`](Conditional-statements:-if,-elif,-else)- [Loops](Loops) - [`for` loops](for-loops) - [`break`, `continue` and `pass`](break,-continue-and-pass)- [Functions](Functions) - [Default argument and keyword arguments](Default-argument-and-keyword-arguments) - [`*args` and `*kwargs` parameters](*args-and-*kwargs-parameters) - [Unnamed functions: `lambda` function](Unnamed-functions:-lambda-function)- [Classes](Classes)- [Modules](Modules)- [Exceptions](Exceptions) Here's what we will focus on in the first block:- [Module](Module)- [Help and Descriptions](Help-and-Descriptions)- [Variables and types](Variables-and-types) - [Symbol names](Symbol-names) - [Assignment](Assignment) - [Fundamental types](Fundamental-types)- [Operators and comparisons](Operators-and-comparisons) - [Shortcut math operation and assignment](Shortcut-math-operation-and-assignment)- [Strings, List and dictionaries](Strings,-List-and-dictionaries) - [Strings](Strings) - [List](List) - [Tuples](Tuples) - [Dictionaries](Dictionaries) ModulesMost of the functionality in `Python` is provided by *modules*. To use a module in a Python program it first has to be imported. A module can be imported using the `import` statement. For example, to import the module `math`, which contains many standard mathematical functions, we can do: ###Code import math ###Output _____no_output_____ ###Markdown This includes the whole module and makes it available for use later in the program. For example, we can do: ###Code import math x = math.cos(2 * math.pi) print(x) ###Output 1.0 ###Markdown Importing the whole module us often times unnecessary and can lead to longer loading time or increase the memory consumption. An alternative to the previous method, we can also choose to import only a few selected functions from a module by explicitly listing which ones we want to import: ###Code from math import cos, pi x = cos(2 * pi) print(x) ###Output 1.0 ###Markdown You can make use of `tab` again to get a list of `functions`/`classes`/etc. for a given `module`. Try it out via navigating the cursor behind the `import statement` and press `tab`: ###Code from math import ###Output _____no_output_____ ###Markdown Comparably you can also use the `help` function to find out more about a given `module`: ###Code import math help(math) ###Output _____no_output_____ ###Markdown It is also possible to give an imported module or symbol your own access name with the `as` additional: ###Code import numpy as np from math import pi as number_pi x = np.rad2deg(number_pi) print(x) ###Output 180.0 ###Markdown You can basically provide any name (given it's following `python`/`coding` conventions) but focusing on intelligibility won't be the worst idea: ###Code import matplotlib as pineapple pineapple. ###Output _____no_output_____ ###Markdown Exercise 1.1Import the `max` from `numpy` and find out what it does. ###Code # write your solution in this code cell from numpy import max help(max) ###Output _____no_output_____ ###Markdown Exercise 1.2Import the `scipy` package and assign the access name `middle_earth` and check its `functions`. ###Code # write your solution in this code cell import scipy as middle_earth help(middle_earth) ###Output Help on package scipy: NAME scipy DESCRIPTION SciPy: A scientific computing package for Python ================================================ Documentation is available in the docstrings and online at https://docs.scipy.org. Contents -------- SciPy imports all the functions from the NumPy namespace, and in addition provides: Subpackages ----------- Using any of these subpackages requires an explicit import. For example, ``import scipy.cluster``. :: cluster --- Vector Quantization / Kmeans fft --- Discrete Fourier transforms fftpack --- Legacy discrete Fourier transforms integrate --- Integration routines interpolate --- Interpolation Tools io --- Data input and output linalg --- Linear algebra routines linalg.blas --- Wrappers to BLAS library linalg.lapack --- Wrappers to LAPACK library misc --- Various utilities that don't have another home. ndimage --- N-D image package odr --- Orthogonal Distance Regression optimize --- Optimization Tools signal --- Signal Processing Tools signal.windows --- Window functions sparse --- Sparse Matrices sparse.linalg --- Sparse Linear Algebra sparse.linalg.dsolve --- Linear Solvers sparse.linalg.dsolve.umfpack --- :Interface to the UMFPACK library: Conjugate Gradient Method (LOBPCG) sparse.linalg.eigen --- Sparse Eigenvalue Solvers sparse.linalg.eigen.lobpcg --- Locally Optimal Block Preconditioned Conjugate Gradient Method (LOBPCG) spatial --- Spatial data structures and algorithms special --- Special functions stats --- Statistical Functions Utility tools ------------- :: test --- Run scipy unittests show_config --- Show scipy build configuration show_numpy_config --- Show numpy build configuration __version__ --- SciPy version string __numpy_version__ --- Numpy version string PACKAGE CONTENTS __config__ _build_utils (package) _distributor_init _lib (package) cluster (package) conftest constants (package) fft (package) fftpack (package) integrate (package) interpolate (package) io (package) linalg (package) misc (package) ndimage (package) odr (package) optimize (package) setup signal (package) sparse (package) spatial (package) special (package) stats (package) version DATA test = <scipy._lib._testutils.PytestTester object> VERSION 1.7.1 FILE /Users/peerherholz/anaconda3/envs/pfp_2021/lib/python3.9/site-packages/scipy/__init__.py ###Markdown Exercise 1.3What happens when we try to import a `module` that is either misspelled or doesn't exist in our `environment` or at all?1. `python` provides us a hint that the `module` name might be misspelled2. we'll get an `error` telling us that the `module` doesn't exist3. `python` automatically searches for the `module` and if it exists downloads/installs it ###Code import welovethiscourse ###Output _____no_output_____ ###Markdown Namespaces and imports* Python is **very** serious about maintaining orderly `namespaces`* If you want to use some code outside the current scope, you need to explicitly "`import`" it* Python's import system often annoys beginners, but it substantially increases `code` clarity * Almost completely eliminates naming conflicts and confusion Help and DescriptionsUsing the function `help` we can get a description of almost all functions. ###Code help(math.log) math.log(10) math.log(10, 2) ###Output _____no_output_____ ###Markdown Variables and data types* in programming `variables` are things that store `values`* in `Python`, we declare a `variable` by **assigning** it a `value` with the `=` sign * `name = value` * code `variables` **!=** math variables * in mathematics `=` refers to equality (statement of truth), e.g. `y = 10x + 2` * in coding `=` refers to assignments, e.g. `x = x + 1` * Variables are pointers, not data stores! * `Python` supports a variety of `data types` and `structures`: * `booleans` * `numbers` (`ints`, `floats`, etc.) * `strings` * `lists` * `dictionaries` * many others!* We don't specify a variable's type at assignment Variables and types Symbol names Variable names in Python can contain alphanumerical characters `a-z`, `A-Z`, `0-9` and some special characters such as `_`. Normal variable names must start with a letter. By convention, variable names start with a lower-case letter, and Class names start with a capital letter. In addition, there are a number of Python keywords that cannot be used as variable names. These keywords are: and, as, assert, break, class, continue, def, del, elif, else, except, exec, finally, for, from, global, if, import, in, is, lambda, not, or, pass, print, raise, return, try, while, with, yield Assignment(Not your homework assignment but the operator in `python`.)The assignment operator in `Python` is `=`. `Python` is a `dynamically typed language`, so we do not need to specify the type of a `variable` when we create one.`Assigning` a `value` to a new `variable` _creates_ the `variable`: ###Code # variable assignment x = 1.0 ###Output _____no_output_____ ###Markdown Again, this does not mean that `x` equals `1` but that the `variable` `x` has the `value` `1`. Thus, our `variable` `x` is _stored_ in the respective `namespace`: ###Code x ###Output _____no_output_____ ###Markdown This means that we can directly utilize the `value` of our `variable`: ###Code x + 3 ###Output _____no_output_____ ###Markdown Although not explicitly specified, a `variable` does have a `type` associated with it. The `type` is _derived_ from the `value` it was `assigned`. ###Code type(x) ###Output _____no_output_____ ###Markdown If we `assign` a new `value` to a `variable`, its `type` can change. ###Code x = 1 type(x) ###Output _____no_output_____ ###Markdown This outline one further _very important_ characteristic of `python` (and many other programming languages): `variables` can be directly overwritten by `assigning` them a new `value`. We don't get an error like "This `namespace` is already taken." Thus, always remember/keep track of what `namespaces` were already used to avoid unintentional deletions/errors (reproducibility/replicability much?). ###Code ring_bearer = 'Bilbo' ring_bearer ring_bearer = 'Frodo' ring_bearer ###Output _____no_output_____ ###Markdown If we try to use a variable that has not yet been defined we get an `NameError` (Note for later sessions, that we will use in the notebooks `try/except` blocks to handle the exception, so the notebook doesn't stop. The code below will try to execute `print` function and if the `NameError` occurs the error message will be printed. Otherwise, an error will be raised. You will learn more about exception handling later.): ###Code try: print(Peer) except(NameError) as err: print("NameError", err) else: raise ###Output NameError name 'Peer' is not defined ###Markdown Variable names:* Can include `letters` (A-Z), `digits` (0-9), and `underscores` ( _ )* Cannot start with a `digit`* Are **case sensitive** (questions: where did "lower/upper case" originate?)This means that, for example:* `shire0` is a valid variable name, whereas `0shire` is not* `shire` and `Shire` are different variables Exercise 2.1Create the following `variables` `n_elves`, `n_dwarfs`, `n_humans` with the respective values `3`, `7.0` and `nine`. ###Code # write your solution here n_elves = 3 n_dwarfs = 7.0 n_humans = "nine" ###Output _____no_output_____ ###Markdown Exercise 2.2What's the output of `n_elves + n_dwarfs`?1. `n_elves + n_dwarfs`2. 103. 10.0 ###Code n_elves + n_dwarfs ###Output _____no_output_____ ###Markdown Exercise 2.3Consider the following lines of code. `ring_bearer = 'Gollum'` `ring_bearer` `ring_bearer = 'Bilbo'` `ring_bearer` What is the final output?1. `'Bilbo'`2. `'Gollum'`3. neither, the variable got deleted ###Code ring_bearer = 'Gollum' ring_bearer ring_bearer = 'Bilbo' ring_bearer ###Output _____no_output_____ ###Markdown Fundamental types & data structures* Most code requires more _complex structures_ built out of _basic data `types`_* `data type` refers to the `value` that is `assigned` to a `variable` * `Python` provides built-in support for many common structures * Many additional structures can be found in the [collections](https://docs.python.org/3/library/collections.html) module Most of the time you'll encounter the following `data types`* `integers` (e.g. `1`, `42`, `180`)* `floating-point numbers` (e.g. `1.0`, `42.42`, `180.90`)* `strings` (e.g. `"Rivendell"`, `"Weathertop"`)* `Boolean` (`True`, `False`)If you're unsure about the `data type` of a given `variable`, you can always use the `type()` command. IntegersLets check out the different `data types` in more detail, starting with `integers`. `Intergers` are _natural numbers_ that can be _signed_ (e.g. `1`, `42`, `180`, `-1`, `-42`, `-180`). ###Code x = 1 type(x) n_nazgul = 9 type(n_nazgul) remaining_rings = -1 type(remaining_rings) ###Output _____no_output_____ ###Markdown Floating-point numbersSo what's the difference to `floating-point numbers`? `Floating-point numbers` are _decimal-point number_ that can be _signed_ (e.g. `1.0`, `42.42`, `180.90`, `-1.0`, `-42.42`, `-180.90`). ###Code x_float = 1.0 type(x_float) n_nazgul_float = 9.0 type(n_nazgul_float) remaining_rings_float = -1.0 type(remaining_rings_float) ###Output _____no_output_____ ###Markdown StringsNext up: `strings`. `Strings` are basically `text elements`, from `letters` to `words` to `sentences` all can be/are `strings` in `python`. In order to define a `string`, `Python` needs **quotation marks**, more precisely `strings` start and end with quotation marks, e.g. `"Rivendell"`. You can choose between `"` and `'` as both will work (NB: `python` will put `'` around `strings` even if you specified `"`). However, it is recommended to decide on one and be consistent. ###Code location = "Weathertop" type(location) abbreviation = 'LOTR' type(abbreviation) book_one = "The fellowship of the ring" type(book_one) ###Output _____no_output_____ ###Markdown BooleansHow about some `Boolean`s? At this point it gets a bit more "abstract". While there are many possible `numbers` and `strings`, a Boolean can only have one of two `values`: `True` or `False`. That is, a `Boolean` says something about whether something _is the case or not_. It's easier to understand with some examples. First try the `type()` function with a `Boolean` as an argument. ###Code b1 = True type(b1) b2 = False type(b2) lotr_is_awesome = True type(lotr_is_awesome) ###Output _____no_output_____ ###Markdown Interestingly, `True` and `False` also have `numeric values`! `True` has a value of `1` and `False` has a value of `0`. ###Code True + True False + False ###Output _____no_output_____ ###Markdown Converting data typesAs mentioned before the `data type` is not set when `assigning` a `value` to a `variable` but determined based on its properties. Additionally, the `data type` of a given `value` can also be changed via set of functions.- `int()` -> convert the `value` of a `variable` to an `integer`- `float()` -> convert the `value` of a `variable` to a `floating-point number`- `str()` -> convert the `value` of a `variable` to a `string`- `bool()` -> convert the `value` of a `variable` to a `Boolean` ###Code int("4") float(3) str(2) bool(1) ###Output _____no_output_____ ###Markdown Exercise 3.1Define the following `variables` with the respective `values` and `data types`: `fellowship_n_humans` with a `value` of two as a `float`, `fellowship_n_hobbits` with a `value` of four as a `string` and `fellowship_n_elves` with a value of one as an `integer`. ###Code # write your solution here fellowship_n_humans = 2.0 fellowship_n_hobbits = 'four' fellowship_n_elves = 1 ###Output _____no_output_____ ###Markdown Exercise 3.2What outcome would you expect based on the following lines of code?1. `True - False`2. `type(True)` 1. `1`2. `bool` Exercise 3.3Define two `variables`, `fellowship_n_dwarfs` with a `value` of one as a `string` and `fellowship_n_wizards` with a `value` of one as a `float`. Subsequently, change the `data type` of `fellowship_n_dwarfs` to `integer` and the `data type` of `fellowship_n_wizard` to `string`. ###Code fellowship_n_dwarfs = 1.0 fellowship_n_wizards = '1.0' int(fellowship_n_dwarfs) str(fellowship_n_wizards) ###Output _____no_output_____
Coursera/Intro to TensorFlow/Week-2/Example/b_estimator.ipynb
###Markdown 2b. Machine Learning using tf.estimator In this notebook, we will create a machine learning model using tf.estimator and evaluate its performance. The dataset is rather small (7700 samples), so we can do it all in-memory. We will also simply pass the raw data in as-is. ###Code import datalab.bigquery as bq import tensorflow as tf import pandas as pd import numpy as np import shutil print(tf.__version__) ###Output /usr/local/envs/py3env/lib/python3.5/site-packages/h5py/__init__.py:36: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`. from ._conv import register_converters as _register_converters ###Markdown Read data created in the previous chapter. ###Code # In CSV, label is the first column, after the features, followed by the key CSV_COLUMNS = ['fare_amount', 'pickuplon','pickuplat','dropofflon','dropofflat','passengers', 'key'] FEATURES = CSV_COLUMNS[1:len(CSV_COLUMNS) - 1] LABEL = CSV_COLUMNS[0] df_train = pd.read_csv('./taxi-train.csv', header = None, names = CSV_COLUMNS) df_valid = pd.read_csv('./taxi-valid.csv', header = None, names = CSV_COLUMNS) df_test = pd.read_csv('./taxi-test.csv', header = None, names = CSV_COLUMNS) ###Output _____no_output_____ ###Markdown Train and eval input functions to read from Pandas Dataframe ###Code def make_train_input_fn(df, num_epochs): return tf.estimator.inputs.pandas_input_fn( x = df, y = df[LABEL], batch_size = 128, num_epochs = num_epochs, shuffle = True, queue_capacity = 1000 ) def make_eval_input_fn(df): return tf.estimator.inputs.pandas_input_fn( x = df, y = df[LABEL], batch_size = 128, shuffle = False, queue_capacity = 1000 ) ###Output _____no_output_____ ###Markdown Our input function for predictions is the same except we don't provide a label ###Code def make_prediction_input_fn(df): return tf.estimator.inputs.pandas_input_fn( x = df, y = None, batch_size = 128, shuffle = False, queue_capacity = 1000 ) ###Output _____no_output_____ ###Markdown Create feature columns for estimator ###Code def make_feature_cols(): input_columns = [tf.feature_column.numeric_column(k) for k in FEATURES] return input_columns ###Output _____no_output_____ ###Markdown Linear Regression with tf.Estimator framework ###Code tf.logging.set_verbosity(tf.logging.INFO) OUTDIR = 'taxi_trained' shutil.rmtree(OUTDIR, ignore_errors = True) # start fresh each time model = tf.estimator.LinearRegressor( feature_columns = make_feature_cols(), model_dir = OUTDIR) model.train(input_fn = make_train_input_fn(df_train, num_epochs = 10)) ###Output INFO:tensorflow:Using default config. INFO:tensorflow:Using config: {'_save_checkpoints_steps': None, '_evaluation_master': '', '_keep_checkpoint_every_n_hours': 10000, '_is_chief': True, '_service': None, '_cluster_spec': <tensorflow.python.training.server_lib.ClusterSpec object at 0x7fbf55a983c8>, '_master': '', '_global_id_in_cluster': 0, '_num_ps_replicas': 0, '_task_type': 'worker', '_model_dir': 'taxi_trained', '_num_worker_replicas': 1, '_save_checkpoints_secs': 600, '_tf_random_seed': None, '_log_step_count_steps': 100, '_save_summary_steps': 100, '_task_id': 0, '_train_distribute': None, '_session_config': None, '_keep_checkpoint_max': 5} INFO:tensorflow:Calling model_fn. INFO:tensorflow:Done calling model_fn. INFO:tensorflow:Create CheckpointSaverHook. INFO:tensorflow:Graph was finalized. INFO:tensorflow:Running local_init_op. INFO:tensorflow:Done running local_init_op. INFO:tensorflow:Saving checkpoints for 1 into taxi_trained/model.ckpt. INFO:tensorflow:loss = 23227.691, step = 1 INFO:tensorflow:global_step/sec: 230.586 INFO:tensorflow:loss = 8995.027, step = 101 (0.438 sec) INFO:tensorflow:global_step/sec: 279.531 INFO:tensorflow:loss = 9363.185, step = 201 (0.357 sec) INFO:tensorflow:global_step/sec: 258.107 INFO:tensorflow:loss = 10644.762, step = 301 (0.388 sec) INFO:tensorflow:global_step/sec: 280.008 INFO:tensorflow:loss = 5163.014, step = 401 (0.357 sec) INFO:tensorflow:global_step/sec: 278.033 INFO:tensorflow:loss = 7394.787, step = 501 (0.360 sec) INFO:tensorflow:global_step/sec: 252.291 INFO:tensorflow:loss = 10883.856, step = 601 (0.396 sec) INFO:tensorflow:Saving checkpoints for 608 into taxi_trained/model.ckpt. INFO:tensorflow:Loss for final step: 112.020546. ###Markdown Evaluate on the validation data (we should defer using the test data to after we have selected a final model). ###Code def print_rmse(model, df): metrics = model.evaluate(input_fn = make_eval_input_fn(df)) print('RMSE on dataset = {}'.format(np.sqrt(metrics['average_loss']))) print_rmse(model, df_valid) ###Output INFO:tensorflow:Calling model_fn. INFO:tensorflow:Done calling model_fn. INFO:tensorflow:Starting evaluation at 2018-11-21-03:41:27 INFO:tensorflow:Graph was finalized. INFO:tensorflow:Restoring parameters from taxi_trained/model.ckpt-608 INFO:tensorflow:Running local_init_op. INFO:tensorflow:Done running local_init_op. INFO:tensorflow:Finished evaluation at 2018-11-21-03:41:27 INFO:tensorflow:Saving dict for global step 608: average_loss = 109.164764, global_step = 608, loss = 12982.81 RMSE on dataset = 10.44819450378418 ###Markdown This is nowhere near our benchmark (RMSE of $6 or so on this data), but it serves to demonstrate what TensorFlow code looks like. Let's use this model for prediction. ###Code predictions = model.predict(input_fn = make_prediction_input_fn(df_test)) for items in predictions: print(items) ###Output INFO:tensorflow:Calling model_fn. INFO:tensorflow:Done calling model_fn. INFO:tensorflow:Graph was finalized. INFO:tensorflow:Restoring parameters from taxi_trained/model.ckpt-608 INFO:tensorflow:Running local_init_op. INFO:tensorflow:Done running local_init_op. {'predictions': array([11.043583], dtype=float32)} {'predictions': array([11.041], dtype=float32)} {'predictions': array([11.041875], dtype=float32)} {'predictions': array([11.039393], dtype=float32)} {'predictions': array([11.043319], dtype=float32)} {'predictions': array([11.043096], dtype=float32)} {'predictions': array([11.0416565], dtype=float32)} {'predictions': array([11.041688], dtype=float32)} {'predictions': array([11.043474], dtype=float32)} {'predictions': array([11.041298], dtype=float32)} {'predictions': array([11.043584], dtype=float32)} {'predictions': array([11.043752], dtype=float32)} {'predictions': array([11.036937], dtype=float32)} {'predictions': array([11.040911], dtype=float32)} {'predictions': array([11.098388], dtype=float32)} {'predictions': array([11.041732], dtype=float32)} {'predictions': array([11.042587], dtype=float32)} {'predictions': array([11.099429], dtype=float32)} {'predictions': array([11.042919], dtype=float32)} {'predictions': array([11.040049], dtype=float32)} {'predictions': array([11.04131], dtype=float32)} {'predictions': array([11.037168], dtype=float32)} {'predictions': array([11.041463], dtype=float32)} {'predictions': array([11.100605], dtype=float32)} {'predictions': array([11.044191], dtype=float32)} {'predictions': array([11.033843], dtype=float32)} {'predictions': array([11.151931], dtype=float32)} {'predictions': array([11.325656], dtype=float32)} {'predictions': array([11.04468], dtype=float32)} {'predictions': array([11.045712], dtype=float32)} {'predictions': array([11.269823], dtype=float32)} {'predictions': array([11.040107], dtype=float32)} {'predictions': array([11.040774], dtype=float32)} {'predictions': array([11.04356], dtype=float32)} {'predictions': array([11.153978], dtype=float32)} {'predictions': array([11.042667], dtype=float32)} {'predictions': array([11.040906], dtype=float32)} {'predictions': array([11.037081], dtype=float32)} {'predictions': array([11.210749], dtype=float32)} {'predictions': array([11.037425], dtype=float32)} {'predictions': array([11.042201], dtype=float32)} {'predictions': array([11.040187], dtype=float32)} {'predictions': array([11.039892], dtype=float32)} {'predictions': array([11.040304], dtype=float32)} {'predictions': array([11.097316], dtype=float32)} {'predictions': array([11.042182], dtype=float32)} {'predictions': array([11.095655], dtype=float32)} {'predictions': array([11.266023], dtype=float32)} {'predictions': array([11.042469], dtype=float32)} {'predictions': array([11.040689], dtype=float32)} {'predictions': array([11.044618], dtype=float32)} {'predictions': array([11.2669], dtype=float32)} {'predictions': array([11.041177], dtype=float32)} {'predictions': array([11.04227], dtype=float32)} {'predictions': array([11.040629], dtype=float32)} {'predictions': array([11.041837], dtype=float32)} {'predictions': array([11.040626], dtype=float32)} {'predictions': array([11.270249], dtype=float32)} {'predictions': array([11.041254], dtype=float32)} {'predictions': array([11.094373], dtype=float32)} {'predictions': array([11.043522], dtype=float32)} {'predictions': array([11.156254], dtype=float32)} {'predictions': array([11.040959], dtype=float32)} {'predictions': array([11.042159], dtype=float32)} {'predictions': array([11.153913], dtype=float32)} {'predictions': array([11.04108], dtype=float32)} {'predictions': array([11.042675], dtype=float32)} {'predictions': array([11.044163], dtype=float32)} {'predictions': array([11.041129], dtype=float32)} {'predictions': array([11.041718], dtype=float32)} {'predictions': array([11.036959], dtype=float32)} {'predictions': array([11.042114], dtype=float32)} {'predictions': array([11.098], dtype=float32)} {'predictions': array([11.041291], dtype=float32)} {'predictions': array([11.04135], dtype=float32)} {'predictions': array([11.041912], dtype=float32)} {'predictions': array([11.044057], dtype=float32)} {'predictions': array([11.041034], dtype=float32)} {'predictions': array([11.270366], dtype=float32)} {'predictions': array([11.097284], dtype=float32)} {'predictions': array([11.042729], dtype=float32)} {'predictions': array([11.041282], dtype=float32)} {'predictions': array([11.041289], dtype=float32)} {'predictions': array([11.093856], dtype=float32)} {'predictions': array([11.03641], dtype=float32)} {'predictions': array([11.154336], dtype=float32)} {'predictions': array([11.267577], dtype=float32)} {'predictions': array([11.268383], dtype=float32)} {'predictions': array([11.043536], dtype=float32)} {'predictions': array([11.042795], dtype=float32)} {'predictions': array([11.038993], dtype=float32)} {'predictions': array([11.042381], dtype=float32)} {'predictions': array([11.251331], dtype=float32)} {'predictions': array([11.09615], dtype=float32)} {'predictions': array([11.035989], dtype=float32)} {'predictions': array([11.041703], dtype=float32)} {'predictions': array([11.042155], dtype=float32)} {'predictions': array([11.267194], dtype=float32)} {'predictions': array([11.041633], dtype=float32)} {'predictions': array([11.049363], dtype=float32)} {'predictions': array([11.04054], dtype=float32)} {'predictions': array([11.098466], dtype=float32)} {'predictions': array([11.038484], dtype=float32)} {'predictions': array([11.03915], dtype=float32)} {'predictions': array([11.037566], dtype=float32)} {'predictions': array([11.042184], dtype=float32)} {'predictions': array([11.268783], dtype=float32)} {'predictions': array([11.039331], dtype=float32)} {'predictions': array([11.1555395], dtype=float32)} {'predictions': array([11.043388], dtype=float32)} {'predictions': array([11.041855], dtype=float32)} {'predictions': array([11.041406], dtype=float32)} {'predictions': array([11.042376], dtype=float32)} {'predictions': array([11.043866], dtype=float32)} {'predictions': array([11.043215], dtype=float32)} {'predictions': array([11.040164], dtype=float32)} {'predictions': array([11.041501], dtype=float32)} {'predictions': array([11.04311], dtype=float32)} {'predictions': array([11.0432], dtype=float32)} {'predictions': array([11.040465], dtype=float32)} {'predictions': array([11.042335], dtype=float32)} {'predictions': array([11.09958], dtype=float32)} {'predictions': array([11.154657], dtype=float32)} {'predictions': array([11.042291], dtype=float32)} {'predictions': array([11.041546], dtype=float32)} {'predictions': array([11.041478], dtype=float32)} {'predictions': array([11.040216], dtype=float32)} {'predictions': array([11.041989], dtype=float32)} {'predictions': array([11.043239], dtype=float32)} {'predictions': array([11.039685], dtype=float32)} {'predictions': array([11.0423565], dtype=float32)} {'predictions': array([11.039973], dtype=float32)} {'predictions': array([11.040326], dtype=float32)} {'predictions': array([11.042178], dtype=float32)} {'predictions': array([11.040891], dtype=float32)} {'predictions': array([11.100114], dtype=float32)} {'predictions': array([11.038007], dtype=float32)} {'predictions': array([11.023517], dtype=float32)} {'predictions': array([11.043558], dtype=float32)} {'predictions': array([11.042251], dtype=float32)} {'predictions': array([11.041613], dtype=float32)} {'predictions': array([11.326094], dtype=float32)} {'predictions': array([11.041692], dtype=float32)} {'predictions': array([11.04058], dtype=float32)} {'predictions': array([11.154365], dtype=float32)} {'predictions': array([11.0423], dtype=float32)} {'predictions': array([11.095227], dtype=float32)} {'predictions': array([11.042067], dtype=float32)} {'predictions': array([11.041041], dtype=float32)} {'predictions': array([11.099038], dtype=float32)} {'predictions': array([11.044602], dtype=float32)} {'predictions': array([11.155069], dtype=float32)} {'predictions': array([11.04067], dtype=float32)} {'predictions': array([11.042636], dtype=float32)} {'predictions': array([11.097184], dtype=float32)} {'predictions': array([11.04221], dtype=float32)} {'predictions': array([11.041746], dtype=float32)} {'predictions': array([11.042782], dtype=float32)} {'predictions': array([11.268113], dtype=float32)} {'predictions': array([11.0409775], dtype=float32)} {'predictions': array([11.040616], dtype=float32)} {'predictions': array([11.09759], dtype=float32)} {'predictions': array([11.0425625], dtype=float32)} {'predictions': array([11.042155], dtype=float32)} {'predictions': array([11.04171], dtype=float32)} {'predictions': array([11.041484], dtype=float32)} {'predictions': array([11.041098], dtype=float32)} {'predictions': array([11.043688], dtype=float32)} {'predictions': array([11.099511], dtype=float32)} {'predictions': array([11.040143], dtype=float32)} {'predictions': array([11.269108], dtype=float32)} {'predictions': array([11.099256], dtype=float32)} {'predictions': array([11.04424], dtype=float32)} {'predictions': array([11.040928], dtype=float32)} {'predictions': array([11.038647], dtype=float32)} {'predictions': array([11.04202], dtype=float32)} {'predictions': array([11.098649], dtype=float32)} {'predictions': array([11.041494], dtype=float32)} {'predictions': array([11.156475], dtype=float32)} {'predictions': array([11.041796], dtype=float32)} {'predictions': array([11.044157], dtype=float32)} {'predictions': array([11.041313], dtype=float32)} {'predictions': array([11.04304], dtype=float32)} {'predictions': array([11.044129], dtype=float32)} {'predictions': array([11.157306], dtype=float32)} {'predictions': array([11.041753], dtype=float32)} {'predictions': array([11.041914], dtype=float32)} {'predictions': array([11.098388], dtype=float32)} {'predictions': array([11.1455555], dtype=float32)} {'predictions': array([11.043727], dtype=float32)} {'predictions': array([11.324195], dtype=float32)} {'predictions': array([11.041725], dtype=float32)} {'predictions': array([11.042469], dtype=float32)} {'predictions': array([11.043445], dtype=float32)} {'predictions': array([11.039519], dtype=float32)} {'predictions': array([11.036065], dtype=float32)} {'predictions': array([11.327052], dtype=float32)} {'predictions': array([11.040879], dtype=float32)} {'predictions': array([11.041282], dtype=float32)} {'predictions': array([11.043299], dtype=float32)} {'predictions': array([11.038628], dtype=float32)} {'predictions': array([11.035419], dtype=float32)} {'predictions': array([11.039596], dtype=float32)} {'predictions': array([11.042705], dtype=float32)} {'predictions': array([11.041706], dtype=float32)} {'predictions': array([11.099128], dtype=float32)} {'predictions': array([11.04482], dtype=float32)} {'predictions': array([11.041065], dtype=float32)} {'predictions': array([11.0440645], dtype=float32)} {'predictions': array([11.0415745], dtype=float32)} {'predictions': array([11.2684555], dtype=float32)} {'predictions': array([11.041735], dtype=float32)} {'predictions': array([11.041736], dtype=float32)} {'predictions': array([11.041774], dtype=float32)} {'predictions': array([11.037273], dtype=float32)} {'predictions': array([11.042819], dtype=float32)} {'predictions': array([11.042189], dtype=float32)} {'predictions': array([11.042265], dtype=float32)} {'predictions': array([11.041152], dtype=float32)} {'predictions': array([11.040235], dtype=float32)} {'predictions': array([11.040625], dtype=float32)} {'predictions': array([11.042375], dtype=float32)} {'predictions': array([11.042636], dtype=float32)} {'predictions': array([11.043384], dtype=float32)} {'predictions': array([11.040967], dtype=float32)} {'predictions': array([11.042022], dtype=float32)} {'predictions': array([11.044832], dtype=float32)} {'predictions': array([11.055435], dtype=float32)} {'predictions': array([11.09761], dtype=float32)} {'predictions': array([11.266256], dtype=float32)} {'predictions': array([11.040114], dtype=float32)} {'predictions': array([11.042281], dtype=float32)} {'predictions': array([11.269469], dtype=float32)} {'predictions': array([11.040534], dtype=float32)} {'predictions': array([11.04277], dtype=float32)} {'predictions': array([11.098203], dtype=float32)} {'predictions': array([11.098971], dtype=float32)} {'predictions': array([11.153664], dtype=float32)} {'predictions': array([11.041707], dtype=float32)} {'predictions': array([11.26889], dtype=float32)} {'predictions': array([11.040719], dtype=float32)} {'predictions': array([11.154336], dtype=float32)} {'predictions': array([11.036353], dtype=float32)} {'predictions': array([11.033702], dtype=float32)} {'predictions': array([11.039275], dtype=float32)} {'predictions': array([11.040946], dtype=float32)} {'predictions': array([11.0233345], dtype=float32)} {'predictions': array([11.041219], dtype=float32)} {'predictions': array([11.026675], dtype=float32)} {'predictions': array([11.097633], dtype=float32)} {'predictions': array([11.043438], dtype=float32)} {'predictions': array([11.042915], dtype=float32)} {'predictions': array([11.09916], dtype=float32)} {'predictions': array([11.093188], dtype=float32)} {'predictions': array([11.100141], dtype=float32)} {'predictions': array([11.04003], dtype=float32)} {'predictions': array([11.037086], dtype=float32)} {'predictions': array([11.155109], dtype=float32)} {'predictions': array([11.03974], dtype=float32)} {'predictions': array([11.043239], dtype=float32)} {'predictions': array([11.042431], dtype=float32)} {'predictions': array([11.099305], dtype=float32)} {'predictions': array([11.154876], dtype=float32)} {'predictions': array([11.041865], dtype=float32)} {'predictions': array([11.212911], dtype=float32)} {'predictions': array([11.096779], dtype=float32)} {'predictions': array([11.042095], dtype=float32)} {'predictions': array([11.03896], dtype=float32)} {'predictions': array([11.027762], dtype=float32)} {'predictions': array([11.040745], dtype=float32)} {'predictions': array([11.09949], dtype=float32)} {'predictions': array([11.042129], dtype=float32)} {'predictions': array([11.266614], dtype=float32)} {'predictions': array([11.042362], dtype=float32)} {'predictions': array([11.043208], dtype=float32)} {'predictions': array([11.043883], dtype=float32)} {'predictions': array([11.041439], dtype=float32)} {'predictions': array([11.038167], dtype=float32)} {'predictions': array([11.041547], dtype=float32)} {'predictions': array([11.041623], dtype=float32)} {'predictions': array([11.157918], dtype=float32)} {'predictions': array([11.036847], dtype=float32)} {'predictions': array([11.212773], dtype=float32)} {'predictions': array([11.040875], dtype=float32)} {'predictions': array([11.043244], dtype=float32)} {'predictions': array([11.039541], dtype=float32)} {'predictions': array([11.098044], dtype=float32)} {'predictions': array([11.041725], dtype=float32)} {'predictions': array([11.32348], dtype=float32)} {'predictions': array([11.0412], dtype=float32)} {'predictions': array([11.041936], dtype=float32)} {'predictions': array([11.040797], dtype=float32)} {'predictions': array([11.098888], dtype=float32)} {'predictions': array([11.045219], dtype=float32)} {'predictions': array([11.101174], dtype=float32)} {'predictions': array([11.098267], dtype=float32)} {'predictions': array([11.266715], dtype=float32)} {'predictions': array([11.140086], dtype=float32)} {'predictions': array([11.042646], dtype=float32)} {'predictions': array([11.211947], dtype=float32)} {'predictions': array([11.0987], dtype=float32)} {'predictions': array([11.098557], dtype=float32)} {'predictions': array([11.039739], dtype=float32)} {'predictions': array([11.043261], dtype=float32)} {'predictions': array([11.042076], dtype=float32)} {'predictions': array([11.155607], dtype=float32)} {'predictions': array([11.09966], dtype=float32)} {'predictions': array([11.027762], dtype=float32)} {'predictions': array([11.04462], dtype=float32)} {'predictions': array([11.041652], dtype=float32)} {'predictions': array([11.040624], dtype=float32)} {'predictions': array([11.043413], dtype=float32)} {'predictions': array([11.04267], dtype=float32)} {'predictions': array([11.098037], dtype=float32)} {'predictions': array([11.154424], dtype=float32)} {'predictions': array([11.041417], dtype=float32)} {'predictions': array([11.0415], dtype=float32)} {'predictions': array([11.042503], dtype=float32)} {'predictions': array([11.042451], dtype=float32)} {'predictions': array([11.155456], dtype=float32)} {'predictions': array([11.03874], dtype=float32)} {'predictions': array([11.042431], dtype=float32)} {'predictions': array([11.155425], dtype=float32)} {'predictions': array([11.041129], dtype=float32)} {'predictions': array([11.042279], dtype=float32)} {'predictions': array([11.14149], dtype=float32)} {'predictions': array([11.263701], dtype=float32)} {'predictions': array([10.983853], dtype=float32)} {'predictions': array([11.027713], dtype=float32)} {'predictions': array([11.03692], dtype=float32)} {'predictions': array([11.041049], dtype=float32)} {'predictions': array([11.040301], dtype=float32)} {'predictions': array([11.042241], dtype=float32)} {'predictions': array([11.099024], dtype=float32)} {'predictions': array([11.03748], dtype=float32)} {'predictions': array([11.041721], dtype=float32)} {'predictions': array([11.100791], dtype=float32)} {'predictions': array([11.263174], dtype=float32)} {'predictions': array([11.266427], dtype=float32)} {'predictions': array([11.042035], dtype=float32)} {'predictions': array([11.042427], dtype=float32)} ###Markdown This explains why the RMSE was so high -- the model essentially predicts the same amount for every trip. Would a more complex model help? Let's try using a deep neural network. The code to do this is quite straightforward as well. Deep Neural Network regression ###Code tf.logging.set_verbosity(tf.logging.INFO) shutil.rmtree(OUTDIR, ignore_errors = True) # start fresh each time model = tf.estimator.DNNRegressor(hidden_units = [32, 8, 2], feature_columns = make_feature_cols(), model_dir = OUTDIR) model.train(input_fn = make_train_input_fn(df_train, num_epochs = 100)); print_rmse(model, df_valid) ###Output INFO:tensorflow:Using default config. INFO:tensorflow:Using config: {'_save_checkpoints_steps': None, '_evaluation_master': '', '_keep_checkpoint_every_n_hours': 10000, '_is_chief': True, '_service': None, '_cluster_spec': <tensorflow.python.training.server_lib.ClusterSpec object at 0x7fbf306e6ef0>, '_master': '', '_global_id_in_cluster': 0, '_num_ps_replicas': 0, '_task_type': 'worker', '_model_dir': 'taxi_trained', '_num_worker_replicas': 1, '_save_checkpoints_secs': 600, '_tf_random_seed': None, '_log_step_count_steps': 100, '_save_summary_steps': 100, '_task_id': 0, '_train_distribute': None, '_session_config': None, '_keep_checkpoint_max': 5} INFO:tensorflow:Calling model_fn. INFO:tensorflow:Done calling model_fn. INFO:tensorflow:Create CheckpointSaverHook. INFO:tensorflow:Graph was finalized. INFO:tensorflow:Running local_init_op. INFO:tensorflow:Done running local_init_op. INFO:tensorflow:Saving checkpoints for 1 into taxi_trained/model.ckpt. INFO:tensorflow:loss = 29842.77, step = 1 INFO:tensorflow:global_step/sec: 210.928 INFO:tensorflow:loss = 9002.392, step = 101 (0.481 sec) INFO:tensorflow:global_step/sec: 272.966 INFO:tensorflow:loss = 8324.836, step = 201 (0.365 sec) INFO:tensorflow:global_step/sec: 258.214 INFO:tensorflow:loss = 8521.68, step = 301 (0.387 sec) INFO:tensorflow:global_step/sec: 243.378 INFO:tensorflow:loss = 8178.3984, step = 401 (0.411 sec) INFO:tensorflow:global_step/sec: 278.198 INFO:tensorflow:loss = 9353.998, step = 501 (0.359 sec) INFO:tensorflow:global_step/sec: 246.354 INFO:tensorflow:loss = 12876.638, step = 601 (0.406 sec) INFO:tensorflow:global_step/sec: 263.237 INFO:tensorflow:loss = 7959.91, step = 701 (0.380 sec) INFO:tensorflow:global_step/sec: 253.149 INFO:tensorflow:loss = 8682.23, step = 801 (0.395 sec) INFO:tensorflow:global_step/sec: 197.901 INFO:tensorflow:loss = 12169.1, step = 901 (0.505 sec) INFO:tensorflow:global_step/sec: 239.283 INFO:tensorflow:loss = 8918.199, step = 1001 (0.419 sec) INFO:tensorflow:global_step/sec: 221.856 INFO:tensorflow:loss = 15695.652, step = 1101 (0.451 sec) INFO:tensorflow:global_step/sec: 241.302 INFO:tensorflow:loss = 5930.3394, step = 1201 (0.414 sec) INFO:tensorflow:global_step/sec: 227.39 INFO:tensorflow:loss = 11537.228, step = 1301 (0.440 sec) INFO:tensorflow:global_step/sec: 251.364 INFO:tensorflow:loss = 14412.359, step = 1401 (0.398 sec) INFO:tensorflow:global_step/sec: 256.601 INFO:tensorflow:loss = 17942.309, step = 1501 (0.393 sec) INFO:tensorflow:global_step/sec: 223.185 INFO:tensorflow:loss = 14170.855, step = 1601 (0.445 sec) INFO:tensorflow:global_step/sec: 247.665 INFO:tensorflow:loss = 16183.543, step = 1701 (0.403 sec) INFO:tensorflow:global_step/sec: 235.695 INFO:tensorflow:loss = 7810.726, step = 1801 (0.427 sec) INFO:tensorflow:global_step/sec: 234.881 INFO:tensorflow:loss = 6149.2754, step = 1901 (0.423 sec) INFO:tensorflow:global_step/sec: 232.834 INFO:tensorflow:loss = 11574.231, step = 2001 (0.429 sec) INFO:tensorflow:global_step/sec: 239.702 INFO:tensorflow:loss = 14717.008, step = 2101 (0.417 sec) INFO:tensorflow:global_step/sec: 237.606 INFO:tensorflow:loss = 9005.641, step = 2201 (0.421 sec) INFO:tensorflow:global_step/sec: 230.737 INFO:tensorflow:loss = 12744.685, step = 2301 (0.433 sec) INFO:tensorflow:global_step/sec: 260.326 INFO:tensorflow:loss = 13345.156, step = 2401 (0.384 sec) INFO:tensorflow:global_step/sec: 225.729 INFO:tensorflow:loss = 6819.132, step = 2501 (0.443 sec) INFO:tensorflow:global_step/sec: 258.975 INFO:tensorflow:loss = 8946.41, step = 2601 (0.386 sec) INFO:tensorflow:global_step/sec: 256.185 INFO:tensorflow:loss = 9691.637, step = 2701 (0.396 sec) INFO:tensorflow:global_step/sec: 226.054 INFO:tensorflow:loss = 9392.388, step = 2801 (0.437 sec) INFO:tensorflow:global_step/sec: 256.864 INFO:tensorflow:loss = 11355.194, step = 2901 (0.389 sec) INFO:tensorflow:global_step/sec: 247.804 INFO:tensorflow:loss = 10107.482, step = 3001 (0.403 sec) INFO:tensorflow:global_step/sec: 256.41 INFO:tensorflow:loss = 7983.508, step = 3101 (0.390 sec) INFO:tensorflow:global_step/sec: 246.065 INFO:tensorflow:loss = 15476.807, step = 3201 (0.406 sec) INFO:tensorflow:global_step/sec: 220.103 INFO:tensorflow:loss = 5103.0156, step = 3301 (0.454 sec) INFO:tensorflow:global_step/sec: 245.347 INFO:tensorflow:loss = 6437.1484, step = 3401 (0.407 sec) INFO:tensorflow:global_step/sec: 226.535 INFO:tensorflow:loss = 10064.092, step = 3501 (0.443 sec) INFO:tensorflow:global_step/sec: 253.54 INFO:tensorflow:loss = 15250.171, step = 3601 (0.393 sec) INFO:tensorflow:global_step/sec: 230.22 INFO:tensorflow:loss = 13603.143, step = 3701 (0.435 sec) INFO:tensorflow:global_step/sec: 235.699 INFO:tensorflow:loss = 17845.951, step = 3801 (0.424 sec) INFO:tensorflow:global_step/sec: 245.393 INFO:tensorflow:loss = 12553.659, step = 3901 (0.411 sec) INFO:tensorflow:global_step/sec: 215.13 INFO:tensorflow:loss = 12433.715, step = 4001 (0.461 sec) INFO:tensorflow:global_step/sec: 244.656 INFO:tensorflow:loss = 9361.515, step = 4101 (0.410 sec) INFO:tensorflow:global_step/sec: 228.47 INFO:tensorflow:loss = 9883.629, step = 4201 (0.436 sec) INFO:tensorflow:global_step/sec: 247.574 INFO:tensorflow:loss = 7398.719, step = 4301 (0.404 sec) INFO:tensorflow:global_step/sec: 232.703 INFO:tensorflow:loss = 6454.9185, step = 4401 (0.434 sec) INFO:tensorflow:global_step/sec: 255.993 INFO:tensorflow:loss = 7323.9966, step = 4501 (0.386 sec) INFO:tensorflow:global_step/sec: 269.523 INFO:tensorflow:loss = 9618.355, step = 4601 (0.371 sec) INFO:tensorflow:global_step/sec: 248.743 INFO:tensorflow:loss = 10290.853, step = 4701 (0.402 sec) INFO:tensorflow:global_step/sec: 261.354 INFO:tensorflow:loss = 5912.894, step = 4801 (0.382 sec) INFO:tensorflow:global_step/sec: 240.777 INFO:tensorflow:loss = 14062.918, step = 4901 (0.416 sec) INFO:tensorflow:global_step/sec: 244.952 INFO:tensorflow:loss = 14587.598, step = 5001 (0.411 sec) INFO:tensorflow:global_step/sec: 262.682 INFO:tensorflow:loss = 15330.396, step = 5101 (0.377 sec) INFO:tensorflow:global_step/sec: 240.852 INFO:tensorflow:loss = 13032.351, step = 5201 (0.415 sec) INFO:tensorflow:global_step/sec: 275.118 INFO:tensorflow:loss = 15112.017, step = 5301 (0.364 sec) INFO:tensorflow:global_step/sec: 270.734 INFO:tensorflow:loss = 13179.615, step = 5401 (0.369 sec) INFO:tensorflow:global_step/sec: 253.145 INFO:tensorflow:loss = 12603.519, step = 5501 (0.395 sec) INFO:tensorflow:global_step/sec: 273.417 INFO:tensorflow:loss = 3724.6326, step = 5601 (0.366 sec) INFO:tensorflow:global_step/sec: 245.797 INFO:tensorflow:loss = 17414.2, step = 5701 (0.407 sec) INFO:tensorflow:global_step/sec: 264.72 INFO:tensorflow:loss = 9366.868, step = 5801 (0.378 sec) INFO:tensorflow:global_step/sec: 254.335 INFO:tensorflow:loss = 10954.008, step = 5901 (0.393 sec) INFO:tensorflow:global_step/sec: 236.256 INFO:tensorflow:loss = 8344.541, step = 6001 (0.423 sec) INFO:tensorflow:Saving checkpoints for 6071 into taxi_trained/model.ckpt. INFO:tensorflow:Loss for final step: 4420.172. INFO:tensorflow:Calling model_fn. INFO:tensorflow:Done calling model_fn. INFO:tensorflow:Starting evaluation at 2018-11-21-03:42:10 INFO:tensorflow:Graph was finalized. INFO:tensorflow:Restoring parameters from taxi_trained/model.ckpt-6071 INFO:tensorflow:Running local_init_op. INFO:tensorflow:Done running local_init_op. INFO:tensorflow:Finished evaluation at 2018-11-21-03:42:10 INFO:tensorflow:Saving dict for global step 6071: average_loss = 109.05948, global_step = 6071, loss = 12970.288 RMSE on dataset = 10.443154335021973 ###Markdown We are not beating our benchmark with either model ... what's up? Well, we may be using TensorFlow for Machine Learning, but we are not yet using it well. That's what the rest of this course is about!But, for the record, let's say we had to choose between the two models. We'd choose the one with the lower validation error. Finally, we'd measure the RMSE on the test data with this chosen model. Benchmark dataset Let's do this on the benchmark dataset. ###Code import datalab.bigquery as bq import numpy as np import pandas as pd def create_query(phase, EVERY_N): """ phase: 1 = train 2 = valid """ base_query = """ SELECT (tolls_amount + fare_amount) AS fare_amount, CONCAT(STRING(pickup_datetime), STRING(pickup_longitude), STRING(pickup_latitude), STRING(dropoff_latitude), STRING(dropoff_longitude)) AS key, DAYOFWEEK(pickup_datetime)*1.0 AS dayofweek, HOUR(pickup_datetime)*1.0 AS hourofday, pickup_longitude AS pickuplon, pickup_latitude AS pickuplat, dropoff_longitude AS dropofflon, dropoff_latitude AS dropofflat, passenger_count*1.0 AS passengers, FROM [nyc-tlc:yellow.trips] WHERE trip_distance > 0 AND fare_amount >= 2.5 AND pickup_longitude > -78 AND pickup_longitude < -70 AND dropoff_longitude > -78 AND dropoff_longitude < -70 AND pickup_latitude > 37 AND pickup_latitude < 45 AND dropoff_latitude > 37 AND dropoff_latitude < 45 AND passenger_count > 0 """ if EVERY_N == None: if phase < 2: # Training query = "{0} AND ABS(HASH(pickup_datetime)) % 4 < 2".format(base_query) else: # Validation query = "{0} AND ABS(HASH(pickup_datetime)) % 4 == {1}".format(base_query, phase) else: query = "{0} AND ABS(HASH(pickup_datetime)) % {1} == {2}".format(base_query, EVERY_N, phase) return query query = create_query(2, 100000) df = bq.Query(query).to_dataframe() print_rmse(model, df) ###Output INFO:tensorflow:Calling model_fn. INFO:tensorflow:Done calling model_fn. INFO:tensorflow:Starting evaluation at 2018-11-21-03:42:52 INFO:tensorflow:Graph was finalized. INFO:tensorflow:Restoring parameters from taxi_trained/model.ckpt-6071 INFO:tensorflow:Running local_init_op. INFO:tensorflow:Done running local_init_op. INFO:tensorflow:Finished evaluation at 2018-11-21-03:42:53 INFO:tensorflow:Saving dict for global step 6071: average_loss = 88.57236, global_step = 6071, loss = 11257.773 RMSE on dataset = 9.41128921508789
Eye Detection.ipynb
###Markdown ROI Region Of Interestt ###Code import cv2 import numpy as np # Load the Haar cascade files for face and eye face_cascade = cv2.CascadeClassifier('haar_cascade_files/haarcascade_frontalface_default.xml') eye_cascade = cv2.CascadeClassifier('haar_cascade_files/haarcascade_eye.xml') # Check if the face cascade file has been loaded correctly if face_cascade.empty(): raise IOError('Unable to load the face cascade classifier xml file') # Check if the eye cascade file has been loaded correctly if eye_cascade.empty(): raise IOError('Unable to load the eye cascade classifier xml file') # Initialize the video capture object cap = cv2.VideoCapture(0) # Define the scaling factor ds_factor = 0.5 # Iterate until the user hits the 'Esc' key while True: # Capture the current frame _, frame = cap.read() # Resize the frame frame = cv2.resize(frame, None, fx=ds_factor, fy=ds_factor, interpolation=cv2.INTER_AREA) # Convert to grayscale gray = cv2.cvtColor(frame, cv2.COLOR_BGR2GRAY) # Run the face detector on the grayscale image faces = face_cascade.detectMultiScale(gray, 1.3, 5) # For each face that's detected, run the eye detector #xy,wh, son las coordenadas de la cara, #usando la imagen completa en gris, de ahi usas las coordenadas [gray] #Asi mismo con la de color [frame] #dentro de esta cara, se corre el cascade de ojos, y se pinta circulos verdes for (x,y,w,h) in faces: # Extract the grayscale face ROI roi_gray = gray[y:y+h, x:x+w] # Extract the color face ROI roi_color = frame[y:y+h, x:x+w] # Run the eye detector on the grayscale ROI eyes = eye_cascade.detectMultiScale(roi_gray) # Draw circles around the eyes for (x_eye,y_eye,w_eye,h_eye) in eyes: center = (int(x_eye + 0.5*w_eye), int(y_eye + 0.5*h_eye)) radius = int(0.3 * (w_eye + h_eye)) color = (0, 255, 0) thickness = 3 cv2.circle(roi_color, center, radius, color, thickness) # Display the output cv2.imshow('Eye Detector', frame) # Check if the user hit the 'Esc' key c = cv2.waitKey(1) if c == 27: break # Release the video capture object cap.release() # Close all the windows cv2.destroyAllWindows() ###Output _____no_output_____
Lec1.ipynb
###Markdown Edureka - Machine Learning ###Code from sklearn import datasets iris_dataset = datasets.load_iris() x = iris_dataset.data[:,:10] count = len(x.flat) print(x) print(count) min = x[:,0].min()-.5 max = x[:,0].max()+.5 count, min, max s="csdcd" a=9.5 print('%s%f',s,a) ###Output %s%f csdcd 9.5
2016/tutorial_final/55/PCA.ipynb
###Markdown Principal Component Analysis --- Gautam Arakalgud (garakalg) IntroductionThis tutorial is designed to give the reader an intuitive understanding of principal component analysis (PCA) and its applications in analyzing large datasets as well as data compression. While the tutorial will largely focus on the high level ideas behind PCA, it also gives you a chance to understand the mathematics behind the theory.In today's world, it is very important that we are able to use exisiting computing resources to store, analyze and work with growing data. For instance, we could be estimating facial expressions from images shot with a 1MP camera. Treating raw pixel values as image features, we have a million features for every image. Running standard machine learning algorithms on these images would require a great deal of computing power and time. However, it is often found that most features in high dimensional datasets are redundant and this _information_ can be contained in a much smaller number of features. PCA deals with effectively trying to identify those features that can capture the _essence_ of the data and represent the data in this lower dimensional feature space. ContentsThis tutorial will cover the following topics:- [The Olivetti Faces Dataset](The-Olivetti-Faces-Dataset)- [Visualizing the dataset](Visualizing-the-dataset)- [Math for PCA](Math-for-PCA)- [Computing eigen values and vectors](Computing-eigen-vectors-and-corresponding-eigen-values)- [Projecting data](Projecting-data-onto-the-principal-vectors-and-Reconstruction)- [Visualizing reconstructed data](Visualizing-reconstructed-images)- [Gram Matrix Trick](Gram-Matrix-Trick)- [Using PCA for classification](Using-Dimensionality-reduction-for-classification) The Olivetti Faces DatasetThe [Olivetti Database of Faces][odb] consists of 10 images each of 40 distinct subjects. The images were captured in natural environments at different times, with different lighting, facial expressions(smiling/not smiling) and facial details(glasses/no glasses).Each image in the dataset is of size 64x64 pixels. The image is quantized to 256 grayscale levels, but the loader converts this to floating point values in the interval [0 1]. We shall apply PCA and some basic (supervised) machine learning algorithms to this dataset and the see the effect of PCA on accuracy, and dependency on computing power and time. The dataset is loaded from the sklearn datasets.We shall first import the required libraries and load the database with the images arranged in a random permutation. This is achieved by setting "shuffle=True" while loading the dataset.[odb]: http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html ###Code import numpy as np from sklearn.preprocessing import normalize from sklearn.datasets import fetch_olivetti_faces from matplotlib import pyplot as plt import scipy.sparse as sp rnd = np.random.RandomState(2) # Use a fixed seed for consistency dataset = fetch_olivetti_faces(shuffle=True, random_state=rnd) # Load the Olivetti_Faces_dataset image_shape = (64, 64) ###Output _____no_output_____ ###Markdown Visualizing the datasetThe first step towards addressing any datascience problem must always be to visualize the dataset and understand it before running algorithms on it, and that is what is done here. A small helper function is written to plot a gallery of images in a 3x3 grid. We can then see a small preview of the kind of images this dataset contains. We plot 9 randomly chosen images from the dataset. ###Code def plot_gallery(title, subtitle, images, n_col, n_row, n): plt.figure(figsize=(2. * n_col, 2.26 * n_row)) plt.suptitle(title, size=16) for i, comp in enumerate(images): ax = plt.subplot(n_row, n_col, i + 1) vmax = max(comp.max(), -comp.min()) ax.set_title("{} {}".format(subtitle, (i+1)*n)) ax.imshow(comp.reshape(image_shape), cmap=plt.cm.gray) plt.xticks(()) plt.yticks(()) plt.subplots_adjust(0.01, 0.05, 0.99, 0.93, 0.20, 0.) plot_gallery("A preview of the Olivetti Faces Dataset", "Face", dataset.data[:9, :], 3, 3, 1) plt.show() ###Output _____no_output_____ ###Markdown Math for PCALet's take a small and quick detour to write down and intuitively understand some of the mathematics behind the PCA formulation. We can think of our data (images) as points in a high dimensional (64\*64 to be precise) space. We have 400 data points in a 4096 dimensional feature space. The core idea behind PCA is to pick **principle vectors** in the original high dimensional feature space and project our data onto these vectors, thus going down to a lower dimensional space (described by these vectors) while preserving most of the _important information_ in our data. The way PCA achieves this, is by picking out the vectors whose directions in the original feature space represent the maximum variance of the data. Hence PCA uses **variance** as a primary measure in compressing data into a smaller feature space. To apply PCA, we commonly calculate the covariance matrix of the data. Notation:$d$: Number of features $N$: Number of samples $X$: Data (Arranged dxN) $\omega$: Principal vector (1xd) $\mu$: Mean of the data (1xd)**Important**: Variance (or covariance) is always defined for data that is centered (meaned) at the origin. We must thus subtract the mean before running PCA on the data. Computing eigen vectors and corresponding eigen valuesThe PCA objective is that we want to find projections of the data (i.e directions onto which the data can be projected) that describe the maximum variance or scatter in the data. Let us look at this as an optimization problem. $$\begin{align*}max. \>\> J(\omega) & = var(\omega^T*X) \\& = E(\omega^TX - \omega^T\mu)^2 \\& = E(\omega^TX - \omega^T\mu)(X^T\omega - \mu^T \omega) \\& = \omega^TE[(X-\mu)(X-\mu)^T]\omega \\& = \omega^T \Sigma \omega\end{align*}$$where $\Sigma = E[(X-\mu)(X-\mu)^T]$ is the covariance matrix.We maximize $J(\omega)$ subject to the contstraint that the projection vectors are unit norm. ($\implies ||\omega||=1$)We solve the Legrangian formulation$$L(\omega, \lambda) = \omega^T\Sigma\omega - \lambda(\omega^T\omega - 1) \\\frac{\partial L(\omega, \lambda)}{\partial \omega} = 2\Sigma\omega - 2\lambda\omega = 0 \\\Sigma\omega = \lambda \omega \\$$This is a standard eigenvalue / eigen vector problem. The vectors $\omega$ are in fact the eigen vectors of the covariance matrix $\Sigma$ and the vectors corresponding to the highest eigen values are the ones that descibe the highest variation in the original data. Let us implement this with the face dataset. ###Code ''' eig() returns a sorted numpy array of eigen values and corresponding normalized eigen vectors as the columns of a square matrix. ''' def covariance_eig(data): covariance_data = data.dot(data.T) eig, V = np.linalg.eig(covariance_data) eig = np.real(eig) V = np.real(V) return (eig, V) img = dataset.data # Load the data in the format that we need img_mean = img.mean(axis=0) # We store the mean of the image because we need it to reconstruct the image. centered_img = img - img_mean # Remove the mean from every image. %time eig, V = covariance_eig(centered_img.T) # Calculate the covariance matrix and principal vectors of the data ###Output CPU times: user 9min 17s, sys: 4min 48s, total: 14min 5s Wall time: 4min 56s ###Markdown Now that we have our eigen vectors and their corresponding eigen values, we can pick the top _k_ eigen vectors that maximize the variance in the data and project our data along these vectors. By doing this, we are effectively reducing the dimentionality of our data from the original high dimensional space to k dimensions.The question of how many eigen vectors to pick is an important one. It largely depends on what purpose the PCA is trying to achieve. For instance, you could be performing PCA for data compression, for classification, etc. One good approach is to list the eigen values in descending order and calculate the amount of variance retained by picking the top k eigen vectors. This is known as the reconstruction capability and explains how well the chosen principal vectors can reconstruct the original data. By "good reconstruction", we usually mean reducing the mean square errors between the original and reconsructed data.$$Reconstruction \> \% = \frac{\sum_{i=1}^k \lambda_i}{\sum_{i=1}^N \lambda_i}$$We can write a small script that plots the reconstruction capability against the number of eigen vectors chosen. ###Code cumulative_eig = np.cumsum(eig) cumulative_eig = cumulative_eig/float(cumulative_eig[-1]) plt.plot(range(1, 201), cumulative_eig[:200]) plt.show() print "Reconstruction capability of the first 200 eigen vectors: {}%".format(cumulative_eig[199]*100) ###Output _____no_output_____ ###Markdown We notice something very interesting by just plotting the reconstruction capability for the first 200 (out of a potential 4096) eigen vectors. The first 200 principal vectors describe ~98% of the variance of the data. What this means is that we could reduce the dimentionality of our data from the original 4096 dimensional space to a 200 dimentional space with just 2% loss in information. We are effectively reducing the size of our data by almost 200 times! Sometimes, we may not even need these many principal vectors. To get a reconstruction capability of 80%, we'd need only 27 eigen vectors. Projecting data onto the principal vectors and ReconstructionWe can write a function that takes in as input the data matrix and the principal vectors and projects the data onto these vectors. ###Code def project(data, principal_vectors): data = data - img_mean project_data = data.dot(principal_vectors) return project_data ###Output _____no_output_____ ###Markdown Now that we have the new data projected onto the low dimensional feature space, we can store or run algorithms on this low compressed data. The next step should be to visually see close the compressed data is to the actual data and calculate the mean square error between the original and reconstructed data.**Important:** We subtracted the mean from the original image and projected the centered_image. Remember to add the mean back to the reconstructed data. ###Code def reconstruct(project_data, principal_vectors): reconstruct_data = project_data.dot(principal_vectors.T) reconstruct_data = reconstruct_data + img_mean return reconstruct_data ###Output _____no_output_____ ###Markdown Visualizing reconstructed imagesWe begin by visualizing the most dominant principal vectors and see what kind of information these vectors contain. We continue this analysis by choosing an image and projecting it on the first k principal vectors and vary k. ###Code eigen_components = np.zeros((9, dataset.data.shape[1])) eigen_components = V[:, :9].T plot_gallery("Principal components of the data", "Eigen vector", eigen_components, 3, 3, 1) plt.show() dom_eigen = np.zeros((9, dataset.data.shape[1])) print V.shape print dataset.data[0].shape for i in xrange(9): project_data = project(dataset.data[0], V[:, :5*(i+1)]) reconstruct_data = reconstruct(project_data, V[:, :5*(i+1)]) dom_eigen[i] = reconstruct_data plot_gallery("Image reconstruction with k principal components", "k =", dom_eigen, 3, 3, 5) plt.show() ###Output (4096, 4096) (4096,) ###Markdown While it seems obvious that increasing the number of principal vectors gives us better reconstruction, we can visually see that this is true. MSE of reconstructed imagesA good metric for the amount of information loss is to calculate the mean square error between the original and reconstructed images. We take 4 cases, by choosing 5, 20, 50 and 400 principal vectors. ###Code rec_images = np.zeros((4, dataset.data.shape[1])) principal_vec = [5, 20, 50, 400] for i,k in enumerate(principal_vec): project_data = project(dataset.data[0], V[:, :k]) reconstruct_data = reconstruct(project_data, V[:, :k]) rec_images[i] = reconstruct_data mse = np.sum(((rec_images-dataset.data[0])**2), axis=1)/float(rec_images.shape[1]) print mse ###Output [ 9.46010545e-03 4.56969746e-03 2.24048652e-03 4.96826193e-14] ###Markdown We get good reconstruction with k = 5, 20 and 50 and near perfect reconstruction by using 400 eigen vectors (as expected). Gram Matrix TrickIn this example we are working with images of size 64x64 pixels, and this gives us feature vectors of size 4096 (d). However, we just have 400 (N) sample images. This is most often the case when we deal with computer vision problems. The feature size far exceeds the number of samples we have. The covariance matrix that we compute is correspondingly of size 4096x4096. As the size of these images increases, it becomes computationally infeasible to compute these large matrices. For instance, if we were dealing with 1MP images, we'd be generating a $10^6*10^6$ covaraince matrix, and we wouldn't be able to store a matrix this large on our computer. Added to this is the fact that we have only N samples (N << d) and can have at most N-1 eigenvectors (assuming that our data is linearly independent). Hence, it would make more sense to calculate an $N*N$ covariance matrix and work with this matrix.We know that $$\Sigma = E[(X-\mu)(X-\mu)^T]$$We must solve for $$\begin{align*}\Sigma v & = \lambda v \\XX^Tv & = \lambda v \\X^TXX^Tv & = \lambda X^Tv \> \> \text{(pre-mult by }X^T) \\\text{Setting }(v' & = X^Tv), \\X^TXv' & = \lambda v' \implies \text{Eigen value problem}\end{align*}$$However, once we solve for the eigen values and eigen vectors, we must get back the principal vectors of the original data, as these are the vectors on which we can project our data onto.$$\begin{align*}XX^Tv & = \lambda v \\v' & = X^Tv \\\implies Xv' & = \lambda v \\\implies v & = \frac{1}{\lambda}Xv'\end{align*}$$We can now go ahead and solve for the new covariance matrix and the corresponding principal vectors. ###Code gram_img = centered_img print gram_img.shape %time eig_gram, V_gram = covariance_eig(gram_img) print V_gram.shape ###Output (400, 4096) CPU times: user 1.9 s, sys: 3.04 s, total: 4.94 s Wall time: 1.46 s (400, 400) ###Markdown We can compare the time that it took for the computation of the covariance and eigen vector matrices for the original data and using the gram matrix trick. Using the gram matrix trick and computing the covariance matrix and principal components for the transposed data matrix took 1s as against 2.5 minutes for computing it for the original data. This is a pretty big speedup. Further, we're now just storing 400 principal vectors rather than the original 4096 principal vectors. It's easy to see why this computation is beneficial. We can now project and reconstruct the same image that we had earlier to see if we get similar results. Recovering original eigen vectorsBefore we can project our data, we must recover the original principal vectors. This is done by following the equation derived above. ###Code def recover(data, V_gram): V_orig = data.T.dot(V_gram) V_orig = normalize(V_orig, axis=0) return V_orig ###Output _____no_output_____ ###Markdown We can verify that the recovered eigen vectors from the gram matrix trick are identical to the original principal vectors generated from the data itself. In short, V_orig must be identical to V. We can resontruct the same image that we'd chosen earlier to verifiy this. ###Code dom_eigen = np.zeros((9, dataset.data.shape[1])) for i in xrange(9): project_data = project(dataset.data[0], V_orig[:, :5*(i+1)]) reconstruct_data = reconstruct(project_data, V_orig[:, :5*(i+1)]) dom_eigen[i] = reconstruct_data plot_gallery("Image reconstruction with k principal components", "k =", dom_eigen, 3, 3, 5) plt.show() ###Output _____no_output_____ ###Markdown Using Dimensionality reduction for classificationLet us now train and build a classification algorithm to classify these images. The dataset contains 10 images each of 40 distinct people and we can use it to perform face recognition. We use the softmax loss for this multiclass classification in a similar fashion as we did to classify handwritten digits in assignment 4.Let us define the loss and gradient functions ###Code def softmax_loss_grad(X, Theta, y): m = y.shape[0] y_mat = sp.csr_matrix((np.ones(m), (y, np.arange(m)))) y_mat = np.array(y_mat.todense()).T hyp = X.dot(Theta) prob = ((np.exp(hyp)).T/(np.sum(np.exp(hyp), axis=1))).T loss = -sum((sum(y_mat*np.log(prob))))/m grad = (-1/float(m))*X.T.dot(y_mat-prob) return (loss, grad) def softmax_gd(X, y, lam=1e-5, alpha=1.0): theta = np.zeros((X.shape[1], len(np.unique(y)))) prev_loss = np.inf while True: loss, grad = softmax_loss_grad(X, theta, y) grad += lam*theta theta = theta - alpha*grad if abs(prev_loss-loss) < 1e-5: break; prev_loss = loss; return theta def predict(X_test, theta): hyp = X_test.dot(theta) pred = hyp.argmax(axis=1) return pred ###Output _____no_output_____ ###Markdown We can use 80% (320 images) as our training set and the remaining 20% (80 images) as our testing set. With PCA: ###Code project_data = project(dataset.data, V[:, :50]) X_train = project_data[:320] X_test = project_data[320:400] y_train = dataset.target[:320] y_test = dataset.target[320:400] theta = softmax_gd(X_train, y_train) pred = predict(X_test, theta) err = len(np.where(pred != y_test)[0])/float(len(y_test)) print err ###Output 0.025 ###Markdown Without PCA: ###Code X_train = dataset.data[:320] X_test = dataset.data[320:400] theta = softmax_gd(X_train, y_train, alpha=0.05) pred = predict(X_test, theta) err = len(np.where(pred != y_test)[0])/float(len(y_test)) print err ###Output 0.022
tv-script-generation/dlnd_tv_script_generation-solution.ipynb
###Markdown TV Script GenerationIn this project, you'll generate your own [Simpsons](https://en.wikipedia.org/wiki/The_Simpsons) TV scripts using RNNs. You'll be using part of the [Simpsons dataset](https://www.kaggle.com/wcukierski/the-simpsons-by-the-data) of scripts from 27 seasons. The Neural Network you'll build will generate a new TV script for a scene at [Moe's Tavern](https://simpsonswiki.com/wiki/Moe's_Tavern). Get the DataThe data is already provided for you. You'll be using a subset of the original dataset. It consists of only the scenes in Moe's Tavern. This doesn't include other versions of the tavern, like "Moe's Cavern", "Flaming Moe's", "Uncle Moe's Family Feed-Bag", etc.. ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ import helper data_dir = './data/simpsons/moes_tavern_lines.txt' text = helper.load_data(data_dir) # Ignore notice, since we don't use it for analysing the data text = text[81:] ###Output _____no_output_____ ###Markdown Explore the DataPlay around with `view_sentence_range` to view different parts of the data. ###Code view_sentence_range = (0, 10) """ DON'T MODIFY ANYTHING IN THIS CELL """ import numpy as np print('Dataset Stats') print('Roughly the number of unique words: {}'.format(len({word: None for word in text.split()}))) scenes = text.split('\n\n') print('Number of scenes: {}'.format(len(scenes))) sentence_count_scene = [scene.count('\n') for scene in scenes] print('Average number of sentences in each scene: {}'.format(np.average(sentence_count_scene))) sentences = [sentence for scene in scenes for sentence in scene.split('\n')] print('Number of lines: {}'.format(len(sentences))) word_count_sentence = [len(sentence.split()) for sentence in sentences] print('Average number of words in each line: {}'.format(np.average(word_count_sentence))) print() print('The sentences {} to {}:'.format(*view_sentence_range)) print('\n'.join(text.split('\n')[view_sentence_range[0]:view_sentence_range[1]])) ###Output Dataset Stats Roughly the number of unique words: 11492 Number of scenes: 262 Average number of sentences in each scene: 15.248091603053435 Number of lines: 4257 Average number of words in each line: 11.50434578341555 The sentences 0 to 10: Moe_Szyslak: (INTO PHONE) Moe's Tavern. Where the elite meet to drink. Bart_Simpson: Eh, yeah, hello, is Mike there? Last name, Rotch. Moe_Szyslak: (INTO PHONE) Hold on, I'll check. (TO BARFLIES) Mike Rotch. Mike Rotch. Hey, has anybody seen Mike Rotch, lately? Moe_Szyslak: (INTO PHONE) Listen you little puke. One of these days I'm gonna catch you, and I'm gonna carve my name on your back with an ice pick. Moe_Szyslak: What's the matter Homer? You're not your normal effervescent self. Homer_Simpson: I got my problems, Moe. Give me another one. Moe_Szyslak: Homer, hey, you should not drink to forget your problems. Barney_Gumble: Yeah, you should only drink to enhance your social skills. ###Markdown Implement Preprocessing FunctionsThe first thing to do to any dataset is preprocessing. Implement the following preprocessing functions below:- Lookup Table- Tokenize Punctuation Lookup TableTo create a word embedding, you first need to transform the words to ids. In this function, create two dictionaries:- Dictionary to go from the words to an id, we'll call `vocab_to_int`- Dictionary to go from the id to word, we'll call `int_to_vocab`Return these dictionaries in the following tuple `(vocab_to_int, int_to_vocab)` ###Code import numpy as np import problem_unittests as tests def create_lookup_tables(text): """ Create lookup tables for vocabulary :param text: The text of tv scripts split into words :return: A tuple of dicts (vocab_to_int, int_to_vocab) """ # TODO: Implement Function vocab_to_int = dict() int_to_vocab = dict() words = set(text) for i, word in enumerate(words): vocab_to_int[word] = i int_to_vocab[i] = word return (vocab_to_int, int_to_vocab) """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_create_lookup_tables(create_lookup_tables) ###Output Tests Passed ###Markdown Tokenize PunctuationWe'll be splitting the script into a word array using spaces as delimiters. However, punctuations like periods and exclamation marks make it hard for the neural network to distinguish between the word "bye" and "bye!".Implement the function `token_lookup` to return a dict that will be used to tokenize symbols like "!" into "||Exclamation_Mark||". Create a dictionary for the following symbols where the symbol is the key and value is the token:- Period ( . )- Comma ( , )- Quotation Mark ( " )- Semicolon ( ; )- Exclamation mark ( ! )- Question mark ( ? )- Left Parentheses ( ( )- Right Parentheses ( ) )- Dash ( -- )- Return ( \n )This dictionary will be used to token the symbols and add the delimiter (space) around it. This separates the symbols as it's own word, making it easier for the neural network to predict on the next word. Make sure you don't use a token that could be confused as a word. Instead of using the token "dash", try using something like "||dash||". ###Code def token_lookup(): """ Generate a dict to turn punctuation into a token. :return: Tokenize dictionary where the key is the punctuation and the value is the token """ # TODO: Implement Function return { '.':'||PERIOD||', ',':'||COMMA||', '"':'||QUOTATION_MARK||', ';':'||SEMICOLON||', '!':'||EXCLAMATION_MARK||', '?':'||QUESTION_MARK||', '(':'||LEFT_PARANTHESES||', ')':'||RIGHT_PARANTHESES||', '--':'||DASH||', '\n':'||RETURN||' } """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_tokenize(token_lookup) ###Output Tests Passed ###Markdown Preprocess all the data and save itRunning the code cell below will preprocess all the data and save it to file. ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ # Preprocess Training, Validation, and Testing Data helper.preprocess_and_save_data(data_dir, token_lookup, create_lookup_tables) ###Output _____no_output_____ ###Markdown Check PointThis is your first checkpoint. If you ever decide to come back to this notebook or have to restart the notebook, you can start from here. The preprocessed data has been saved to disk. ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ import helper import numpy as np import problem_unittests as tests int_text, vocab_to_int, int_to_vocab, token_dict = helper.load_preprocess() ###Output _____no_output_____ ###Markdown Build the Neural NetworkYou'll build the components necessary to build a RNN by implementing the following functions below:- get_inputs- get_init_cell- get_embed- build_rnn- build_nn- get_batches Check the Version of TensorFlow and Access to GPU ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ from distutils.version import LooseVersion import warnings import tensorflow as tf # Check TensorFlow Version assert LooseVersion(tf.__version__) >= LooseVersion('1.0'), 'Please use TensorFlow version 1.0 or newer' print('TensorFlow Version: {}'.format(tf.__version__)) # Check for a GPU if not tf.test.gpu_device_name(): warnings.warn('No GPU found. Please use a GPU to train your neural network.') else: print('Default GPU Device: {}'.format(tf.test.gpu_device_name())) ###Output TensorFlow Version: 1.1.0-rc1 ###Markdown InputImplement the `get_inputs()` function to create TF Placeholders for the Neural Network. It should create the following placeholders:- Input text placeholder named "input" using the [TF Placeholder](https://www.tensorflow.org/api_docs/python/tf/placeholder) `name` parameter.- Targets placeholder- Learning Rate placeholderReturn the placeholders in the following tuple `(Input, Targets, LearningRate)` ###Code def get_inputs(): """ Create TF Placeholders for input, targets, and learning rate. :return: Tuple (input, targets, learning rate) """ # TODO: Implement Function inputs = tf.placeholder(tf.int32, [None, None], name='input') targets = tf.placeholder(tf.int32, [None, None], name='targets') learn_rate = tf.placeholder(tf.float32, name='learning_rate') return inputs, targets, learn_rate """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_get_inputs(get_inputs) ###Output Tests Passed ###Markdown Build RNN Cell and InitializeStack one or more [`BasicLSTMCells`](https://www.tensorflow.org/api_docs/python/tf/contrib/rnn/BasicLSTMCell) in a [`MultiRNNCell`](https://www.tensorflow.org/api_docs/python/tf/contrib/rnn/MultiRNNCell).- The Rnn size should be set using `rnn_size`- Initalize Cell State using the MultiRNNCell's [`zero_state()`](https://www.tensorflow.org/api_docs/python/tf/contrib/rnn/MultiRNNCellzero_state) function - Apply the name "initial_state" to the initial state using [`tf.identity()`](https://www.tensorflow.org/api_docs/python/tf/identity)Return the cell and initial state in the following tuple `(Cell, InitialState)` ###Code def get_init_cell(batch_size, rnn_size): """ Create an RNN Cell and initialize it. :param batch_size: Size of batches :param rnn_size: Size of RNNs :return: Tuple (cell, initialize state) """ # TODO: Implement Function lstm = tf.contrib.rnn.BasicLSTMCell(rnn_size) cell = tf.contrib.rnn.MultiRNNCell([lstm]) initial_state = cell.zero_state(batch_size, tf.float32) return cell, tf.identity(initial_state, name='initial_state') """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_get_init_cell(get_init_cell) ###Output Tests Passed ###Markdown Word EmbeddingApply embedding to `input_data` using TensorFlow. Return the embedded sequence. ###Code def get_embed(input_data, vocab_size, embed_dim): """ Create embedding for <input_data>. :param input_data: TF placeholder for text input. :param vocab_size: Number of words in vocabulary. :param embed_dim: Number of embedding dimensions :return: Embedded input. """ # TODO: Implement Function embedding = tf.Variable(tf.random_uniform((vocab_size, embed_dim), -1, 1)) embed = tf.nn.embedding_lookup(embedding, input_data) return embed """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_get_embed(get_embed) ###Output Tests Passed ###Markdown Build RNNYou created a RNN Cell in the `get_init_cell()` function. Time to use the cell to create a RNN.- Build the RNN using the [`tf.nn.dynamic_rnn()`](https://www.tensorflow.org/api_docs/python/tf/nn/dynamic_rnn) - Apply the name "final_state" to the final state using [`tf.identity()`](https://www.tensorflow.org/api_docs/python/tf/identity)Return the outputs and final_state state in the following tuple `(Outputs, FinalState)` ###Code def build_rnn(cell, inputs): """ Create a RNN using a RNN Cell :param cell: RNN Cell :param inputs: Input text data :return: Tuple (Outputs, Final State) """ # TODO: Implement Function outputs, final_state = tf.nn.dynamic_rnn(cell, inputs, dtype=tf.float32) final_state = tf.identity(final_state, name='final_state') return outputs, final_state """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_build_rnn(build_rnn) ###Output Tests Passed ###Markdown Build the Neural NetworkApply the functions you implemented above to:- Apply embedding to `input_data` using your `get_embed(input_data, vocab_size, embed_dim)` function.- Build RNN using `cell` and your `build_rnn(cell, inputs)` function.- Apply a fully connected layer with a linear activation and `vocab_size` as the number of outputs.Return the logits and final state in the following tuple (Logits, FinalState) ###Code def fully_conn_linear(x_tensor, num_outputs): num_inputs = x_tensor.shape.as_list() weights = tf.Variable(tf.truncated_normal(num_inputs + [num_outputs], stddev=0.1)) bias = tf.Variable(tf.zeros(num_outputs)) # fc_layer = tf.reshape(x_tensor, [-1, num_inputs]) fc_layer = tf.add(tf.matmul(x_tensor, weights), bias) return fc_layer def build_nn(cell, rnn_size, input_data, vocab_size, embed_dim): """ Build part of the neural network :param cell: RNN cell :param rnn_size: Size of rnns :param input_data: Input data :param vocab_size: Vocabulary size :param embed_dim: Number of embedding dimensions :return: Tuple (Logits, FinalState) """ # TODO: Implement Function embedding = get_embed(input_data, vocab_size, embed_dim) outputs, final_state = build_rnn(cell, embedding) logits = tf.layers.dense(outputs, vocab_size, activation=None) return logits, final_state """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_build_nn(build_nn) ###Output Tests Passed ###Markdown BatchesImplement `get_batches` to create batches of input and targets using `int_text`. The batches should be a Numpy array with the shape `(number of batches, 2, batch size, sequence length)`. Each batch contains two elements:- The first element is a single batch of **input** with the shape `[batch size, sequence length]`- The second element is a single batch of **targets** with the shape `[batch size, sequence length]`If you can't fill the last batch with enough data, drop the last batch.For exmple, `get_batches([1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20], 3, 2)` would return a Numpy array of the following:```[ First Batch [ Batch of Input [[ 1 2], [ 7 8], [13 14]] Batch of targets [[ 2 3], [ 8 9], [14 15]] ] Second Batch [ Batch of Input [[ 3 4], [ 9 10], [15 16]] Batch of targets [[ 4 5], [10 11], [16 17]] ] Third Batch [ Batch of Input [[ 5 6], [11 12], [17 18]] Batch of targets [[ 6 7], [12 13], [18 1]] ]]```Notice that the last target value in the last batch is the first input value of the first batch. In this case, `1`. This is a common technique used when creating sequence batches, although it is rather unintuitive. ###Code # Note, this was my first attempt and to be honest I don't quite understand the batching. # As you should note below is that my code does produce the example output, then also passes # the majority of unit tests, but fails where it tests for the last item. # I think the unit tests are not designed correctly, because I do set the last target to # the first input. The problem is along a different axis which is not covered by the unit test # and neither the example explains what's wrong. # I think in the example it is a bit unfortunate that you have 3x2 batch_size and seq_length, but also # when I tried to figure out why after [1 2] it follows [7 8] the only reason I could find is because it's # continuing the sequence along the first vertical [1] [2] [3] [4] [5] [6], which happens to be 6=2*3. # In the unit test this "shift" is 35, which I don't understand and so the formulas don't work. # When I looked up the error in the forums I saw a non-working code which just had the very last element # wrong so I used and fixed that, see next code box... def get_batches(int_text, batch_size, seq_length, test=False): """ Return batches of input and target :param int_text: Text with the words replaced by their ids :param batch_size: The size of batch :param seq_length: The length of sequence :return: Batches as a Numpy array """ # TODO: Implement Function N = len(int_text) count = 3 if test else int(N/batch_size/seq_length) delta = 6 if test else 35 x = [] print('N',N) print('batch_size',batch_size) print('seq_length',seq_length) print('delta',delta) print('count',count) for c in range(count): i = 2*c batch = [] for b in range(batch_size): idx = i+b*delta batch += int_text[ idx : idx+seq_length ] for b in range(batch_size): idx = i+b*delta+1 batch += int_text[ idx : idx+seq_length ] if (c == count-1): batch[-1] = int_text[0] #print(batch) x += batch print(len(x)) batches = np.array(x).flatten().reshape((count, 2, batch_size, seq_length)) return batches print(get_batches([1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20], 3, 2, test=True)) """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_get_batches(get_batches) def get_batches(int_text, batch_size, seq_length): """ Return batches of input and target :param int_text: Text with the words replaced by their ids :param batch_size: The size of batch :param seq_length: The length of sequence :return: Batches as a Numpy array """ # TODO: Implement Function n_batches = int(len(int_text) / (batch_size * seq_length)) # Drop the last few characters to make only full batches xdata = np.array(int_text[: n_batches * batch_size * seq_length]) ydata = np.array(int_text[1: n_batches * batch_size * seq_length + 1]) ydata[-1] = xdata[0] x_batches = np.split(xdata.reshape(batch_size, -1), n_batches, 1) y_batches = np.split(ydata.reshape(batch_size, -1), n_batches, 1) return np.array(list(zip(x_batches, y_batches))) """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_get_batches(get_batches) ###Output Tests Passed ###Markdown Neural Network Training HyperparametersTune the following parameters:- Set `num_epochs` to the number of epochs.- Set `batch_size` to the batch size.- Set `rnn_size` to the size of the RNNs.- Set `embed_dim` to the size of the embedding.- Set `seq_length` to the length of sequence.- Set `learning_rate` to the learning rate.- Set `show_every_n_batches` to the number of batches the neural network should print progress. ###Code # Number of Epochs num_epochs = 50 # Batch Size batch_size = 256 # RNN Size rnn_size = 512 # Embedding Dimension Size embed_dim = 200 # Sequence Length seq_length = 20 # Learning Rate learning_rate = 0.01 # Show stats for every n number of batches show_every_n_batches = 13 """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ save_dir = './save' ###Output _____no_output_____ ###Markdown Build the GraphBuild the graph using the neural network you implemented. ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ from tensorflow.contrib import seq2seq train_graph = tf.Graph() with train_graph.as_default(): vocab_size = len(int_to_vocab) input_text, targets, lr = get_inputs() input_data_shape = tf.shape(input_text) cell, initial_state = get_init_cell(input_data_shape[0], rnn_size) logits, final_state = build_nn(cell, rnn_size, input_text, vocab_size, embed_dim) # Probabilities for generating words probs = tf.nn.softmax(logits, name='probs') # Loss function cost = seq2seq.sequence_loss( logits, targets, tf.ones([input_data_shape[0], input_data_shape[1]])) # Optimizer optimizer = tf.train.AdamOptimizer(lr) # Gradient Clipping gradients = optimizer.compute_gradients(cost) capped_gradients = [(tf.clip_by_value(grad, -1., 1.), var) for grad, var in gradients if grad is not None] train_op = optimizer.apply_gradients(capped_gradients) ###Output _____no_output_____ ###Markdown TrainTrain the neural network on the preprocessed data. If you have a hard time getting a good loss, check the [forms](https://discussions.udacity.com/) to see if anyone is having the same problem. ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ batches = get_batches(int_text, batch_size, seq_length) with tf.Session(graph=train_graph) as sess: sess.run(tf.global_variables_initializer()) for epoch_i in range(num_epochs): state = sess.run(initial_state, {input_text: batches[0][0]}) for batch_i, (x, y) in enumerate(batches): feed = { input_text: x, targets: y, initial_state: state, lr: learning_rate} train_loss, state, _ = sess.run([cost, final_state, train_op], feed) # Show every <show_every_n_batches> batches if (epoch_i * len(batches) + batch_i) % show_every_n_batches == 0: print('Epoch {:>3} Batch {:>4}/{} train_loss = {:.3f}'.format( epoch_i, batch_i, len(batches), train_loss)) # Save Model saver = tf.train.Saver() saver.save(sess, save_dir) print('Model Trained and Saved') ###Output Epoch 0 Batch 0/13 train_loss = 8.823 Epoch 1 Batch 0/13 train_loss = 5.566 Epoch 2 Batch 0/13 train_loss = 4.892 Epoch 3 Batch 0/13 train_loss = 4.435 Epoch 4 Batch 0/13 train_loss = 4.004 Epoch 5 Batch 0/13 train_loss = 3.604 Epoch 6 Batch 0/13 train_loss = 3.283 Epoch 7 Batch 0/13 train_loss = 3.027 Epoch 8 Batch 0/13 train_loss = 2.738 Epoch 9 Batch 0/13 train_loss = 2.463 Epoch 10 Batch 0/13 train_loss = 2.253 Epoch 11 Batch 0/13 train_loss = 2.056 Epoch 12 Batch 0/13 train_loss = 1.887 Epoch 13 Batch 0/13 train_loss = 1.730 Epoch 14 Batch 0/13 train_loss = 1.547 Epoch 15 Batch 0/13 train_loss = 1.432 Epoch 16 Batch 0/13 train_loss = 1.352 Epoch 17 Batch 0/13 train_loss = 1.184 Epoch 18 Batch 0/13 train_loss = 1.012 Epoch 19 Batch 0/13 train_loss = 0.921 Epoch 20 Batch 0/13 train_loss = 0.795 Epoch 21 Batch 0/13 train_loss = 0.692 Epoch 22 Batch 0/13 train_loss = 0.615 Epoch 23 Batch 0/13 train_loss = 0.534 Epoch 24 Batch 0/13 train_loss = 0.474 Epoch 25 Batch 0/13 train_loss = 0.432 Epoch 26 Batch 0/13 train_loss = 0.388 Epoch 27 Batch 0/13 train_loss = 0.357 Epoch 28 Batch 0/13 train_loss = 0.327 Epoch 29 Batch 0/13 train_loss = 0.295 Epoch 30 Batch 0/13 train_loss = 0.279 Epoch 31 Batch 0/13 train_loss = 0.247 Epoch 32 Batch 0/13 train_loss = 0.225 Epoch 33 Batch 0/13 train_loss = 0.211 Epoch 34 Batch 0/13 train_loss = 0.192 Epoch 35 Batch 0/13 train_loss = 0.183 Epoch 36 Batch 0/13 train_loss = 0.180 Epoch 37 Batch 0/13 train_loss = 0.175 Epoch 38 Batch 0/13 train_loss = 0.167 Epoch 39 Batch 0/13 train_loss = 0.166 Epoch 40 Batch 0/13 train_loss = 0.162 Epoch 41 Batch 0/13 train_loss = 0.156 Epoch 42 Batch 0/13 train_loss = 0.152 Epoch 43 Batch 0/13 train_loss = 0.150 Epoch 44 Batch 0/13 train_loss = 0.148 Epoch 45 Batch 0/13 train_loss = 0.147 Epoch 46 Batch 0/13 train_loss = 0.147 Epoch 47 Batch 0/13 train_loss = 0.145 Epoch 48 Batch 0/13 train_loss = 0.145 Epoch 49 Batch 0/13 train_loss = 0.144 Model Trained and Saved ###Markdown Save ParametersSave `seq_length` and `save_dir` for generating a new TV script. ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ # Save parameters for checkpoint helper.save_params((seq_length, save_dir)) ###Output _____no_output_____ ###Markdown Checkpoint ###Code """ DON'T MODIFY ANYTHING IN THIS CELL """ import tensorflow as tf import numpy as np import helper import problem_unittests as tests _, vocab_to_int, int_to_vocab, token_dict = helper.load_preprocess() seq_length, load_dir = helper.load_params() ###Output _____no_output_____ ###Markdown Implement Generate Functions Get TensorsGet tensors from `loaded_graph` using the function [`get_tensor_by_name()`](https://www.tensorflow.org/api_docs/python/tf/Graphget_tensor_by_name). Get the tensors using the following names:- "input:0"- "initial_state:0"- "final_state:0"- "probs:0"Return the tensors in the following tuple `(InputTensor, InitialStateTensor, FinalStateTensor, ProbsTensor)` ###Code def get_tensors(loaded_graph): """ Get input, initial state, final state, and probabilities tensor from <loaded_graph> :param loaded_graph: TensorFlow graph loaded from file :return: Tuple (InputTensor, InitialStateTensor, FinalStateTensor, ProbsTensor) """ # TODO: Implement Function return loaded_graph.get_tensor_by_name(name='input:0'), \ loaded_graph.get_tensor_by_name(name='initial_state:0'), \ loaded_graph.get_tensor_by_name(name='final_state:0'), \ loaded_graph.get_tensor_by_name(name='probs:0') """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_get_tensors(get_tensors) ###Output Tests Passed ###Markdown Choose WordImplement the `pick_word()` function to select the next word using `probabilities`. ###Code def pick_word(probabilities, int_to_vocab): """ Pick the next word in the generated text :param probabilities: Probabilites of the next word :param int_to_vocab: Dictionary of word ids as the keys and words as the values :return: String of the predicted word """ # TODO: Implement Function p = np.squeeze(probabilities) p = p/np.sum(p) c = np.random.choice(len(p), 1, p=p)[0] return int_to_vocab[c] """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ tests.test_pick_word(pick_word) ###Output Tests Passed ###Markdown Generate TV ScriptThis will generate the TV script for you. Set `gen_length` to the length of TV script you want to generate. ###Code gen_length = 200 # homer_simpson, moe_szyslak, or Barney_Gumble prime_word = 'moe_szyslak' """ DON'T MODIFY ANYTHING IN THIS CELL THAT IS BELOW THIS LINE """ loaded_graph = tf.Graph() with tf.Session(graph=loaded_graph) as sess: # Load saved model loader = tf.train.import_meta_graph(load_dir + '.meta') loader.restore(sess, load_dir) # Get Tensors from loaded model input_text, initial_state, final_state, probs = get_tensors(loaded_graph) # Sentences generation setup gen_sentences = [prime_word + ':'] prev_state = sess.run(initial_state, {input_text: np.array([[1]])}) # Generate sentences for n in range(gen_length): # Dynamic Input dyn_input = [[vocab_to_int[word] for word in gen_sentences[-seq_length:]]] dyn_seq_length = len(dyn_input[0]) # Get Prediction probabilities, prev_state = sess.run( [probs, final_state], {input_text: dyn_input, initial_state: prev_state}) pred_word = pick_word(probabilities[dyn_seq_length-1], int_to_vocab) gen_sentences.append(pred_word) # Remove tokens tv_script = ' '.join(gen_sentences) for key, token in token_dict.items(): ending = ' ' if key in ['\n', '(', '"'] else '' tv_script = tv_script.replace(' ' + token.lower(), key) tv_script = tv_script.replace('\n ', '\n') tv_script = tv_script.replace('( ', '(') print(tv_script) ###Output INFO:tensorflow:Restoring parameters from ./save moe_szyslak: hey, hey, hey, hey! that plank's only for comin' in! moe_szyslak:(gently) you want to buy the best japanese yellow we're homer. lisa_simpson:(thrilled) the wordloaf festival doesn't do! moe_szyslak:(to homer) easy there, habitrail. lisa_simpson:(quickly) i'm not here. moe_szyslak: that's the worst name i ever heard. barney_gumble:(calling after him) hey, joey joe joe junior... oh, nuts. i forgot. all i can think of hope. moe_szyslak: nah, nah, no. makin' polenta less be in here. waylon_smithers: huh? huh? listen, what if i helped you say heavyweight championship. kent_brockman:(to himself, disgusted) my thesaurus. homer_simpson: quit changing the subject. how do you feel about me right now? moe_szyslak: oh, no. oh lisa. homer_simpson: oh man, i love our valentine's day to play with me, i just hope you do. homer_simpson: chief! thank god, i always
Copy_of_Hello_ML_World.ipynb
###Markdown The Hello World of Deep Learning with Neural Networks Like every first app you should start with something super simple that shows the overall scaffolding for how your code works. In the case of creating neural networks, the sample I like to use is one where it learns the relationship between two numbers. So, for example, if you were writing code for a function like this, you already know the 'rules' -- ```float my_function(float x){``` This is formatted as code``` float y = (3 * x)+1; return y;}```---``` This is formatted as code```So how would you train a neural network to do the equivalent task? Using data! By feeding it with a set of Xs, and a set of Ys, it should be able to figure out the relationship between them. This is obviously a very different paradigm than what you might be used to, so let's step through it piece by piece. New Section ImportsLet's start with our imports. Here we are importing TensorFlow and calling it tf for ease of use.We then import a library called numpy, which helps us to represent our data as lists easily and quickly.The framework for defining a neural network as a set of Sequential layers is called keras, so we import that too. ###Code import tensorflow as tf import numpy as np from tensorflow import keras ###Output _____no_output_____ ###Markdown Define and Compile the Neural NetworkNext we will create the simplest possible neural network. It has 1 layer, and that layer has 1 neuron, and the input shape to it is just 1 value. ###Code model = tf.keras.Sequential([keras.layers.Dense(units=1, input_shape=[1])]) ###Output _____no_output_____ ###Markdown Now we compile our Neural Network. When we do so, we have to specify 2 functions, a loss and an optimizer.If you've seen lots of math for machine learning, here's where it's usually used, but in this case it's nicely encapsulated in functions for you. But what happens here -- let's explain...We know that in our function, the relationship between the numbers is y=3x+1. When the computer is trying to 'learn' that, it makes a guess...maybe y=10x+10. The LOSS function measures the guessed answers against the known correct answers and measures how well or how badly it did.It then uses the OPTIMIZER function to make another guess. Based on how the loss function went, it will try to minimize the loss. At that point maybe it will come up with somehting like y=5x+5, which, while still pretty bad, is closer to the correct result (i.e. the loss is lower)It will repeat this for the number of EPOCHS which you will see shortly. But first, here's how we tell it to use 'MEAN SQUARED ERROR' for the loss and 'STOCHASTIC GRADIENT DESCENT' for the optimizer. You don't need to understand the math for these yet, but you can see that they work! :)Over time you will learn the different and appropriate loss and optimizer functions for different scenarios. ###Code model.compile(optimizer='sgd', loss='mean_squared_error') ###Output _____no_output_____ ###Markdown Providing the DataNext up we'll feed in some data. In this case we are taking 6 xs and 6ys. You can see that the relationship between these is that y=2x-1, so where x = -1, y=-3 etc. etc. A python library called 'Numpy' provides lots of array type data structures that are a defacto standard way of doing it. We declare that we want to use these by specifying the values asn an np.array[] ###Code xs = np.array([-1.0, 0.0, 1.0, 2.0, 3.0, 4.0,-2.0,-3.0,-4.0], dtype=float) ys = np.array([-1.0, 0.0, 3.0, 5.0, 7.0, 9.0,-3.0,-5.0,-7.0], dtype=float) ###Output _____no_output_____ ###Markdown Training the Neural Network The process of training the neural network, where it 'learns' the relationship between the Xs and Ys is in the **model.fit** call. This is where it will go through the loop we spoke about above, making a guess, measuring how good or bad it is (aka the loss), using the opimizer to make another guess etc. It will do it for the number of epochs you specify. When you run this code, you'll see the loss on the right hand side. ###Code model.fit(xs, ys, epochs=1000) ###Output Epoch 1/1000 1/1 [==============================] - 0s 1ms/step - loss: 5.5134 Epoch 2/1000 1/1 [==============================] - 0s 828us/step - loss: 4.3312 Epoch 3/1000 1/1 [==============================] - 0s 1ms/step - loss: 3.4366 Epoch 4/1000 1/1 [==============================] - 0s 1ms/step - loss: 2.7584 Epoch 5/1000 1/1 [==============================] - 0s 1ms/step - loss: 2.2429 Epoch 6/1000 1/1 [==============================] - 0s 1ms/step - loss: 1.8499 Epoch 7/1000 1/1 [==============================] - 0s 1ms/step - loss: 1.5492 Epoch 8/1000 1/1 [==============================] - 0s 1ms/step - loss: 1.3180 Epoch 9/1000 1/1 [==============================] - 0s 847us/step - loss: 1.1391 Epoch 10/1000 1/1 [==============================] - 0s 945us/step - loss: 0.9999 Epoch 11/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.8906 Epoch 12/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.8039 Epoch 13/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.7344 Epoch 14/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.6780 Epoch 15/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.6317 Epoch 16/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.5930 Epoch 17/1000 1/1 [==============================] - 0s 743us/step - loss: 0.5602 Epoch 18/1000 1/1 [==============================] - 0s 793us/step - loss: 0.5320 Epoch 19/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.5073 Epoch 20/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.4856 Epoch 21/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.4660 Epoch 22/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.4483 Epoch 23/1000 1/1 [==============================] - 0s 718us/step - loss: 0.4321 Epoch 24/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.4171 Epoch 25/1000 1/1 [==============================] - 0s 773us/step - loss: 0.4032 Epoch 26/1000 1/1 [==============================] - 0s 724us/step - loss: 0.3901 Epoch 27/1000 1/1 [==============================] - 0s 975us/step - loss: 0.3778 Epoch 28/1000 1/1 [==============================] - 0s 760us/step - loss: 0.3662 Epoch 29/1000 1/1 [==============================] - 0s 854us/step - loss: 0.3552 Epoch 30/1000 1/1 [==============================] - 0s 706us/step - loss: 0.3447 Epoch 31/1000 1/1 [==============================] - 0s 709us/step - loss: 0.3347 Epoch 32/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.3252 Epoch 33/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.3161 Epoch 34/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.3074 Epoch 35/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.2991 Epoch 36/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.2911 Epoch 37/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.2834 Epoch 38/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2761 Epoch 39/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2690 Epoch 40/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2623 Epoch 41/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2558 Epoch 42/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2495 Epoch 43/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2436 Epoch 44/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2378 Epoch 45/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2323 Epoch 46/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.2270 Epoch 47/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.2219 Epoch 48/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.2171 Epoch 49/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.2124 Epoch 50/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.2079 Epoch 51/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.2036 Epoch 52/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1994 Epoch 53/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1954 Epoch 54/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1916 Epoch 55/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1879 Epoch 56/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1844 Epoch 57/1000 1/1 [==============================] - 0s 994us/step - loss: 0.1810 Epoch 58/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1777 Epoch 59/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1746 Epoch 60/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1716 Epoch 61/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1687 Epoch 62/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1660 Epoch 63/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1633 Epoch 64/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1607 Epoch 65/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1583 Epoch 66/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1559 Epoch 67/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1537 Epoch 68/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1515 Epoch 69/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1494 Epoch 70/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1474 Epoch 71/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1455 Epoch 72/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1436 Epoch 73/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1418 Epoch 74/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1401 Epoch 75/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1385 Epoch 76/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1369 Epoch 77/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1354 Epoch 78/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1340 Epoch 79/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1326 Epoch 80/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1312 Epoch 81/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1299 Epoch 82/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1287 Epoch 83/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1275 Epoch 84/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1264 Epoch 85/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1253 Epoch 86/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1242 Epoch 87/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1232 Epoch 88/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1223 Epoch 89/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1213 Epoch 90/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1204 Epoch 91/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1196 Epoch 92/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1188 Epoch 93/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1180 Epoch 94/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1172 Epoch 95/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1165 Epoch 96/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1158 Epoch 97/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1151 Epoch 98/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1145 Epoch 99/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1138 Epoch 100/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1132 Epoch 101/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1127 Epoch 102/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1121 Epoch 103/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1116 Epoch 104/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1111 Epoch 105/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1106 Epoch 106/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1101 Epoch 107/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1097 Epoch 108/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1092 Epoch 109/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1088 Epoch 110/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1084 Epoch 111/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1080 Epoch 112/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1077 Epoch 113/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1073 Epoch 114/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1070 Epoch 115/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1067 Epoch 116/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1063 Epoch 117/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1060 Epoch 118/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1058 Epoch 119/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1055 Epoch 120/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1052 Epoch 121/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1050 Epoch 122/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1047 Epoch 123/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1045 Epoch 124/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1043 Epoch 125/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1040 Epoch 126/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1038 Epoch 127/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1036 Epoch 128/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1034 Epoch 129/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1032 Epoch 130/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1031 Epoch 131/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1029 Epoch 132/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1027 Epoch 133/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1026 Epoch 134/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1024 Epoch 135/1000 1/1 [==============================] - 0s 955us/step - loss: 0.1023 Epoch 136/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1021 Epoch 137/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1020 Epoch 138/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1019 Epoch 139/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1018 Epoch 140/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1016 Epoch 141/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1015 Epoch 142/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1014 Epoch 143/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1013 Epoch 144/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1012 Epoch 145/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1011 Epoch 146/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1010 Epoch 147/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1009 Epoch 148/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1008 Epoch 149/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1008 Epoch 150/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1007 Epoch 151/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1006 Epoch 152/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1005 Epoch 153/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1005 Epoch 154/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1004 Epoch 155/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1003 Epoch 156/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1003 Epoch 157/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1002 Epoch 158/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1002 Epoch 159/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1001 Epoch 160/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.1000 Epoch 161/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.1000 Epoch 162/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0999 Epoch 163/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0999 Epoch 164/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0999 Epoch 165/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0998 Epoch 166/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0998 Epoch 167/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0997 Epoch 168/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0997 Epoch 169/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0997 Epoch 170/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0996 Epoch 171/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0996 Epoch 172/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0996 Epoch 173/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0995 Epoch 174/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0995 Epoch 175/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0995 Epoch 176/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0994 Epoch 177/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0994 Epoch 178/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0994 Epoch 179/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0994 Epoch 180/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0993 Epoch 181/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0993 Epoch 182/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0993 Epoch 183/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0993 Epoch 184/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0993 Epoch 185/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0992 Epoch 186/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0992 Epoch 187/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0992 Epoch 188/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0992 Epoch 189/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0992 Epoch 190/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0991 Epoch 191/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0991 Epoch 192/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0991 Epoch 193/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0991 Epoch 194/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0991 Epoch 195/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0991 Epoch 196/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0991 Epoch 197/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0991 Epoch 198/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 199/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 200/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 201/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 202/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 203/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 204/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 205/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0990 Epoch 206/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0990 Epoch 207/1000 1/1 [==============================] - 0s 8ms/step - loss: 0.0990 Epoch 208/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 209/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0989 Epoch 210/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0989 Epoch 211/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 212/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0989 Epoch 213/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 214/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 215/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 216/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0989 Epoch 217/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 218/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 219/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0989 Epoch 220/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 221/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 222/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 223/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 224/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 225/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 226/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 227/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0989 Epoch 228/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 229/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 230/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 231/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 232/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 233/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 234/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 235/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 236/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 237/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 238/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 239/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 240/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 241/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 242/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 243/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 244/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 245/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 246/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 247/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 248/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 249/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 250/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 251/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 252/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 253/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 254/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 255/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 256/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 257/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 258/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 259/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 260/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 261/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 262/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 263/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 264/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 265/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 266/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 267/1000 1/1 [==============================] - 0s 5ms/step - loss: 0.0988 Epoch 268/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 269/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 270/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 271/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 272/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 273/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 274/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 275/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 276/1000 1/1 [==============================] - 0s 933us/step - loss: 0.0988 Epoch 277/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 278/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 279/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 280/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 281/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 282/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 283/1000 1/1 [==============================] - 0s 996us/step - loss: 0.0988 Epoch 284/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 285/1000 1/1 [==============================] - 0s 948us/step - loss: 0.0988 Epoch 286/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 287/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 288/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 289/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 290/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 291/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 292/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 293/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 294/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 295/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 296/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 297/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 298/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 299/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 300/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 301/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 302/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 303/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 304/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 305/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 306/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 307/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 308/1000 1/1 [==============================] - 0s 936us/step - loss: 0.0988 Epoch 309/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 310/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 311/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 312/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 313/1000 1/1 [==============================] - 0s 902us/step - loss: 0.0988 Epoch 314/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 315/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 316/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 317/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 318/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 319/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 320/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 321/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 322/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 323/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 324/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 325/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 326/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 327/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 328/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 329/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 330/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 331/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 332/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 333/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 334/1000 1/1 [==============================] - 0s 902us/step - loss: 0.0988 Epoch 335/1000 1/1 [==============================] - 0s 817us/step - loss: 0.0988 Epoch 336/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 337/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 338/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 339/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 340/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 341/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 342/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 343/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 344/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 345/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 346/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 347/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 348/1000 1/1 [==============================] - 0s 5ms/step - loss: 0.0988 Epoch 349/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 350/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 351/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 352/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 353/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 354/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 355/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 356/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 357/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 358/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 359/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 360/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 361/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 362/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 363/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 364/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 365/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 366/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 367/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 368/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 369/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 370/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 371/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 372/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 373/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 374/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 375/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 376/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 377/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 378/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 379/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 380/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 381/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 382/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 383/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 384/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 385/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 386/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 387/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 388/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 389/1000 1/1 [==============================] - 0s 858us/step - loss: 0.0988 Epoch 390/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 391/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 392/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 393/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 394/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 395/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 396/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 397/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 398/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 399/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 400/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 401/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 402/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 403/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 404/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 405/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 406/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 407/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.0988 Epoch 408/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 409/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 410/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 411/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 412/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 413/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 414/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 415/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 416/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 417/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 418/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 419/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 420/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 421/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 422/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 423/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 424/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 425/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 426/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 427/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 428/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 429/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 430/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 431/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 432/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 433/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 434/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 435/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 436/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 437/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 438/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 439/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 440/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 441/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 442/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 443/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 444/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 445/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 446/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 447/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 448/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 449/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 450/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.0988 Epoch 451/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 452/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 453/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 454/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 455/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 456/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 457/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 458/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 459/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 460/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 461/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 462/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 463/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 464/1000 1/1 [==============================] - 0s 777us/step - loss: 0.0988 Epoch 465/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 466/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 467/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 468/1000 1/1 [==============================] - 0s 799us/step - loss: 0.0988 Epoch 469/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 470/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 471/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 472/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 473/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 474/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 475/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 476/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 477/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 478/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 479/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 480/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 481/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 482/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 483/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 484/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 485/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.0988 Epoch 486/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 487/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 488/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 489/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 490/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 491/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 492/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 493/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 494/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 495/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 496/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 497/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 498/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 499/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 500/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 501/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 502/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 503/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 504/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 505/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 506/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 507/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 508/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 509/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 510/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 511/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 512/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 513/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 514/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 515/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 516/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 517/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 518/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 519/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 520/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 521/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 522/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 523/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 524/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 525/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 526/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 527/1000 1/1 [==============================] - 0s 6ms/step - loss: 0.0988 Epoch 528/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 529/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 530/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 531/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 532/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 533/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 534/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 535/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 536/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 537/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 538/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 539/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 540/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 541/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 542/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 543/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 544/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 545/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.0988 Epoch 546/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 547/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 548/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 549/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 550/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 551/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 552/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 553/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 554/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 555/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 556/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 557/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 558/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 559/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 560/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 561/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 562/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 563/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 564/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 565/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 566/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 567/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 568/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 569/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 570/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 571/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 572/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 573/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 574/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 575/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 576/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 577/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 578/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 579/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 580/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 581/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 582/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 583/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 584/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 585/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 586/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 587/1000 1/1 [==============================] - 0s 848us/step - loss: 0.0988 Epoch 588/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 589/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 590/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 591/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 592/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 593/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 594/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 595/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 596/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 597/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 598/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.0988 Epoch 599/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 600/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 601/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 602/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 603/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 604/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 605/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 606/1000 1/1 [==============================] - 0s 5ms/step - loss: 0.0988 Epoch 607/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 608/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 609/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 610/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 611/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 612/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 613/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 614/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 615/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 616/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 617/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 618/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 619/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 620/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 621/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 622/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 623/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 624/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 625/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 626/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 627/1000 1/1 [==============================] - 0s 874us/step - loss: 0.0988 Epoch 628/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 629/1000 1/1 [==============================] - 0s 914us/step - loss: 0.0988 Epoch 630/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 631/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 632/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 633/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 634/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 635/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 636/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 637/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 638/1000 1/1 [==============================] - 0s 999us/step - loss: 0.0988 Epoch 639/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 640/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 641/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 642/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 643/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 644/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 645/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 646/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 647/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 648/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 649/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 650/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 651/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 652/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 653/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 654/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 655/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 656/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 657/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 658/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 659/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 660/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 661/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 662/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 663/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 664/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 665/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 666/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 667/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 668/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 669/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 670/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 671/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 672/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 673/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 674/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 675/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 676/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 677/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 678/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 679/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 680/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 681/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 682/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 683/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 684/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 685/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 686/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 687/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 688/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 689/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 690/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 691/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 692/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 693/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 694/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 695/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 696/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 697/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 698/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 699/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 700/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 701/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 702/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 703/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 704/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 705/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 706/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 707/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 708/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 709/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 710/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 711/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 712/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 713/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 714/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 715/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 716/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 717/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 718/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 719/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 720/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 721/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 722/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 723/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 724/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 725/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 726/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 727/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 728/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 729/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 730/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 731/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 732/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 733/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 734/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 735/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 736/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 737/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 738/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 739/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 740/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 741/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 742/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 743/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 744/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 745/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 746/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 747/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 748/1000 1/1 [==============================] - 0s 6ms/step - loss: 0.0988 Epoch 749/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 750/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 751/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 752/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 753/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 754/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 755/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 756/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 757/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 758/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 759/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 760/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 761/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 762/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 763/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 764/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 765/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 766/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 767/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 768/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 769/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 770/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 771/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 772/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 773/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 774/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 775/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 776/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 777/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 778/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 779/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 780/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 781/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 782/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 783/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 784/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 785/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 786/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 787/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 788/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 789/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 790/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 791/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 792/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 793/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 794/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 795/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 796/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 797/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 798/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 799/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 800/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 801/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 802/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 803/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 804/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 805/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 806/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 807/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 808/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 809/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 810/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 811/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 812/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 813/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 814/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 815/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 816/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 817/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 818/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 819/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 820/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 821/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 822/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 823/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 824/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 825/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 826/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 827/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 828/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 829/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 830/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 831/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 832/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 833/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 834/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 835/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 836/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 837/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 838/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 839/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 840/1000 1/1 [==============================] - 0s 856us/step - loss: 0.0988 Epoch 841/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 842/1000 1/1 [==============================] - 0s 803us/step - loss: 0.0988 Epoch 843/1000 1/1 [==============================] - 0s 962us/step - loss: 0.0988 Epoch 844/1000 1/1 [==============================] - 0s 849us/step - loss: 0.0988 Epoch 845/1000 1/1 [==============================] - 0s 884us/step - loss: 0.0988 Epoch 846/1000 1/1 [==============================] - 0s 866us/step - loss: 0.0988 Epoch 847/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 848/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 849/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 850/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 851/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 852/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 853/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 854/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 855/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 856/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 857/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 858/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 859/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 860/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 861/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 862/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 863/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 864/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 865/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 866/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 867/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 868/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 869/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 870/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 871/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 872/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 873/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 874/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 875/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 876/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 877/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 878/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 879/1000 1/1 [==============================] - 0s 4ms/step - loss: 0.0988 Epoch 880/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 881/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 882/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 883/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 884/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 885/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 886/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 887/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 888/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 889/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 890/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 891/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 892/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 893/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 894/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 895/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 896/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 897/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 898/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 899/1000 1/1 [==============================] - 0s 7ms/step - loss: 0.0988 Epoch 900/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 901/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 902/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 903/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 904/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 905/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 906/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 907/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 908/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 909/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 910/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 911/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 912/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 913/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 914/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 915/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 916/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 917/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 918/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 919/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 920/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 921/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 922/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 923/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 924/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 925/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 926/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 927/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 928/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 929/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 930/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 931/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 932/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 933/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 934/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 935/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 936/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 937/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 938/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 939/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 940/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 941/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 942/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 943/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 944/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 945/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 946/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 947/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 948/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 949/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 950/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 951/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 952/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 953/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 954/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 955/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 956/1000 1/1 [==============================] - 0s 895us/step - loss: 0.0988 Epoch 957/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 958/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 959/1000 1/1 [==============================] - 0s 944us/step - loss: 0.0988 Epoch 960/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 961/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 962/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 963/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 964/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 965/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 966/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 967/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 968/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 969/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 970/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 971/1000 1/1 [==============================] - 0s 6ms/step - loss: 0.0988 Epoch 972/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 973/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 974/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 975/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 976/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 977/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 978/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 979/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 980/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 981/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 982/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 983/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 984/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 985/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 986/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 987/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 988/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 989/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 990/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 991/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 992/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 993/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 994/1000 1/1 [==============================] - 0s 3ms/step - loss: 0.0988 Epoch 995/1000 1/1 [==============================] - 0s 7ms/step - loss: 0.0988 Epoch 996/1000 1/1 [==============================] - 0s 1ms/step - loss: 0.0988 Epoch 997/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 998/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 999/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 Epoch 1000/1000 1/1 [==============================] - 0s 2ms/step - loss: 0.0988 ###Markdown Ok, now you have a model that has been trained to learn the relationshop between X and Y. You can use the **model.predict** method to have it figure out the Y for a previously unknown X. So, for example, if X = 10, what do you think Y will be? Take a guess before you run this code: ###Code print(model.predict([21.0])) ###Output [[42.88888]]
notebooks/.ipynb_checkpoints/0-Background-checkpoint.ipynb
###Markdown Simple Transverse WavesHere we will consider plane waves moving in opposite direction. Wave 1: $E1=cos(k x-\omega t)$ &rarr; Wave 2: $E2=cos(-k x -\omega t)$ &larr; Addition of the two wave: $E_r=cos(k x - \omega t) + cos(-k x - \omega t)$ $ = 2cos(kx)cos(-\omega t)$ Average square of the total field (detectable power): $\langle E_r^2 \rangle = 2+2cos(2kx)$ Standing Wave Combination of Two Waves Demonstation ###Code # First set up the figure, the axis, and the plot element we want to animate fig, axs = plt.subplots(5,figsize=(16,8)) for ax in axs: ax.set_xlim(( -4, 4)) ax.set_ylim((-1, 1)) #set limit for E_sum E_T^2 and E_T^2 average. axs[2].set_ylim((-2,2)) axs[3].set_ylim((0,4)) axs[4].set_ylim((0,1)) plt.xlabel('microns') axs[0].set_ylabel('E1',rotation=0,labelpad=10) axs[1].set_ylabel('E2',rotation=0,labelpad=10) axs[2].set_ylabel(r'$E_T = E1 + E2$',rotation=0,labelpad=25) axs[3].set_ylabel(r'$E_T^2',rotation=0,labelpad=25) axs[4].set_ylabel(r'$E_T^2 Average',rotation=0,labelpad=45) line1, = axs[0].plot([], [], lw=2) line2, = axs[1].plot([], [], lw=2) line3, = axs[2].plot([], [], lw=2) line4, = axs[3].plot([], [], lw=2) line5, = axs[4].plot([], [], lw=2) # initialization function: plot the background of each frame def init(): line1.set_data([], []) line2.set_data([], []) line3.set_data([], []) line4.set_data([], []) line5.set_data([], []) return (line1, line2, line3, line4, line5,) # animation function. This is called sequentially def animate(i): x = np.linspace(-4, 4, 1000) y1 = np.cos(2 * np.pi * (x - 0.04 * i)) line1.set_data(x, y1) y2 = np.cos(2 * np.pi * (-x - 0.04 * i)) line2.set_data(x, y2) y3 = y1+y2 line3.set_data(x,y3) y4 =y3**2 line4.set_data(x,y4) y5 = (2+2*np.cos(4*np.pi*x))/4 #Normalized Average line5.set_data(x,y5) return (line1,line2,line3,line4,line5,) # call the animator. blit=True means only re-draw the parts that # have changed. anim = animation.FuncAnimation(fig, animate, init_func=init, frames=25, interval=40, blit=True) #Can show by running "anim" anim.save('img/combination-of-two-waves.gif', writer='imagemagick', fps=30) Image(url='img/combination-of-two-waves.gif') ###Output _____no_output_____
.ipynb_checkpoints/B+Tree - The good one-checkpoint.ipynb
###Markdown B+Tree ###Code # El módulo abc obliga a las clases derivadas a implemtar # un método particular utilizando @abstractmethod from abc import ABCMeta, abstractmethod # Permite trabajar mejor con listas, insertar conservando el orden import bisect from math import floor as floor class Node(metaclass=ABCMeta): def __init__(self): self.num = 0 self.keys = [] @abstractmethod def getLoc(self, key): """Return null si no split. Si no retorna la info del split.""" pass @abstractmethod def insert(self, key,value): pass @abstractmethod def display(self): pass class Split: # TODO: Refactor Split def __init__(self, key, node_leaf, node_right): self.key = key self.left = node_leaf self.right = node_right class LNode(Node): def __init__(self): super(LNode, self).__init__(); self.values = [] def getLoc(self, key): for i, key_i in enumerate(self.keys): if key_i >= key: return i # Si no se encontró una posición retorna la posición final return len(self.keys) #return bisect.bisect_left(self.keys, key) def insert(self, key, value): i = self.getLoc(key) if self.num == M: # Nodo llego, se debe dividir mid = floor((M + 1) / 2) sNum = self.num - mid # Creamos un nodo Hermano, y a este le asignamos la mitad de los elementos sibling = LNode() sibling.num = sNum sibling.keys = self.keys[mid:] sibling.values = self.values[mid:] self.keys = self.keys[:mid] self.values = self.values[:mid] self.num = mid if i < mid: self.insertNonFull(key, value, i) else: sibling.insertNonFull(key, value, i-mid) # notificar al nodo padre result = Split(sibling.keys[0], self, sibling) return result else: self.insertNonFull(key, value, i) return None def insertNonFull(self, key, value, i): #print("key: {}, val: {}, i: {}".format(key, value, i)) self.keys.insert(i, key) self.values.insert(i, value) self.num += 1 def display(self): print('\t<LNode>\t', end='') for key, val in zip(self.keys, self.values): print("{} -> {}\t".format(key, val), end='') print('') def showChildren(self): print("Children: {}".format(len(self.values))) print("Children: {}".format(self.num)) print("") class INode(Node): def __init__(self): super(INode, self).__init__(); self.children = [] def getLoc(self, key): for i, key_i in enumerate(self.keys): if key_i >= key: return i # Si no se encontró una posición retorna la posición final return len(self.keys) #return bisect.bisect_left(self.keys, key) def insert(self, key, value): if self.num == N: # Dividir mid = floor((N + 1) / 2) sNum = self.num - mid sibling = INode() sibling.num = sNum sibling.keys = self.keys[mid:] sibling.children = self.children[mid:] self.keys = self.keys[:mid] self.children = self.children[:mid] self.num = mid - 1 # El elemento del extremo izq se envia a la parte superior result = Split(self.keys[mid-1], self, sibling) # insertar en el lado apropiado if key < result.key: # menor que cero self.insertNonFull(key, value) else: sibling.insertNonFull(key, value) return result else: self.insertNonFull(key, value) return None def insertNonFull(self, key, value): i = self.getLoc(key) result = self.children[i].insert(key, value) if result is not None: # Caso contrario no se debe modicar nada if i == self.num: # Insertamos a la derecha #self.keys[i] = result.key #self.children[i] = result.left #self.children[i+1] = result.right self.keys.insert(i, result.key) self.children.insert(i, result.left) self.children.insert(i+1, result.right) self.num += 1 else: self.children.insert(i, result.left) self.children.insert(i + 1, result.right) self.keys.insert(i, result.key) self.num += 1; def display(self): print("Displaying INode") print('<INode>\t', end='') for key in self.keys: print('{}\t'.format(key), end='') print("") for child in self.children: if isinstance(child, INode): print("") child.display() def showChildren(self): print("Children INode: {}".format(len(self.children))) print("Children INode: {}".format(self.num)) self.display() print("") class BTree: def __init__(self, degree): self.max_leafs = degree - 1 self.max_inner_nodes = degree M = self.max_leafs N = self.max_inner_nodes self.root = LNode() def insert(self, key, value): print("BEFORE ADD") self.root.showChildren() print("Insertar [{}]={}".format(key, value)) result = self.root.insert(key, value) if result is not None: # Se dividió la raiz # Se crea una nueva raiz _root = INode() _root.num = 2 _root.keys.insert(0, result.key) _root.children.insert(0, result.left) _root.children.insert(1, result.right) self.root = _root print("AFTER ADD") self.root.showChildren() print("") print("- - -") print("") def find(self, key): node = self.root while isinstance(node, INode): idx = node.getLoc(key) node = node.children[idx] # Estamos en la hoja idx = node.getLoc(key) if idx < node.num and node.keys[idx] == key: return node.values[idx] else: return None def display(self): self.root.display() tree = BTree(3) # Maximo número de hojas M = 2; # Maximo número de nodos en nodo intermedio N = 3; tree.insert(1, "1111") tree.insert(2, "2222") tree.insert(4, "4444") tree.insert(3, "3333") print(tree.find(1)) tree.display() tree.insert(5, "5555") tree.display() tree.insert(6, "6666") tree.display() tree.insert(9, "9999") tree.display() ###Output Displaying INode <INode> 3 5 Displaying INode <INode> 2 3 <LNode> 1 -> 1111 <LNode> 2 -> 2222 Displaying INode <INode> 4 5 <LNode> 3 -> 3333 <LNode> 4 -> 4444 Displaying INode <INode> 6 <LNode> 5 -> 5555 <LNode> 6 -> 6666 9 -> 9999 <LNode> 5 -> 5555 <LNode> 4 -> 4444 <LNode> 3 -> 3333 <LNode> 2 -> 2222 Displaying INode <INode> 4 5 <LNode> 3 -> 3333 <LNode> 4 -> 4444
Natural Language Processing/4-Topic-Modelling.ipynb
###Markdown Topic Modeling IntroductionTopic modeling is another popular text analysis technique. The goal of topic modeling is to find various topics that are present in your corpus. Each document in the corpus will be made up of at least one topic, if not multiple topics.In this notebook, I will apply Latent Dirichlet Allocation (LDA), which is one of many topic modeling techniques. It was specifically designed for text data.Latent means hidden (topics) and Dirichlet refers to a probability distrubution technique. To use a topic modeling technique, we need to provide1. a document-term matrix and 2. the number of topics you would like the algorithm to pick up.Once the topic modeling technique is applied, the job is to interpret the results and see if the mix of words in each topic make sense. If they don't make sense, we can try changing up the number of topics, the terms in the document-term matrix, model parameters, or even try a different model. Topic Modeling - Attempt 1 (All Text) ###Code # reading the document term matrix import pandas as pd import pickle data = pd.read_pickle('pickle files/dtm_stop.pkl') data.head() # importing necessary modules for LDA with gensim from gensim import matutils, models import scipy.sparse # we can also use log to debug that helps in later hyperpaarameter tuning # import logging # logging.basicConfig(format = '%(asctime)s : %(levelname)s : %(message)s',level = logging.INFO) # we need term document matrix tdm = data.transpose() tdm.head() # putting tdm into a new gensim format # df -----> sparse matrix ----->gensim corpus sparse_count = scipy.sparse.csr_matrix(tdm) corpus = matutils.Sparse2Corpus(sparse_count) # Gensim also requires dictionary of all the terms and # their respective location in the tdm cv = pickle.load(open("pickle files/cv_stop.pkl", "rb")) id2word = dict((value, key) for key, value in cv.vocabulary_.items()) ###Output _____no_output_____ ###Markdown Now that we have the corpus(term document matrix) and id2word(dictionary of (location: term), we need to specify two more parameters - the number of topics and the numbers of passes(iterations). Let's take the number of topics as 2, and check the results. ###Code # specifying no. of topics ad no. of passes lda = models.LdaModel(corpus = corpus, id2word = id2word, num_topics = 2, passes = 10) lda.print_topics() # increasing number of topics for better insights lda = models.LdaModel(corpus = corpus, id2word = id2word, num_topics = 3, passes = 10) lda.print_topics() ###Output _____no_output_____ ###Markdown Clearly, there are no meaningful results.This time I had modified the parameters, now I have tried to modify the term list as well. Topic modeling - Attempt 2(Nouns only) ###Code # function to pull oout all the nouns from the text string from nltk import word_tokenize, pos_tag def nouns(text): # string of text ----> tokenize the text ----> pull out nouns is_noun = lambda pos: pos[:2] == 'NN' tokenized = word_tokenize(text) all_nouns = [word for (word, pos) in pos_tag(tokenized) if is_noun(pos)] return ' '.join(all_nouns) # loading the cleaned text data before count vectorization data_clean = pd.read_pickle('pickle files/clean.pkl') data_clean # applying nouns function to the trascripts to filter only on nouns data_nouns = pd.DataFrame(data_clean.transcript.apply(nouns)) data_nouns # creating a new document-term matrix using only nouns from sklearn.feature_extraction import text from sklearn.feature_extraction.text import CountVectorizer # re- adding stop_words since we are creating new dtm # defining some additional stop words add_stop_words = ['like', 'im', 'know', 'just', 'dont', 'thats', 'right', 'people', 'youre', 'got', 'gonna', 'time', 'think', 'yeah', 'said'] stop_words = text.ENGLISH_STOP_WORDS.union(add_stop_words) # creating a dtm with only nouns cvn = CountVectorizer(stop_words = stop_words) model = cvn.fit_transform(data_nouns.transcript) data_new_dtm = pd.DataFrame(model.toarray(), columns = cvn.get_feature_names()) data_new_dtm.index = data_nouns.index data_new_dtm # creating gensim corpus new_corpus = matutils.Sparse2Corpus(scipy.sparse.csr_matrix(data_new_dtm.transpose())) # creating vocabulary dictionary new_id2word = dict((v, k) for k, v in cvn.vocabulary_.items()) # starting with no of topics = 2 new_lda = models.LdaModel(corpus = new_corpus, id2word = new_id2word, num_topics = 2 ,passes = 10) new_lda.print_topics() # no of topics = 3 new_lda = models.LdaModel(corpus = new_corpus, id2word = new_id2word, num_topics = 3 ,passes = 10) new_lda.print_topics() ###Output _____no_output_____ ###Markdown Topic Modeling - Attempt 3 (Nouns and Adjectives) ###Code # creating a function to pull out nouns from a string of text def nouns_adj(text): '''Given a string of text, tokenize the text and pull out only the nouns and adjectives.''' is_noun_adj = lambda pos: pos[:2] == 'NN' or pos[:2] == 'JJ' tokenized = word_tokenize(text) nouns_adj = [word for (word, pos) in pos_tag(tokenized) if is_noun_adj(pos)] return ' '.join(nouns_adj) # Applying the nouns function to the transcripts to filter only on nouns data_nouns_adj = pd.DataFrame(data_clean.transcript.apply(nouns_adj)) data_nouns_adj # Create a new document-term matrix using only nouns and adjectives, also remove common words with max_df cvna = CountVectorizer(stop_words=stop_words, max_df=.8) data_cvna = cvna.fit_transform(data_nouns_adj.transcript) data_dtmna = pd.DataFrame(data_cvna.toarray(), columns=cvna.get_feature_names()) data_dtmna.index = data_nouns_adj.index data_dtmna # Creating the gensim corpus corpusna = matutils.Sparse2Corpus(scipy.sparse.csr_matrix(data_dtmna.transpose())) # Creating the vocabulary dictionary id2wordna = dict((v, k) for k, v in cvna.vocabulary_.items()) # no of topics = 2 ldana = models.LdaModel(corpus=corpusna, num_topics=2, id2word=id2wordna, passes=10) ldana.print_topics() # no of topics = 3 ldana = models.LdaModel(corpus=corpusna, num_topics=3, id2word=id2wordna, passes=10) ldana.print_topics() # no of topics = 4 ldana = models.LdaModel(corpus=corpusna, num_topics=4, id2word=id2wordna, passes=10) ldana.print_topics() ###Output _____no_output_____ ###Markdown Identify Topics in Each DocumentOut of the various topic models we looked at, the nouns and adjectives, 4 topic one made the most sense. So pulling that down here and running it through some more iterations to get more fine-tuned topics. ###Code # final LDA model (for now) with more no of passes ldana = models.LdaModel(corpus=corpusna, num_topics=4, id2word=id2wordna, passes=80) ldana.print_topics() ###Output _____no_output_____ ###Markdown Trying to settle on these 4 topics as of now - Topic 0: guns - Topic 1: husband - Topic 2: mom, parents - Topic 3: grandma, friends ###Code # checking which topic does each transcript contains corpus_transformed = ldana[corpusna] list(zip([a for [(a,b)] in corpus_transformed], data_dtmna.index)) ###Output _____no_output_____
ex_10_ray_serve_example.ipynb
###Markdown Ray Serve - Model Serving© 2019-2022, Anyscale. All Rights Reserved Now we'll explore a short example for Ray Serve. This example is from the Ray Serve [scikit-learn example.](https://docs.ray.io/en/latest/serve/tutorials/sklearn.html)See also the Serve documentation's [mini-tutorials](https://docs.ray.io/en/latest/serve/tutorials/index.html) for using Serve with various frameworks. ###Code import ray from ray import serve import requests # for making web requests import tempfile import os import pickle import json import numpy as np serve.start() ###Output 2022-04-19 11:29:41,124 INFO services.py:1460 -- View the Ray dashboard at http://127.0.0.1:8268 (ServeController pid=3675) 2022-04-19 11:29:45,067 INFO checkpoint_path.py:15 -- Using RayInternalKVStore for controller checkpoint and recovery. (ServeController pid=3675) 2022-04-19 11:29:45,172 INFO http_state.py:106 -- Starting HTTP proxy with name 'SERVE_CONTROLLER_ACTOR:BkVkza:SERVE_PROXY_ACTOR-node:127.0.0.1-0' on node 'node:127.0.0.1-0' listening on '127.0.0.1:8000' 2022-04-19 11:29:46,505 INFO api.py:797 -- Started Serve instance in namespace 'serve'. ###Markdown Create a Model to Serve We'll begin by training a classifier with the Iris data we used before, this time using [scikit-learn](https://scikit-learn.org/stable/). The details aren't too important for our purposes, except for the fact we'll save the trained model to disk for subsequent serving. ###Code import sklearn from sklearn.datasets import load_iris from sklearn.ensemble import GradientBoostingClassifier from sklearn.metrics import mean_squared_error # Load data iris_dataset = load_iris() data, target, target_names = iris_dataset["data"], iris_dataset[ "target"], iris_dataset["target_names"] # Instantiate model model = GradientBoostingClassifier() # Training and validation split data, target = sklearn.utils.shuffle(data, target) train_x, train_y = data[:100], target[:100] val_x, val_y = data[100:], target[100:] # Train and evaluate models model.fit(train_x, train_y) print("MSE:", mean_squared_error(model.predict(val_x), val_y)) ###Output MSE: 0.02 ###Markdown Save the model and label to file. This could also be S3 or other "global" place or fetched from the model regsitry. ###Code MODEL_PATH = os.path.join(tempfile.gettempdir(), "iris_model_logistic_regression.pkl") LABEL_PATH = os.path.join(tempfile.gettempdir(), "iris_labels.json") with open(MODEL_PATH, "wb") as f: pickle.dump(model, f) with open(LABEL_PATH, "w") as f: json.dump(target_names.tolist(), f) ###Output _____no_output_____ ###Markdown Create a Deployment and Serve ItNext, we define a servable model by instantiating a class and defining the `__call__` method that Ray Serve will use. ###Code @serve.deployment(route_prefix="/regressor", num_replicas=2) class BoostingModel: def __init__(self): with open(MODEL_PATH, "rb") as f: self.model = pickle.load(f) with open(LABEL_PATH) as f: self.label_list = json.load(f) # async allows us to have this call concurrently async def __call__(self, starlette_request): payload = await starlette_request.json() print("Worker: received starlette request with data", payload) input_vector = [ payload["sepal length"], payload["sepal width"], payload["petal length"], payload["petal width"], ] prediction = self.model.predict([input_vector])[0] human_name = self.label_list[prediction] return {"result": human_name} ###Output _____no_output_____ ###Markdown Deploy the model ###Code BoostingModel.deploy() ###Output 2022-04-19 11:29:48,245 INFO api.py:618 -- Updating deployment 'BoostingModel'. component=serve deployment=BoostingModel (ServeController pid=3675) 2022-04-19 11:29:48,348 INFO deployment_state.py:1210 -- Adding 2 replicas to deployment 'BoostingModel'. component=serve deployment=BoostingModel 2022-04-19 11:29:50,252 INFO api.py:633 -- Deployment 'BoostingModel' is ready at `http://127.0.0.1:8000/regressor`. component=serve deployment=BoostingModel ###Markdown Score the modelInternally, Serve stores the model as a Ray actor and routes traffic to it as the endpoint is queried, in this case over HTTP. Now let’s query the endpoint to see results. ###Code sample_request_input = { "sepal length": 1.2, "sepal width": 1.0, "petal length": 1.1, "petal width": 0.9, } ###Output _____no_output_____ ###Markdown We can now send HTTP requests to our route `route_prefix=/regressor` at the default port 8000 ###Code response = requests.get( "http://localhost:8000/regressor", json=sample_request_input) print(response.text) response = requests.get("http://localhost:8000/regressor", json=sample_request_input).json() print(response) deployments = serve.list_deployments() print(f'deployments: {deployments}') serve.shutdown() ###Output (ServeController pid=3675) 2022-04-19 11:30:02,607 INFO deployment_state.py:1236 -- Removing 2 replicas from deployment 'BoostingModel'. component=serve deployment=BoostingModel
02-novice/050EarthquakesExercise.ipynb
###Markdown Extended exercise: the biggest earthquake in the UK this century USGS earthquake catalog[GeoJSON](https://en.wikipedia.org/wiki/GeoJSON) is a JSON-based file format for sharing geographic data. One example dataset available in the GeoJSON format is the [USGS earthquake catalog](https://www.usgs.gov/natural-hazards/earthquake-hazards/earthquakes). A [web service *application programming interface* (API)](https://earthquake.usgs.gov/fdsnws/event/1/) is provided for programatically accessing events in the earthquake catalog. Specifically the `query` method allows querying the catalog for events with the [query parameters](https://earthquake.usgs.gov/fdsnws/event/1/parameters) passed as `key=value` pairs.We can use the [`requests` Python library](https://docs.python-requests.org/en/latest/) to simplify constructing the appropriate query string to add to the URL and to deal with sending the HTTP request. ###Code import requests ###Output _____no_output_____ ###Markdown We first define a variable URL for the earthquake catalog web service API. ###Code earthquake_catalog_api_url = "http://earthquake.usgs.gov/fdsnws/event/1/query" ###Output _____no_output_____ ###Markdown We now need to define the parameters of our query. We want to get the data in GeoJSON format for all events in the earthquake catalog with date on or after 1st January 2000 and with location within [a bounding box covering the UK](http://bboxfinder.com/49.877000,-8.182000,60.830000,1.767000). We will filter the events to only request those with magnitude greater or equal to 1 to avoid downloading responses for more frequent small magnitude events. Finally we want the results to be returned in order of ascending time. ###Code query_parameters = { "format": "geojson", "starttime": "2000-01-01", "maxlatitude": "60.830", "minlatitude": "49.877", "maxlongitude": "1.767", "minlongitude": "-8.182", "minmagnitude": "1", "orderby": "time-asc" } ###Output _____no_output_____ ###Markdown We now can execute the API query using the [`requests.get` function](https://docs.python-requests.org/en/latest/api/requests.get). This takes as arguments the URL to make the request from and, optionally, a dictionary argument `params` containing the parameters to send in the query string for the request. A [`requests.Response` object](https://docs.python-requests.org/en/latest/api/requests.Response) is returned, which represents the server's response to the HTTP request made. ###Code quakes_response = requests.get(earthquake_catalog_api_url, params=query_parameters) ###Output _____no_output_____ ###Markdown The response object has various attributes and methods. A useful attribute to check is the `ok` attribute which will be `False` if the status code for the response to the request corresponds to a client or server error and `True` otherwise. ###Code quakes_response.ok ###Output _____no_output_____ ###Markdown We can also check specifically that the status code corresponds to [the expected `200 OK`](https://en.wikipedia.org/wiki/List_of_HTTP_status_codes2xx_success) using the `status_code` attribute ###Code quakes_response.status_code == 200 ###Output _____no_output_____ ###Markdown The actual content of the response can be accessed in various formats. The `content` attribute gives the content of the response as [bytes](https://docs.python.org/3/library/stdtypes.htmlbytes). As here we expect the response content to be Unicode-encoded text, the `text` attribute is more relevant as it gives the response content as a Python string. We can display the first 100 characters of the response content as follows ###Code print(quakes_response.text[:100]) ###Output _____no_output_____
notebooks/Using Jupyter/interactive-dashboards-with-jupyter.ipynb
###Markdown Interactive Dashboards with JupyterLet's say that you have to regularly send a [Folium](https://blog.dominodatalab.com/creating-interactive-crime-maps-with-folium/)map to your colleague's email with all the earthquakes of the past day.To be able to do that, you first need an earthquake data set that updates regularly (at least daily).A data feed that updates every 5 minutes can be found at the [USGS website](https://earthquake.usgs.gov/earthquakes/feed/v1.0/csv.php).Then, you can use Jupyter to write the code to load this data and create the map. FoliumFolium is a powerful Python library that helps you create several types of [Leaflet](http://leafletjs.com/) maps.The fact that the Folium results are interactive makes this library very useful for dashboard building.To get an idea, just zoom/click around on the next map to get an impression.The [Folium github](https://github.com/python-visualization/folium) contains many other examples.By default, Folium creates a map in a separate HTML file.In case you use Jupyter, you might prefer to get inline maps.This Jupyter example shows how to display maps inline. ###Code import folium from IPython.display import HTML def display(m, height=300): """Takes a folium instance and embed HTML.""" m._build_map() srcdoc = m.HTML.replace('"', '&quot;') embed = HTML('<iframe srcdoc="{0}" ' 'style="width: 100%; height: {1}px; ' 'border: none"></iframe>'.format(srcdoc, height)) return embed # print version number of your Folium package print("Folium Version: ", folium.__version__) map = folium.Map(location=[37.76, -122.45]) map.simple_marker([37.76, -122.45]) display(map) ###Output _____no_output_____ ###Markdown Dashboard ###Code import pandas as pd import folium from matplotlib.colors import Normalize, rgb2hex import matplotlib.cm as cm data = pd.read_csv('http://earthquake.usgs.gov/earthquakes/feed/v1.0/summary/all_day.csv') norm = Normalize(data['mag'].min(), data['mag'].max()) map = folium.Map(location=[48, -102], zoom_start=3) for eq in data.iterrows(): color = rgb2hex(cm.OrRd(norm(float(eq[1]['mag'])))) map.circle_marker([eq[1]['latitude'], eq[1]['longitude']], popup=eq[1]['place'], radius=20000*float(eq[1]['mag']), line_color=color, fill_color=color) map.create_map(path='results/earthquake.html') # need to replace CDN with https URLs with open('results/earthquake.html', 'r') as f: contents = f.read() contents = contents.replace("http://cdn.leafletjs.com/leaflet-0.5/", "//cdnjs.cloudflare.com/ajax/libs/leaflet/0.7.7/") with open('results/earthquake2.html', 'w') as f: f.writelines(contents) %%HTML <iframe width="100%" height="350" src="https://app.dominodatalab.com/r00sj3/jupyter/raw/latest/results/earthquake2.html?inline=true"></iframe> ###Output _____no_output_____
Kaggle-Competitions/Home Insurance/Exploratory Analysis.ipynb
###Markdown Exploratory Data Analysis ###Code # take a look at some of the examples train.head() train[train.QuoteConversion_Flag==0].head() train[train.QuoteConversion_Flag==1].head() # see 5 number summary train.describe() train.QuoteConversion_Flag.value_counts().plot(kind='bar'); def num_zero_features(row): return list(row).count(0) train['count_zero'] = train.apply(num_zero_features, axis=1) train = train.fillna(-1) train['count_missing'] = train.apply(lambda x: list(x).count(-1), axis=1) train.groupby(['QuoteConversion_Flag', 'count_zero']).size() train.boxplot(column='count_zero', by='QuoteConversion_Flag') train.groupby(['QuoteConversion_Flag', 'count_missing']).size() ###Output _____no_output_____
docs/gallery/plot_density.ipynb
###Markdown Density histogramm examples ###Code import pandas as pd import toto import matplotlib.pyplot as plt from toto.inputs.txt import TXTfile import os # read the file hindcast='https://raw.githubusercontent.com/calypso-science/Toto/master/_tests/txt_file/tahuna_hindcast.txt' measured='https://raw.githubusercontent.com/calypso-science/Toto/master/_tests/txt_file/tahuna_measured.txt' os.system('wget %s ' % hindcast) os.system('wget %s ' % measured) hd=TXTfile(['tahuna_hindcast.txt'],colNamesLine=1,skiprows=1,unitNamesLine=0,time_col_name={'Year':'year','Month':'month','Day':'day','Hour':'hour','Min':'Minute'}) hd.reads() hd.read_time() hd=hd._toDataFrame() # # Processing hd[0].StatPlots.density_diagramm(X='tp',Y='hs',args={ 'X name':'Wave period', 'Y name':'Significant wave height', 'Y unit':'m', 'X unit':'s', 'Y limits':[0,5], 'X limits':[0,20], 'display':'On', }) ###Output _____no_output_____
Data Visualization/exercise-bar-charts-and-heatmaps.ipynb
###Markdown **This notebook is an exercise in the [Data Visualization](https://www.kaggle.com/learn/data-visualization) course. You can reference the tutorial at [this link](https://www.kaggle.com/alexisbcook/bar-charts-and-heatmaps).**--- In this exercise, you will use your new knowledge to propose a solution to a real-world scenario. To succeed, you will need to import data into Python, answer questions using the data, and generate **bar charts** and **heatmaps** to understand patterns in the data. ScenarioYou've recently decided to create your very own video game! As an avid reader of [IGN Game Reviews](https://www.ign.com/reviews/games), you hear about all of the most recent game releases, along with the ranking they've received from experts, ranging from 0 (_Disaster_) to 10 (_Masterpiece_).![ex2_ign](https://i.imgur.com/Oh06Fu1.png)You're interested in using [IGN reviews](https://www.ign.com/reviews/games) to guide the design of your upcoming game. Thankfully, someone has summarized the rankings in a really useful CSV file that you can use to guide your analysis. SetupRun the next cell to import and configure the Python libraries that you need to complete the exercise. ###Code import pandas as pd pd.plotting.register_matplotlib_converters() import matplotlib.pyplot as plt %matplotlib inline import seaborn as sns print("Setup Complete") ###Output Setup Complete ###Markdown The questions below will give you feedback on your work. Run the following cell to set up our feedback system. ###Code # Set up code checking import os if not os.path.exists("../input/ign_scores.csv"): os.symlink("../input/data-for-datavis/ign_scores.csv", "../input/ign_scores.csv") from learntools.core import binder binder.bind(globals()) from learntools.data_viz_to_coder.ex3 import * print("Setup Complete") ###Output Setup Complete ###Markdown Step 1: Load the dataRead the IGN data file into `ign_data`. Use the `"Platform"` column to label the rows. ###Code # Path of the file to read ign_filepath = "../input/ign_scores.csv" # Fill in the line below to read the file into a variable ign_data ign_data = pd.read_csv(ign_filepath, index_col='Platform') # Run the line below with no changes to check that you've loaded the data correctly step_1.check() # Lines below will give you a hint or solution code #step_1.hint() # step_1.solution() ###Output _____no_output_____ ###Markdown Step 2: Review the dataUse a Python command to print the entire dataset. ###Code # Print the data ____ # Your code here ign_data ###Output _____no_output_____ ###Markdown The dataset that you've just printed shows the average score, by platform and genre. Use the data to answer the questions below. ###Code # Fill in the line below: What is the highest average score received by PC games, # for any platform? high_score = 7.759930 # Fill in the line below: On the Playstation Vita platform, which genre has the # lowest average score? Please provide the name of the column, and put your answer # in single quotes (e.g., 'Action', 'Adventure', 'Fighting', etc.) worst_genre = 'Simulation' # Check your answers step_2.check() # Lines below will give you a hint or solution code #step_2.hint() # step_2.solution() ###Output _____no_output_____ ###Markdown Step 3: Which platform is best?Since you can remember, your favorite video game has been [**Mario Kart Wii**](https://www.ign.com/games/mario-kart-wii), a racing game released for the Wii platform in 2008. And, IGN agrees with you that it is a great game -- their rating for this game is a whopping 8.9! Inspired by the success of this game, you're considering creating your very own racing game for the Wii platform. Part ACreate a bar chart that shows the average score for **racing** games, for each platform. Your chart should have one bar for each platform. ###Code # Bar chart showing average score for racing games by platform ____ # Your code here # set the width and height of the figure plt.figure(figsize=(25,8)) # Add Title to project plt.title("Average Score for Racing Games") # Bar Chart Showing the average score for racing games, for each platform sns.barplot(x=ign_data.index, y=ign_data['Racing']) # Add label to vertical axis plt.ylabel("Average Score") # Check your answer step_3.a.check() # Lines below will give you a hint or solution code #step_3.a.hint() # step_3.a.solution_plot() ###Output _____no_output_____ ###Markdown Part BBased on the bar chart, do you expect a racing game for the **Wii** platform to receive a high rating? If not, what gaming platform seems to be the best alternative? ###Code # step_3.b.hint() # Check your answer (Run this code cell to receive credit!) # step_3.b.solution() ###Output _____no_output_____ ###Markdown Step 4: All possible combinations!Eventually, you decide against creating a racing game for Wii, but you're still committed to creating your own video game! Since your gaming interests are pretty broad (_... you generally love most video games_), you decide to use the IGN data to inform your new choice of genre and platform. Part AUse the data to create a heatmap of average score by genre and platform. ###Code # Heatmap showing average game score by platform and genre ____ # Your code here # Set the Width & Height of the Figure plt.figure(figsize=(25,8)) # Add Title to Figure plt.title("Average Score by Genre and Platform") # Heat map showing average score by genre and platform sns.heatmap(data=ign_data, annot=True) # Add Label for h axis plt.xlabel("Genre") # Check your answer step_4.a.check() # Lines below will give you a hint or solution code #step_4.a.hint() # step_4.a.solution_plot() ###Output _____no_output_____ ###Markdown Part BWhich combination of genre and platform receives the highest average ratings? Which combination receives the lowest average rankings? ###Code #step_4.b.hint() # Check your answer (Run this code cell to receive credit!) # step_4.b.solution() ###Output _____no_output_____
notebooks/tutorial_change_params_srd.ipynb
###Markdown On crée un ménage avec un enfant ###Code jean = Person(age=45, earn=40000) jacques = Person(age=40, earn=50000) jeanne = Dependent(age=4, child_care=10000) joaquim = Dependent(age=8, child_care=8000) hh = Hhold(jean, jacques, prov='qc') hh.add_dependent(jeanne, joaquim) ###Output _____no_output_____ ###Markdown On crée une instance du simulateur pour l'année fiscale 2020 et on passe le ménage dans le simulateur ###Code tax_form = tax(2020) tax_form.compute(hh) print(f'revenu disponible familial: {hh.fam_disp_inc}') print(f'crédit pour frais de garde: {jean.qc_chcare + jacques.qc_chcare}') srd.quebec.template.chcare?? vars(tax_form.prov['qc']) tax_form.prov['qc'].chcare_young = 15000 tax_form.prov['qc'].chcare_old = 10000 ###Output _____no_output_____ ###Markdown On recrée le ménage et le passe dans le simulateur avec les nouveaux paramètres. ###Code jean = Person(age=45, earn=40000) jacques = Person(age=40, earn=50000) jeanne = Dependent(age=4, child_care=10000) joaquim = Dependent(age=8, child_care=8000) hh = Hhold(jean, jacques, prov='qc') hh.add_dependent(jeanne, joaquim) tax_form.prov['qc'].chcare_old = 8000 tax_form.compute(hh) print(f'revenu disponible familial: {hh.fam_disp_inc}') print(f'crédit pour frais de garde: {jean.qc_chcare + jacques.qc_chcare}') ###Output _____no_output_____
notebooks/01-KerasPoisonousMushrooms.ipynb
###Markdown Keras versus Poisonous MushroomsThis example demonstrates building a simple dense neural network using Keras. The example uses [Agaricus Lepiota](https://archive.ics.uci.edu/ml/datasets/Mushroom) training data to detect poisonous mushrooms. ###Code from pandas import read_csv srooms_df = read_csv('../data/agaricus-lepiota.data.csv') srooms_df.head() ###Output _____no_output_____ ###Markdown Feature extractionIf we wanted to use all the features in the training set then we would need to map each out. The ```LabelEncoder``` converts T/F data to 1 and 0. The ```LabelBinarizer``` converts categorical data to **one hot encoding**. If we wanted to use all the features in the training set then we would need to map each out:```column_names = srooms_df.axes[1]def get_mapping(name): if(name == 'edibility' or name == 'gill-attachment'): return (name, sklearn.preprocessing.LabelEncoder()) else: return (name, sklearn.preprocessing.LabelBinarizer()) mappings = list(map(lambda name: get_mapping(name), column_names)```We will use a subset of features to make it interesting. Are there simple rules or a handful of features that can be used to test edibility? Lets try a few. ###Code from sklearn_pandas import DataFrameMapper import sklearn import numpy as np mappings = ([ ('edibility', sklearn.preprocessing.LabelEncoder()), ('odor', sklearn.preprocessing.LabelBinarizer()), ('habitat', sklearn.preprocessing.LabelBinarizer()), ('spore-print-color', sklearn.preprocessing.LabelBinarizer()) ]) mapper = DataFrameMapper(mappings) srooms_np = mapper.fit_transform(srooms_df.copy()) ###Output _____no_output_____ ###Markdown Now lets transform the textual data to a vector... The transformed data should have 26 features. The break down is as follows:* Edibility (0 = edible, 1 = poisonous)* odor (9 features): ```[almond=a, creosote=c, foul=f, anise=l, musty=m, none=n, pungent=p, spicy=s, fishy=y]```* habitat (7 features): ```[woods=d, grasses=g, leaves=l, meadows=m, paths=p, urban=u, waste=w]```* spore-print-color (9 features): ```[buff=b, chocolate=h, black=k, brown=n, orange=o, green=r, purple=u, white=w, yellow=y]``` ###Code print(srooms_np.shape) print("Frist sample: {}".format(srooms_np[0])) print(" edibility (poisonous): {}".format(srooms_np[0][0])) print(" ordr (pungent): {}".format(srooms_np[0][1:10])) print(" habitat (urban): {}".format(srooms_np[0][10:17])) print(" spore-print-color (black): {}".format(srooms_np[0][17:])) ###Output (8124, 26) Frist sample: [1 0 0 0 0 0 0 1 0 0 0 0 0 0 0 1 0 0 0 1 0 0 0 0 0 0] edibility (poisonous): 1 ordr (pungent): [0 0 0 0 0 0 1 0 0] habitat (urban): [0 0 0 0 0 1 0] spore-print-color (black): [0 0 1 0 0 0 0 0 0] ###Markdown Before we train the neural network, let's split the data into training and test datasets. ###Code from sklearn.model_selection import train_test_split train, test = train_test_split(srooms_np, test_size = 0.2, random_state=7) train_labels = train[:,0:1] train_data = train[:,1:] test_labels = test[:,0:1] test_data = test[:,1:] print('training data dims: {}, label dims: {}'.format(train_data.shape,train_labels.shape)) print('test data dims: {}, label dims: {}'.format(test_data.shape,test_labels.shape)) ###Output training data dims: (6499, 25), label dims: (6499, 1) test data dims: (1625, 25), label dims: (1625, 1) ###Markdown Model DefinitionWe will create a simple three layer neural network. The network contains two dense layers and a dropout layer (to avoid overfitting). Layer 1: Dense LayerA dense layer applies an activation function to the output of $W \cdot x + b$. If the dense layer only had three inputs and outputs, then the dense layer looks like this... ![Dense Layer](images/DenseLayer.png) Under the covers, keras represents the layer's weights as a matrix. The inputs, outputs, and biases are vectors...$$ \begin{bmatrix} y_1 \\y_2 \\y_3\end{bmatrix}=relu\begin{pmatrix}\begin{bmatrix} W_{1,1} & W_{1,2} & W_{1,3} \\W_{2,1} & W_{2,2} & W_{2,3} \\W_{3,1} & W_{3,2} & W_{3,3}\end{bmatrix}\cdot\begin{bmatrix} x_1 \\x_2 \\x_3\end{bmatrix}+\begin{bmatrix} b_1 \\b_2 \\b_3\end{bmatrix}\end{pmatrix}$$ If this operation was decomposed futher, it would look like this...$$ \begin{bmatrix} y_1 \\y_2 \\y_3\end{bmatrix}=\begin{bmatrix}relu(W_{1,1} x_1 + W_{1,2} x_2 + W_{1,3} x_3 + b_1) \\relu(W_{2,1} x_1 + W_{2,2} x_2 + W_{2,3} x_3 + b_2) \\relu(W_{3,1} x_1 + W_{3,2} x_2 + W_{3,3} x_3 + b_3)\end{bmatrix}$$ The Rectified Linear Unit (RELU) function looks like this...![RELU](images/relu.png) Layer 2: DropoutThe dropout layer prevents overfitting by randomly dropping inputs to the next layer.![dropout layer](./images/dropout.png) Layer 3: Dense LayerThis layer acts like the first one, except this layer applies a sigmod activation function. The output is the probability a mushroom is poisonous. If a sample represents a small probability of poisoning, we'll want to know!![sigmod](./images/sigmod.png)$$y = sigmod(W \cdot x + b)$$ Putting It TogetherFortunately, we don't need to worry about defining the parameters (the weigths and biases) in Keras. We just define the layers in a sequence... ###Code from keras.models import Sequential from keras.layers import Dense, Dropout model = Sequential() model.add(Dense(20, activation='relu', input_dim=25)) model.add(Dropout(0.5)) model.add(Dense(1, activation='sigmoid')) model.summary() ###Output Using TensorFlow backend. ###Markdown Model Training Model ComplieThis step configures the model for training with the following settings:* An optimizier (update the gradients based on a loss function)* A loss function* Metrics to track during training ###Code model.compile(optimizer='adam', loss='binary_crossentropy', metrics=['accuracy']) ###Output _____no_output_____ ###Markdown Keras CallbacksKeras provides callbacks as a means to instrument internal state. In this example, we will write a tensorflow event log. The event log enables a tensorboard visualization of the translated model. The event log also captures key metrics during training. > Note: This step is completely optional and depends on the backend engine. ###Code from keras.callbacks import TensorBoard tensor_board = TensorBoard(log_dir='./logs/keras_srooms', histogram_freq=1) model.fit(train_data, train_labels, epochs=10, batch_size=32, callbacks=[tensor_board]) ###Output INFO:tensorflow:Summary name dense_1/kernel:0 is illegal; using dense_1/kernel_0 instead. INFO:tensorflow:Summary name dense_1/bias:0 is illegal; using dense_1/bias_0 instead. INFO:tensorflow:Summary name dense_2/kernel:0 is illegal; using dense_2/kernel_0 instead. INFO:tensorflow:Summary name dense_2/bias:0 is illegal; using dense_2/bias_0 instead. Epoch 1/10 6499/6499 [==============================] - 0s - loss: 0.4709 - acc: 0.8120 Epoch 2/10 6499/6499 [==============================] - 0s - loss: 0.2020 - acc: 0.9508 Epoch 3/10 6499/6499 [==============================] - 0s - loss: 0.1118 - acc: 0.9729 Epoch 4/10 6499/6499 [==============================] - 0s - loss: 0.0759 - acc: 0.9818 Epoch 5/10 6499/6499 [==============================] - 0s - loss: 0.0596 - acc: 0.9865 Epoch 6/10 6499/6499 [==============================] - 0s - loss: 0.0447 - acc: 0.9885 Epoch 7/10 6499/6499 [==============================] - 0s - loss: 0.0397 - acc: 0.9900 Epoch 8/10 6499/6499 [==============================] - 0s - loss: 0.0339 - acc: 0.9902 Epoch 9/10 6499/6499 [==============================] - 0s - loss: 0.0330 - acc: 0.9892 Epoch 10/10 6499/6499 [==============================] - 0s - loss: 0.0264 - acc: 0.9929 ###Markdown Model Evaluation ###Code score = model.evaluate(test_data, test_labels, batch_size=1625) print(score) ###Output 1625/1625 [==============================] - 0s [0.010582135990262032, 0.99507689476013184] ###Markdown Save/Restore the ModelKeras provides methods to save the models architecture as yaml or json. ###Code print(model.to_yaml()) definition = model.to_yaml() ###Output backend: tensorflow class_name: Sequential config: - class_name: Dense config: activation: relu activity_regularizer: null batch_input_shape: !!python/tuple [null, 25] bias_constraint: null bias_initializer: class_name: Zeros config: {} bias_regularizer: null dtype: float32 kernel_constraint: null kernel_initializer: class_name: VarianceScaling config: {distribution: uniform, mode: fan_avg, scale: 1.0, seed: null} kernel_regularizer: null name: dense_1 trainable: true units: 20 use_bias: true - class_name: Dropout config: {name: dropout_1, rate: 0.5, trainable: true} - class_name: Dense config: activation: sigmoid activity_regularizer: null bias_constraint: null bias_initializer: class_name: Zeros config: {} bias_regularizer: null kernel_constraint: null kernel_initializer: class_name: VarianceScaling config: {distribution: uniform, mode: fan_avg, scale: 1.0, seed: null} kernel_regularizer: null name: dense_2 trainable: true units: 1 use_bias: true keras_version: 2.0.4 ###Markdown We also need to save the *parameters* or weights learns from training. ###Code model.save_weights('/tmp/srmooms.hdf5') ###Output _____no_output_____ ###Markdown Model RestoreWe'll load the definition and parameters... ###Code from keras.models import model_from_yaml new_model = model_from_yaml(definition) new_model.load_weights('/tmp/srmooms.hdf5') ###Output _____no_output_____ ###Markdown Lets run some predictions on the newly initiated model. ###Code predictions = new_model.predict(test_data[0:25]).round() for i in range(25): if predictions[i]: print('Test sample {} is poisonous.'.format(i)) ###Output Test sample 0 is poisonous. Test sample 1 is poisonous. Test sample 8 is poisonous. Test sample 11 is poisonous. Test sample 14 is poisonous. Test sample 15 is poisonous. Test sample 17 is poisonous. ###Markdown Confusion Matrix ###Code predictions = new_model.predict(test_data).round() labels = test_labels[:,0] from sklearn.metrics import confusion_matrix cm = confusion_matrix(labels,predictions) import matplotlib.pyplot as plt import itertools def plot_confusion_matrix(cm, classes, normalize=False, title='Confusion matrix', cmap=plt.cm.Blues): """ This function prints and plots the confusion matrix. Normalization can be applied by setting `normalize=True`. """ plt.imshow(cm, interpolation='nearest', cmap=cmap) plt.title(title) plt.colorbar() tick_marks = np.arange(len(classes)) plt.xticks(tick_marks, classes, rotation=45) plt.yticks(tick_marks, classes) if normalize: cm = cm.astype('float') / cm.sum(axis=1)[:, np.newaxis] print("Normalized confusion matrix") else: print('Confusion matrix, without normalization') print(cm) thresh = cm.max() / 2. for i, j in itertools.product(range(cm.shape[0]), range(cm.shape[1])): plt.text(j, i, cm[i, j], horizontalalignment="center", color="white" if cm[i, j] > thresh else "black") plt.tight_layout() plt.ylabel('True label') plt.xlabel('Predicted label') plot_confusion_matrix(cm,['edible','poisonous']) plt.show() ###Output Confusion matrix, without normalization [[840 0] [ 8 777]]
Assignment+3 correct.ipynb
###Markdown ---_You are currently looking at **version 1.5** of this notebook. To download notebooks and datafiles, as well as get help on Jupyter notebooks in the Coursera platform, visit the [Jupyter Notebook FAQ](https://www.coursera.org/learn/python-data-analysis/resources/0dhYG) course resource._--- Assignment 3 - More PandasThis assignment requires more individual learning then the last one did - you are encouraged to check out the [pandas documentation](http://pandas.pydata.org/pandas-docs/stable/) to find functions or methods you might not have used yet, or ask questions on [Stack Overflow](http://stackoverflow.com/) and tag them as pandas and python related. And of course, the discussion forums are open for interaction with your peers and the course staff. Question 1 (20%)Load the energy data from the file `Energy Indicators.xls`, which is a list of indicators of [energy supply and renewable electricity production](Energy%20Indicators.xls) from the [United Nations](http://unstats.un.org/unsd/environment/excel_file_tables/2013/Energy%20Indicators.xls) for the year 2013, and should be put into a DataFrame with the variable name of **energy**.Keep in mind that this is an Excel file, and not a comma separated values file. Also, make sure to exclude the footer and header information from the datafile. The first two columns are unneccessary, so you should get rid of them, and you should change the column labels so that the columns are:`['Country', 'Energy Supply', 'Energy Supply per Capita', '% Renewable']`Convert `Energy Supply` to gigajoules (there are 1,000,000 gigajoules in a petajoule). For all countries which have missing data (e.g. data with "...") make sure this is reflected as `np.NaN` values.Rename the following list of countries (for use in later questions):```"Republic of Korea": "South Korea","United States of America": "United States","United Kingdom of Great Britain and Northern Ireland": "United Kingdom","China, Hong Kong Special Administrative Region": "Hong Kong"```There are also several countries with numbers and/or parenthesis in their name. Be sure to remove these, e.g. `'Bolivia (Plurinational State of)'` should be `'Bolivia'`, `'Switzerland17'` should be `'Switzerland'`.Next, load the GDP data from the file `world_bank.csv`, which is a csv containing countries' GDP from 1960 to 2015 from [World Bank](http://data.worldbank.org/indicator/NY.GDP.MKTP.CD). Call this DataFrame **GDP**. Make sure to skip the header, and rename the following list of countries:```"Korea, Rep.": "South Korea", "Iran, Islamic Rep.": "Iran","Hong Kong SAR, China": "Hong Kong"```Finally, load the [Sciamgo Journal and Country Rank data for Energy Engineering and Power Technology](http://www.scimagojr.com/countryrank.php?category=2102) from the file `scimagojr-3.xlsx`, which ranks countries based on their journal contributions in the aforementioned area. Call this DataFrame **ScimEn**.Join the three datasets: GDP, Energy, and ScimEn into a new dataset (using the intersection of country names). Use only the last 10 years (2006-2015) of GDP data and only the top 15 countries by Scimagojr 'Rank' (Rank 1 through 15). The index of this DataFrame should be the name of the country, and the columns should be ['Rank', 'Documents', 'Citable documents', 'Citations', 'Self-citations', 'Citations per document', 'H index', 'Energy Supply', 'Energy Supply per Capita', '% Renewable', '2006', '2007', '2008', '2009', '2010', '2011', '2012', '2013', '2014', '2015'].*This function should return a DataFrame with 20 columns and 15 entries.* ###Code import pandas as pd import numpy as np def answer_one(): x = pd.ExcelFile('Energy Indicators.xls') energy = x.parse(skiprows=17,skip_footer=(38)) energy = energy[['Unnamed: 1','Petajoules','Gigajoules','%']] energy.columns = ['Country', 'Energy Supply', 'Energy Supply per Capita', '% Renewable'] energy[['Energy Supply', 'Energy Supply per Capita', '% Renewable']] = energy[['Energy Supply', 'Energy Supply per Capita', '% Renewable']].replace('...',np.NaN).apply(pd.to_numeric) energy['Energy Supply'] = energy['Energy Supply']*1000000 energy['Country'] = energy['Country'].replace({'China, Hong Kong Special Administrative Region':'Hong Kong','United Kingdom of Great Britain and Northern Ireland':'United Kingdom','Republic of Korea':'South Korea','United States of America':'United States','Iran (Islamic Republic of)':'Iran'}) energy['Country'] = energy['Country'].str.replace(r" \(.*\)","") GDP = pd.read_csv('world_bank.csv',skiprows=4) GDP['Country Name'] = GDP['Country Name'].replace('Korea, Rep.','South Korea') GDP['Country Name'] = GDP['Country Name'].replace('Iran, Islamic Rep.','Iran') GDP['Country Name'] = GDP['Country Name'].replace('Hong Kong SAR, China','Hong Kong') GDP = GDP[['Country Name','2006','2007','2008','2009','2010','2011','2012','2013','2014','2015']] GDP.columns = ['Country','2006','2007','2008','2009','2010','2011','2012','2013','2014','2015'] ScimEn = pd.read_excel(io='scimagojr-3.xlsx') ScimEn_m = ScimEn[:15] df = pd.merge(ScimEn_m,energy,how='inner',left_on='Country',right_on='Country') final_df = pd.merge(df,GDP,how='inner',left_on='Country',right_on='Country') final_df = final_df.set_index('Country') return final_df final_df= answer_one() answer_one() ###Output _____no_output_____ ###Markdown Question 2 (6.6%)The previous question joined three datasets then reduced this to just the top 15 entries. When you joined the datasets, but before you reduced this to the top 15 items, how many entries did you lose?*This function should return a single number.* ###Code %%HTML <svg width="800" height="300"> <circle cx="150" cy="180" r="80" fill-opacity="0.2" stroke="black" stroke-width="2" fill="blue" /> <circle cx="200" cy="100" r="80" fill-opacity="0.2" stroke="black" stroke-width="2" fill="red" /> <circle cx="100" cy="100" r="80" fill-opacity="0.2" stroke="black" stroke-width="2" fill="green" /> <line x1="150" y1="125" x2="300" y2="150" stroke="black" stroke-width="2" fill="black" stroke-dasharray="5,3"/> <text x="300" y="165" font-family="Verdana" font-size="35">Everything but this!</text> </svg> import pandas as pd import numpy as np def answer_two(): energy = pd.read_excel("Energy Indicators.xls",skip_footer=38,skip_header=1,skiprows=17) # Skip header and footer energy.drop(energy.columns[[0,1]],axis=1,inplace=True) # Drop first 2 columns energy.columns = ['Country', 'Energy Supply', 'Energy Supply per Capita', '% Renewable'] energy.dropna() # Drop rows with NaN values. energy['Country'] = energy['Country'].str.replace(r'\(.*\)', '') # Remove contents within parenthesis. energy['Country'] = energy['Country'].str.replace('\d+', '') # Remove digits from names energy['Country'] = energy['Country'].str.strip() # This brings the Iran energy values back! # Turn blank values into NaN for col in energy: energy[col] = energy[col].replace('...',np.nan) energy['Country'] = energy['Country'].str.replace('Republic of Korea','South Korea') energy['Country'] = energy['Country'].str.replace('United States of America','United States') energy['Country'] = energy['Country'].str.replace('United Kingdom of Great Britain and Northern Ireland','United Kingdom') energy['Country'] = energy['Country'].str.replace('China, Hong Kong Special Administrative Region','Hong Kong') # GDP: GDP = pd.read_csv('world_bank.csv', skiprows=3) # Skip header # Make first row the column names new_header = GDP.iloc[0] GDP = GDP[1:] GDP.columns = new_header #GDP = GDP.rename(index=str,columns = {"Country Name":"Country"}) GDP['Country Name'] = GDP['Country Name'].str.replace('Korea, Rep.','South Korea') GDP['Country Name'] = GDP['Country Name'].str.replace('Iran, Islamic Rep.','Iran') GDP['Country Name'] = GDP['Country Name'].str.replace('Hong Kong SAR, China','Hong Kong') # Change column name from 'Country Name' to 'Country' for merging 3 files on country name. names = GDP.columns.tolist() names[names.index('Country Name')] = 'Country' GDP.columns = names # Only keep the columns from 2006-15. Drop column number 1 to 50. Don't need country code, etc. GDP = GDP.drop(GDP.iloc[:,1:50], axis=1) GDP.columns = GDP.columns.astype(str).str.split('.').str[0] # Remove '.0' at the end of the year columns. # SCIMEN: ScimEn = pd.read_excel('scimagojr-3.xlsx') # LOST ENTRIES = LEN(OUTER JOIN) - LEN(INNER JOIN) # Need unique entries in all 3 sets so use concat. Can't do that with a left or right outer join! num_outer = len(pd.concat([ScimEn['Country'],energy['Country'],GDP['Country']]).unique()) num_inter = (GDP.merge(energy, left_on='Country', right_on='Country', how='inner').merge(ScimEn, left_on='Country', right_on='Country', how='inner').shape[0]) return num_outer-num_inter answer_two() ###Output _____no_output_____ ###Markdown Answer the following questions in the context of only the top 15 countries by Scimagojr Rank (aka the DataFrame returned by `answer_one()`) Question 3 (6.6%)What is the average GDP over the last 10 years for each country? (exclude missing values from this calculation.)*This function should return a Series named `avgGDP` with 15 countries and their average GDP sorted in descending order.* ###Code import pandas as pd import numpy as np def answer_three(): Top15 = final_df columns = ['2006','2007','2008','2009','2010','2011','2012','2013','2014','2015'] Top15['Mean'] = Top15[columns].mean(axis=1) avgGDP = Top15.sort_values(by = 'Mean', ascending = False)['Mean'] return avgGDP answer_three() ###Output _____no_output_____ ###Markdown Question 4 (6.6%)By how much had the GDP changed over the 10 year span for the country with the 6th largest average GDP?*This function should return a single number.* ###Code import pandas as pd import numpy as np def answer_four(): Top15 = final_df columns = ['2006','2007','2008','2009','2010','2011','2012','2013','2014','2015'] Top15['Mean'] = Top15[columns].mean(axis=1) avgGDP = Top15.sort_values(by = 'Mean', ascending = False)['Mean'] target = avgGDP.index[5] target_data = Top15.loc[target] ans = target_data['2015'] - target_data['2006'] return ans answer_four() ###Output _____no_output_____ ###Markdown Question 5 (6.6%)What is the mean `Energy Supply per Capita`?*This function should return a single number.* ###Code import pandas as pd import numpy as np def answer_five() : Top15 = final_df return Top15['Energy Supply per Capita'].mean() answer_five() ###Output _____no_output_____ ###Markdown Question 6 (6.6%)What country has the maximum % Renewable and what is the percentage?*This function should return a tuple with the name of the country and the percentage.* ###Code import pandas as pd import numpy as np def answer_six() : Top15 = final_df ct = Top15.sort_values(by='% Renewable', ascending=False).iloc[0] return (ct.name, ct['% Renewable']) answer_six() ###Output _____no_output_____ ###Markdown Question 7 (6.6%)Create a new column that is the ratio of Self-Citations to Total Citations. What is the maximum value for this new column, and what country has the highest ratio?*This function should return a tuple with the name of the country and the ratio.* ###Code import pandas as pd import numpy as np def answer_seven(): Top15 = final_df Top15['Citation_ratio'] = Top15['Self-citations']/Top15['Citations'] ct = Top15.sort_values(by='Citation_ratio', ascending=False).iloc[0] return (ct.name, ct['Citation_ratio']) answer_seven() ###Output _____no_output_____ ###Markdown Question 8 (6.6%)Create a column that estimates the population using Energy Supply and Energy Supply per capita. What is the third most populous country according to this estimate?*This function should return a single string value.* ###Code import pandas as pd import numpy as np def answer_eight(): Top15 = final_df Top15['Population'] = Top15['Energy Supply']/Top15['Energy Supply per Capita'] return Top15.sort_values(by='Population', ascending=False).iloc[2].name answer_eight() ###Output _____no_output_____ ###Markdown Question 9 (6.6%)Create a column that estimates the number of citable documents per person. What is the correlation between the number of citable documents per capita and the energy supply per capita? Use the `.corr()` method, (Pearson's correlation).*This function should return a single number.**(Optional: Use the built-in function `plot9()` to visualize the relationship between Energy Supply per Capita vs. Citable docs per Capita)* ###Code import pandas as pd import numpy as np def answer_nine(): Top15 = final_df Top15['Estimate Population'] = Top15['Energy Supply'] / Top15['Energy Supply per Capita'] Top15['avgCiteDocPerPerson'] = Top15['Citable documents'] / Top15['Estimate Population'] return Top15[['Energy Supply per Capita', 'avgCiteDocPerPerson']].corr().ix['Energy Supply per Capita', 'avgCiteDocPerPerson'] answer_nine() ###Output _____no_output_____ ###Markdown Question 10 (6.6%)Create a new column with a 1 if the country's % Renewable value is at or above the median for all countries in the top 15, and a 0 if the country's % Renewable value is below the median.*This function should return a series named `HighRenew` whose index is the country name sorted in ascending order of rank.* ###Code import pandas as pd import numpy as np def answer_ten(): Top15 = final_df mid = Top15['% Renewable'].median() Top15['HighRenew'] = Top15['% Renewable']>=mid Top15['HighRenew'] = Top15['HighRenew'].apply(lambda x:1 if x else 0) Top15.sort_values(by='Rank', inplace=True) return Top15['HighRenew'] answer_ten() ###Output _____no_output_____ ###Markdown Question 11 (6.6%)Use the following dictionary to group the Countries by Continent, then create a dateframe that displays the sample size (the number of countries in each continent bin), and the sum, mean, and std deviation for the estimated population of each country.```pythonContinentDict = {'China':'Asia', 'United States':'North America', 'Japan':'Asia', 'United Kingdom':'Europe', 'Russian Federation':'Europe', 'Canada':'North America', 'Germany':'Europe', 'India':'Asia', 'France':'Europe', 'South Korea':'Asia', 'Italy':'Europe', 'Spain':'Europe', 'Iran':'Asia', 'Australia':'Australia', 'Brazil':'South America'}```*This function should return a DataFrame with index named Continent `['Asia', 'Australia', 'Europe', 'North America', 'South America']` and columns `['size', 'sum', 'mean', 'std']`* ###Code import pandas as pd import numpy as np def answer_eleven(): Top15 = final_df ContinentDict = {'China':'Asia', 'United States':'North America', 'Japan':'Asia', 'United Kingdom':'Europe', 'Russian Federation':'Europe', 'Canada':'North America', 'Germany':'Europe', 'India':'Asia', 'France':'Europe', 'South Korea':'Asia', 'Italy':'Europe', 'Spain':'Europe', 'Iran':'Asia', 'Australia':'Australia', 'Brazil':'South America'} new_df = pd.DataFrame() for i, row in Top15.iterrows(): row['Continent'] = ContinentDict[row.name] new_df = new_df.append(row) return new_df answer_eleven() ###Output _____no_output_____ ###Markdown Question 12 (6.6%)Cut % Renewable into 5 bins. Group Top15 by the Continent, as well as these new % Renewable bins. How many countries are in each of these groups?*This function should return a __Series__ with a MultiIndex of `Continent`, then the bins for `% Renewable`. Do not include groups with no countries.* ###Code import pandas as pd import numpy as np Top15 = final_df ContinentDict = {'China':'Asia', 'United States':'North America', 'Japan':'Asia', 'United Kingdom':'Europe', 'Russian Federation':'Europe', 'Canada':'North America', 'Germany':'Europe', 'India':'Asia', 'France':'Europe', 'South Korea':'Asia', 'Italy':'Europe', 'Spain':'Europe', 'Iran':'Asia', 'Australia':'Australia', 'Brazil':'South America'} Top15 = Top15.reset_index() Top15['Continent'] = [ContinentDict[country] for country in Top15['Country']] Top15['bins'] = pd.cut(Top15['% Renewable'],5) Top15.groupby(['Continent','bins']).size() ###Output _____no_output_____ ###Markdown Question 13 (6.6%)Convert the Population Estimate series to a string with thousands separator (using commas). Do not round the results.e.g. 317615384.61538464 -> 317,615,384.61538464*This function should return a Series `PopEst` whose index is the country name and whose values are the population estimate string.* ###Code import pandas as pd import numpy as np def answer_thirteen(): Top15 = final_df Top15['PopEst'] = (Top15['Energy Supply'] / Top15['Energy Supply per Capita']).astype(float) return Top15['PopEst'].apply(lambda x: '{0:,}'.format(x)) answer_thirteen() ###Output _____no_output_____
notebooks/4_Segmentation.ipynb
###Markdown Quadrant Scan The Quadrant Scan is a method based on recurrence plot that identify tipping points (change points) at which the system changes its behaviour (dynamic). The method is applicable on univariate and multivariate data.The code below is for the Weighted Quadrant Scan (WQS). The inputs of the function are: x : the data, a data frame with columns represent different variables for multivariate application, in this case the normalisation step is required. Or for univariate application, x is the embedded time series, in this case the normalisation step is not required. alpha : is the recurrence threshold, it is a problem dependant parameter, good value to start with is 0.1 m1 and m2 : are the weighting parameters, also they are problem dependant, values to use (m1,m2)=(200,50), (100,25) or (50,10)For referencing:Zaitouny, A., Walker, D.M. and Small, M., 2019. Quadrant scan for multi-scale transition detection. Chaos: An Interdisciplinary Journal of Nonlinear Science, 29(10), p.103117.Zaitouny, A., Small, M., Hill, J., Emelyanova, I. and Clennell, M.B., 2020. Fast automatic detection of geological boundaries from multivariate log data using recurrence. Computers & Geosciences, 135, p.104362. Import libraries and define functions ###Code import scipy.io as sio import time import scipy import numpy as np import pandas as pd import matplotlib.pyplot as plt from scipy.stats import pearsonr import scipy.sparse as sparse from IPython.display import Latex from matplotlib.pyplot import imshow ##Recurrence Plot matrix ## This function is constructing the recurrence plot matrix based on the threshold alpha. alpha is between 0 and 1. def RecurrenceMatrix(x, alpha): distance = scipy.spatial.distance.pdist(x, 'euclidean') #Different norms for different applications #distance = scipy.spatial.distance.pdist(x, 'seuclidean',V=None) #distance = scipy.spatial.distance.pdist(x, 'canberra') #distance = scipy.spatial.distance.pdist(x, 'mahalanobis', VI=None) distance = scipy.spatial.distance.squareform(distance) threshold = alpha*(distance.mean()+3*distance.std()) #Recurrence Plot Matrix RM=np.array(distance<=threshold,dtype='int') return RM ## Normalisation ## This function is for normalising the columns in a multivariate input which is important to avoid dominance ## of variable with larger scale ## This step is not required for univariate application and embedded time series def normalize(df): result = df.copy() for feature_name in df.columns: max_value = df[feature_name].max() min_value = df[feature_name].min() result[feature_name] = (df[feature_name] - min_value) / (max_value - min_value) return result ##Quadrant Scan ## This function is employing the Quadrant Scan technique on input x (univariate or multivariate) using: ## the recuurence plot threshold alpha, the weighting parameters m1 and m2. def WeightQS2(x, alpha, m1, m2): x = normalize(x) # not required for embedded time series x = np.array(x) z = np.array(x[:20]) np.random.shuffle(z) x = np.append(z, x, axis=0) z = np.array(x[len(x)-20:len(x)]) np.random.shuffle(z) x = np.append(x, z, axis=0) x = np.array(x) RM=RecurrenceMatrix(x, alpha) qs=np.zeros(len(x)) for ii in range(1, len(x)): weightp = 0.5*(1-np.tanh((np.arange(1,ii+1)-m1)/m2)) weightp = weightp[::-1] weightf = 0.5*(1-np.tanh((np.arange(1,len(x)-ii+1)-m1)/m2)) weightpp = weightp[:,None]*weightp[None,:] weightpp = weightpp/weightpp[-1,-1] weightff = weightf[:,None]*weightf[None,:] weightff = weightff/weightff[0,0] weightpf=weightp[:,None]*weightf[None,:] weightpf /= weightpf[-1,0] weightfp=weightf[:,None]*weightp[None,:] weightfp /= weightfp[0,-1] pp = RM[:ii,:ii] * weightpp ff = RM[ii-1:-1,ii-1:-1] * weightff pf = RM[:ii,ii-1:-1] * weightpf fp = RM[ii-1:-1,:ii] * weightfp qs1 = np.sum(pp)+np.sum(ff) qs2 = np.sum(pf)+np.sum(fp) qs[ii] = qs1/(qs1+qs2) return qs[20:len(qs)-20] ###Output _____no_output_____ ###Markdown Simple artificial example for demonstration ###Code # Creating a 2D dataframe at which each feature shows a transition in different time x1 = np.random.rand(100)+3 x2 = np.random.rand(100)+5 x = np.concatenate((x1, x2)) # first column y1 = np.random.rand(150)+8 y2 = np.random.rand(50)+2 y = np.concatenate((y1, y2)) # second column df = pd.DataFrame({'Feature1': x, 'Feature2' : y }) # the dataframe df.head() ## implementing the weighted quadrant scan wqs = WeightQS2(df, 0.1, 100, 25) ## Plot the data and the wqs for demonstration fig, ax = plt.subplots(2,1,figsize=(15,10)) ax[0].plot(df['Feature1'] , label = 'Feature1') ax[0].plot(df['Feature2'], label = 'Feature2') ax[0].set_xlabel('Time index') ax[0].set_ylabel('Value') ax[0].set_title('The 2D data') ax[0].legend() ax[1].plot(wqs) ax[1].set_xlabel('Time index') ax[1].set_ylabel('WQS') ax[1].set_title('The weighted quadrant scan results') plt.show ###Output _____no_output_____ ###Markdown Detect lithological boundaries using Quadrant Scan In this notebook, we will test the Quadrant Scan technique to analyse a multivariate, noisy, and nonstationary data set.The data set includes petrophysical profiles from well_log measures. Namely, DTCO, ECGR, HART, PEFZ, RHOZ, TNPH.The data includes 7688 depth samples. Refering to the following petrophysical profiles density, electrical resistivity, sonic velocity, natural radioactivity, mean atomic number and neutron porosityWe have added a column to the data set (last column: "Geological layer") in which we gave a number to each layer type as identified by the geologists.The idea of this exercise is to use the Quadrant Scan to detect transision in the data profiles to identify the lithological boundaries. Reading the data file ###Code Petrophysical_data = pd.read_csv('../data/Petrophysical_data.csv') ## Reading the data set file, fix the directory Petrophysical_data.head() ## Have a look at the data xx = pd.DataFrame(Petrophysical_data[['DTCO','ECGR']]) ## Select two columns from the data file xx.head() ###Output _____no_output_____ ###Markdown Implement the Quadrant ScanThere are three paramteres to set up. alpha --> the recurrence plot threshold, this can be varied for multiscale detection, Good choice for the data at hand is 0.2m1, m2 are the weighting scheme parameters as demonstrated in the slides, an appropraite setting for the the data at hand is m1=200, m2=50. You could try different settings such as m1=100, m2=25 or m1=50, m2=10 Input a single variableChoose one variable and run the Quadrant Scan for 2000 depth samples (between depth index 1000 to depth index 3000)Here, We chose ECGR - Gamma Ray ###Code x1 = pd.DataFrame(Petrophysical_data[['ECGR']]) ## select a variable from the data set QS1 = WeightQS2(x1[1000:3000], 0.2, 200, 50) ## Run the Quadrant Scan on the selected column (variable) ## Plotting the resulted Quadrant Scan with the input data. RM = RecurrenceMatrix(x1[1000:3000], 0.2) ## Estimating the recurrence plot matrix for demonstration. ## You can change the parameter alpha to see how it effects the recurrence plot matrix. (alpha is between 0 to 1) f1 = plt.figure() ## plot the recurrence plot matrix imshow(np.asarray(RM)) plt.title('Recurrence Plot Matrix') plt.xlabel('Depth index') plt.ylabel('Depth index') f2 = plt.figure(figsize=(6,8)) ## plot the Quadrant Scan and the input data plt.subplot(1, 2, 1) plt.plot(QS1, np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.title('Boundary detection') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 2, 2) plt.plot(np.array(Petrophysical_data['ECGR'][1000:3000]), np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.xlabel('ECGR') plt.ylabel('Depth') f2.tight_layout() ###Output _____no_output_____ ###Markdown Now try for another variable Choose another variable and run the Quadrant Scan for 2000 depth samples (between depth index 1000 to depth index 3000)Here, We chose HART - Resistivity ###Code x2 = pd.DataFrame(Petrophysical_data[['HART']]) ## select a variable from the data set QS2 = WeightQS2(x2[1000:3000], 0.2, 200, 50) ## Run the Quadrant Scan on the selected column (variable) ## Plotting the resulted Quadrant Scan with the input data. f1 = plt.figure(figsize=(6,8)) plt.subplot(1, 2, 1) plt.plot(QS2, np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.title('Boundary detection') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 2, 2) plt.plot(np.array(Petrophysical_data['HART'][1000:3000]), np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.xlabel('HART') plt.ylabel('Depth') f1.tight_layout() ###Output _____no_output_____ ###Markdown Compare the detections from different profiles ###Code ## Plotting the resulted Quadrant Scan with the input data. f1 = plt.figure(figsize = (6,8)) plt.plot(QS1, np.array(Petrophysical_data['DEPTH'][1000:3000]), label="QS ECGR") plt.plot(QS2, np.array(Petrophysical_data['DEPTH'][1000:3000]), label="QS HART") plt.gca().invert_yaxis() plt.xlabel('QS') plt.ylabel('Depth') plt.legend() plt.show() ###Output _____no_output_____ ###Markdown Now combine the variables and use multivariate inputLet us start combining the profiles we used above. ###Code x3 = pd.DataFrame(Petrophysical_data[['HART','ECGR']]) ## Select multiple variables QS3 = WeightQS2(x3[1000:3000], 0.2, 200, 50) ## Run the Quadrant Scan on the selected columns (variables) ## Plotting the resulted Quadrant Scan with the input data. f2 = plt.figure(figsize = (10,6)) plt.subplot(1, 3, 1) plt.plot(QS3, np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.title('Boundary detection') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 3, 2) plt.plot(np.array(Petrophysical_data['HART'][1000:3000]), np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.xlabel('HART') plt.ylabel('Depth') plt.subplot(1, 3, 3) plt.plot(np.array(Petrophysical_data['ECGR'][1000:3000]), np.array(Petrophysical_data['DEPTH'][1000:3000])) plt.gca().invert_yaxis() plt.xlabel('ECGR') plt.ylabel('Depth') f2.tight_layout() ###Output _____no_output_____ ###Markdown Compare the results ###Code ## Plotting the Quadrant curves for comparison. f1 = plt.figure(figsize = (6,8)) plt.plot(QS1, np.array(Petrophysical_data['DEPTH'][1000:3000]), label="QS ECGR") plt.plot(QS2, np.array(Petrophysical_data['DEPTH'][1000:3000]), label="QS HART") plt.plot(QS3, np.array(Petrophysical_data['DEPTH'][1000:3000]), label="QS MV") plt.gca().invert_yaxis() plt.xlabel('QS') plt.ylabel('Depth') plt.legend() plt.show() ###Output _____no_output_____ ###Markdown Multiscale detection:Test the performance by varying the Recurrence Plot threshold (alpha) from larger value to smaller value:Say from alpha=0.6 to 0.1You can test using univariate or multivariate input ###Code ## Evaluate the Quadrant Scan for different thresholds. MS_QS6 = WeightQS2(x1[2000:3000], 0.6, 200, 50) MS_QS5 = WeightQS2(x1[2000:3000], 0.5, 200, 50) MS_QS4 = WeightQS2(x1[2000:3000], 0.4, 200, 50) MS_QS3 = WeightQS2(x1[2000:3000], 0.3, 200, 50) MS_QS2 = WeightQS2(x1[2000:3000], 0.2, 200, 50) MS_QS1 = WeightQS2(x1[2000:3000], 0.1, 200, 50) ###Output _____no_output_____ ###Markdown Plotting for comparison ###Code ## Plotting the resulted Quadrant Scan curves for each threshold. f3 = plt.figure(figsize=(15,6)) plt.subplot(1, 6, 1) plt.plot(MS_QS6, np.array(Petrophysical_data['DEPTH'][2000:3000])) plt.gca().invert_yaxis() plt.title('alpha=0.6') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 6, 2) plt.plot(MS_QS5, np.array(Petrophysical_data['DEPTH'][2000:3000])) plt.gca().invert_yaxis() plt.title('alpha=0.5') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 6, 3) plt.plot(MS_QS4, np.array(Petrophysical_data['DEPTH'][2000:3000])) plt.gca().invert_yaxis() plt.title('alpha=0.4') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 6, 4) plt.plot(MS_QS3, np.array(Petrophysical_data['DEPTH'][2000:3000])) plt.gca().invert_yaxis() plt.title('alpha=0.3') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 6, 5) plt.plot(MS_QS2, np.array(Petrophysical_data['DEPTH'][2000:3000])) plt.gca().invert_yaxis() plt.title('alpha=0.2') plt.xlabel('QS') plt.ylabel('Depth') plt.subplot(1, 6, 6) plt.plot(MS_QS1, np.array(Petrophysical_data['DEPTH'][2000:3000])) plt.gca().invert_yaxis() plt.title('alpha=0.1') plt.xlabel('QS') plt.ylabel('Depth') f3.tight_layout() ###Output _____no_output_____
examples/seaborn_01_distribution.ipynb
###Markdown distplotCalling ``sns.distplot`` shows a distribution of the ``target`` colum by default. ###Code df.sns.distplot(); ###Output _____no_output_____ ###Markdown You can specify options as the same as the standard ``seaborn``. ###Code df.sns.distplot(kde=False, rug=True); df.sns.distplot(bins=20, kde=False, rug=True); df.sns.distplot(hist=False, rug=True); ###Output _____no_output_____ ###Markdown Specifying the column name shows distribution of the specified column. ###Code df.sns.distplot('b'); ###Output _____no_output_____ ###Markdown kdeplot``kdeplot`` works almost the same as ``distplot`` ###Code df.sns.kdeplot(shade=True); df.sns.kdeplot() df.sns.kdeplot(bw=.2, label="bw: 0.2") df.sns.kdeplot(bw=2, label="bw: 2") plt.legend(); df.sns.kdeplot('b', shade=True, cut=0) df.sns.rugplot('b'); ###Output _____no_output_____ ###Markdown jointplotYou must specify the label of x-axis via ``x`` keyword. If ``y`` keyword is omitted, it will be a ``target`` column. ###Code df.sns.jointplot('b', size=4); ###Output /Users/sin/miniconda/envs/py27std/lib/python2.7/site-packages/matplotlib/__init__.py:892: UserWarning: axes.color_cycle is deprecated and replaced with axes.prop_cycle; please use the latter. warnings.warn(self.msg_depr % (key, alt_key)) ###Markdown If you specify both ``x`` and ``y``, it will shows the specified columns. ###Code df.sns.jointplot('b', 'c', kind="hex", color="k", size=4); df.sns.jointplot("d", "e", kind="kde", size=4); f, ax = plt.subplots() df.sns.kdeplot('b', 'c', ax=ax) df.sns.rugplot('b', color="g", ax=ax) df.sns.rugplot('c', vertical=True, ax=ax); cmap = df.sns.cubehelix_palette(as_cmap=True, dark=0, light=1, reverse=True) df.sns.kdeplot('b', 'c', cmap=cmap, n_levels=60, shade=True); g = df.sns.jointplot(x="a", y="b", kind="kde", color="m", size=4) g.plot_joint(plt.scatter, c="w", s=30, linewidth=1, marker="+") g.ax_joint.collections[0].set_alpha(0); ###Output _____no_output_____ ###Markdown pairplot ###Code import seaborn as sns iris = sns.load_dataset("iris") iris = pdml.ModelFrame(iris, target='species') iris.head() iris.sns.pairplot(size=1.5); import seaborn as sns g = iris.sns.PairGrid(size=1.5) g.map_diag(iris.sns.kdeplot) g.map_offdiag(iris.sns.kdeplot, cmap="Blues_d", n_levels=6); ###Output /Users/sin/miniconda/envs/py27std/lib/python2.7/site-packages/matplotlib/axes/_axes.py:519: UserWarning: No labelled objects found. Use label='...' kwarg on individual plots. warnings.warn("No labelled objects found. "
tutorials/tutorial06/tutorial06.ipynb
###Markdown [Tutorial 06: Classification](https://franciszheng.com/dsper2020/tutorials/tutorial06/) [[1] Getting Started](https://franciszheng.com/dsper2020/tutorials/tutorial06/getting-started) [[1a] Importing Libraries](https://franciszheng.com/dsper2020/tutorials/tutorial06/importing-libraries) ###Code import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns import statsmodels.api as sm import statsmodels.formula.api as smf from sklearn.discriminant_analysis import LinearDiscriminantAnalysis from sklearn.discriminant_analysis import QuadraticDiscriminantAnalysis from sklearn import neighbors from sklearn import preprocessing from IPython.display import Markdown, display ###Output _____no_output_____ ###Markdown [[1b] Plot and Output Settings](https://franciszheng.com/dsper2020/tutorials/tutorial06/plot-and-output-settings) ###Code # Reset all styles to the default: plt.rcParams.update(plt.rcParamsDefault) # Then make graphs inline: %matplotlib inline # Useful function for Jupyter to display text in bold: def displaybd(text): display(Markdown("**" + text + "**")) # Set custom style settings: # NB: We need to separate "matplotlib inline" call and these settings into different # cells, otherwise the parameters are not set. This is a bug somewhere in Jupyter plt.rcParams['figure.figsize'] = (7, 6) plt.rcParams['font.size'] = 24 plt.rcParams['legend.fontsize'] = 'large' plt.rcParams['figure.titlesize'] = 'large' plt.rcParams['lines.markersize'] = 10 ###Output _____no_output_____ ###Markdown [[1c] Our Dataset](https://franciszheng.com/dsper2020/tutorials/tutorial06/our-dataset) ###Code smarket = pd.read_csv('Smarket.csv', parse_dates=False) # Create direction codes as Up=1 and Down=0 to be sure about the interpretation in regressions: smarket["DirectionCode"] = np.where(smarket["Direction"].str.contains("Up"), 1, 0) display(smarket[1:10]) display(smarket.describe()) displaybd("Correlations matrix:") display(smarket.corr()) smarket["Volume"].plot() plt.xlabel("Day"); plt.ylabel("Volume"); ###Output _____no_output_____ ###Markdown [[2] Logit](https://franciszheng.com/dsper2020/tutorials/tutorial06/logit) [[2a] Running Logit via GLM](https://franciszheng.com/dsper2020/tutorials/tutorial06/running-logit-via-glm) ###Code model = smf.glm("DirectionCode~Lag1+Lag2+Lag3+Lag4+Lag5+Volume", data=smarket, family=sm.families.Binomial()) res = model.fit() display(res.summary()) ###Output _____no_output_____ ###Markdown [[2b] Predicted Probabilities and Confusion Matrix](https://franciszheng.com/dsper2020/tutorials/tutorial06/predicted-probabilities-and-confusion-matrix) ###Code displaybd("Predicted probabilities for the first observations:") DirectionProbs = res.predict() print(DirectionProbs[0:10]) DirectionHat = np.where(DirectionProbs > 0.5, "Up", "Down") confusionDF = pd.crosstab(DirectionHat, smarket["Direction"], rownames=['Predicted'], colnames=['Actual'], margins=True) display(Markdown("***")) displaybd("Confusion matrix:") display(confusionDF) displaybd("Share of correctly predicted market movements:") print(np.mean(smarket['Direction'] == DirectionHat)) ###Output _____no_output_____ ###Markdown [[2c] Estimation of Test Error](https://franciszheng.com/dsper2020/tutorials/tutorial06/estimation-of-test-error) ###Code train = (smarket['Year'] < 2005) smarket2005 = smarket[~train] displaybd("Dimensions of the validation set:") print(smarket2005.shape) model = smf.glm("DirectionCode~Lag1+Lag2+Lag3+Lag4+Lag5+Volume", data=smarket, family=sm.families.Binomial(), subset=train) res = model.fit() DirectionProbsTets = res.predict(smarket2005) DirectionTestHat = np.where(DirectionProbsTets > 0.5, "Up", "Down") displaybd("Share of correctly predicted market movements in 2005:") print(np.mean(smarket2005['Direction'] == DirectionTestHat)) ###Output _____no_output_____ ###Markdown [[3] Linear Discriminant Analysis](https://franciszheng.com/dsper2020/tutorials/tutorial06/linear-discriminant-analysis) [[3a] Custom Output Functions](https://franciszheng.com/dsper2020/tutorials/tutorial06/custom-output-functions) ###Code def printPriorProbabilities(ldaClasses, ldaPriors): priorsDF = pd.DataFrame() for cIdx, cName in enumerate(ldaClasses): priorsDF[cName] = [ldaPriors[cIdx]]; displaybd('Prior probablities of groups:') display(Markdown(priorsDF.to_html(index=False))) def printGroupMeans(ldaClasses, featuresNames, ldaGroupMeans): displaybd("Group means:") groupMeansDF = pd.DataFrame(index=ldaClasses) for fIdx, fName in enumerate(featuresNames): groupMeansDF[fName] = ldaGroupMeans[:, fIdx] display(groupMeansDF) def printLDACoeffs(featuresNames, ldaCoeffs): coeffDF = pd.DataFrame(index=featuresNames) for cIdx in range(ldaCoeffs.shape[0]): colName = "LDA" + str(cIdx + 1) coeffDF[colName] = ldaCoeffs[cIdx] displaybd("Coefficients of linear discriminants:") display(coeffDF) ###Output _____no_output_____ ###Markdown [[3b] Fitting an LDA Model](https://franciszheng.com/dsper2020/tutorials/tutorial06/fitting-an-lda-model) ###Code outcomeName = 'Direction' featuresNames = ['Lag1', 'Lag2']; X_train = smarket.loc[train, featuresNames] y_train = smarket.loc[train, outcomeName] lda = LinearDiscriminantAnalysis() ldaFit = lda.fit(X_train, y_train); printPriorProbabilities(ldaFit.classes_, ldaFit.priors_) printGroupMeans(ldaFit.classes_, featuresNames, ldaFit.means_) printLDACoeffs(featuresNames, ldaFit.coef_) # Coefficients calcualted by Python's LDA are different from R's lda. # But they are proportional: printLDACoeffs(featuresNames, 11.580267503964166 * ldaFit.coef_) # See this: https://stats.stackexchange.com/questions/87479/what-are-coefficients-of-linear-discriminants-in-lda ###Output _____no_output_____ ###Markdown [[3c] LDA Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/lda-predictions) ###Code X_test = smarket2005.loc[~train, featuresNames] y_test = smarket.loc[~train, outcomeName] y_hat = ldaFit.predict(X_test) confusionDF = pd.crosstab(y_hat, y_test, rownames=['Predicted'], colnames=['Actual'], margins=True) displaybd("Confusion matrix:") display(confusionDF) displaybd("Share of correctly predicted market movements:") print(np.mean(y_test == y_hat)) ###Output _____no_output_____ ###Markdown [[3d] Posterior Probabilities](https://franciszheng.com/dsper2020/tutorials/tutorial06/posterior-probabilities) ###Code pred_p = lda.predict_proba(X_test) # pred_p is an array of shape (number of observations) x (number of classes) upNmb = np.sum(pred_p[:, 1] > 0.5) displaybd("Number of upward movements with threshold 0.5: " + str(upNmb)) upNmb = np.sum(pred_p[:, 1] > 0.9) displaybd("Number of upward movements with threshold 0.9: " + str(upNmb)) ###Output _____no_output_____ ###Markdown [[4] Quadratic Discriminant Analysis](https://franciszheng.com/dsper2020/tutorials/tutorial06/quadratic-discriminant-analysis) [[4a] Fitting a QDA Model](https://franciszheng.com/dsper2020/tutorials/tutorial06/fitting-a-qda-model) ###Code qda = QuadraticDiscriminantAnalysis() qdaFit = qda.fit(X_train, y_train); printPriorProbabilities(qdaFit.classes_, qdaFit.priors_) printGroupMeans(qdaFit.classes_, featuresNames, qdaFit.means_) ###Output _____no_output_____ ###Markdown [[4b] QDA Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/qda-predictions) ###Code y_hat = qdaFit.predict(X_test) confusionDF = pd.crosstab(y_hat, y_test, rownames=['Predicted'], colnames=['Actual'], margins=True) displaybd("Confusion matrix:") display(confusionDF) displaybd("Share of correctly predicted market movements:") print(np.mean(y_test == y_hat)) ###Output _____no_output_____ ###Markdown [[5] k-Nearest Neighbors](https://franciszheng.com/dsper2020/tutorials/tutorial06/k-nearest-neighbors) [[5a] One Neighbor](https://franciszheng.com/dsper2020/tutorials/tutorial06/one-neighbor) ###Code knn = neighbors.KNeighborsClassifier(n_neighbors=1) y_hat = knn.fit(X_train, y_train).predict(X_test) confusionDF = pd.crosstab(y_hat, y_test, rownames=['Predicted'], colnames=['Actual'], margins=True) displaybd("Confusion matrix:") display(confusionDF) displaybd("Share of correctly predicted market movements:") print(np.mean(y_test == y_hat)) ###Output _____no_output_____ ###Markdown [[5b] Three Neighbors](https://franciszheng.com/dsper2020/tutorials/tutorial06/three-neighbors) ###Code knn = neighbors.KNeighborsClassifier(n_neighbors=3) y_hat = knn.fit(X_train, y_train).predict(X_test) confusionDF = pd.crosstab(y_hat, y_test, rownames=['Predicted'], colnames=['Actual'], margins=True) displaybd("Confusion matrix:") display(confusionDF) displaybd("Share of correctly predicted market movements:") print(np.mean(y_test == y_hat)) ###Output _____no_output_____ ###Markdown [[6] An Application to Caravan Insurance Data](https://franciszheng.com/dsper2020/tutorials/tutorial06/an-application-to-caravan-insurance-data) [[6a] A New Dataset](https://franciszheng.com/dsper2020/tutorials/tutorial06/a-new-dataset) [[6aa] Loading Our Dataset](https://franciszheng.com/dsper2020/tutorials/tutorial06/loading-our-dataset) ###Code caravan = pd.read_csv('Caravan.csv', index_col=0) display(caravan.describe()) display(caravan.describe(include=[np.object])) ###Output _____no_output_____ ###Markdown [[6ab] Standardizing Our Data](https://franciszheng.com/dsper2020/tutorials/tutorial06/standardizing-our-data) ###Code y = caravan.Purchase X = caravan.drop('Purchase', axis=1).astype('float64') X_scaled = preprocessing.scale(X) ###Output _____no_output_____ ###Markdown [[6ac] Splitting Data into Train and Test Data](https://franciszheng.com/dsper2020/tutorials/tutorial06/splitting-data-into-train-and-test-data) ###Code X_train = X_scaled[1000:,:] y_train = y[1000:] X_test = X_scaled[:1000,:] y_test = y[:1000] ###Output _____no_output_____ ###Markdown [[6b] Using KNN for Prediction](https://franciszheng.com/dsper2020/tutorials/tutorial06/using-knn-for-prediction) ###Code knn = neighbors.KNeighborsClassifier(n_neighbors=1) y_hat = knn.fit(X_train, y_train).predict(X_test) confusionDF = pd.crosstab(y_hat, y_test, rownames=['Predicted'], colnames=['Actual'], margins=True) displaybd("Confusion matrix:") display(confusionDF) displaybd("Share of correctly predicted purchases:") print(np.mean(y_test == y_hat)) ###Output _____no_output_____ ###Markdown [[6c] Logit](https://franciszheng.com/dsper2020/tutorials/tutorial06/logit-1) ###Code X_train_w_constant = sm.add_constant(X_train) X_test_w_constant = sm.add_constant(X_test, has_constant='add') y_train_code = np.where(y_train == "No", 0, 1) res = sm.GLM(y_train_code, X_train_w_constant, family=sm.families.Binomial()).fit() y_hat_code = res.predict(X_test_w_constant) PurchaseHat = np.where(y_hat_code > 0.25, "Yes", "No") confusionDF = pd.crosstab(PurchaseHat, y_test, rownames=['Predicted'], colnames=['Actual'], margins=True) displaybd("Confusion matrix:") display(confusionDF) ###Output _____no_output_____ ###Markdown [[7] More Iris Classification](https://franciszheng.com/dsper2020/tutorials/tutorial06/more-iris-classification) [[7a] Our Dataset](https://franciszheng.com/dsper2020/tutorials/tutorial06/our-dataset-1) [[7aa] Importing Our Dataset](https://franciszheng.com/dsper2020/tutorials/tutorial06/importing-our-dataset) ###Code url = "https://archive.ics.uci.edu/ml/machine-learning-databases/iris/iris.data" names = ['sepal-length', 'sepal-width', 'petal-length', 'petal-width', 'type'] iris_df = pd.read_csv(url, names=names) ###Output _____no_output_____ ###Markdown [[7ab] Splitting Data into Train and Test Data](https://franciszheng.com/dsper2020/tutorials/tutorial06/splitting-data-into-train-and-test-data-1) ###Code X = iris_df.iloc[:, :-1] #attributes, iloc[:, :-1] means until the last column y = iris_df['type'] #labels from sklearn.model_selection import train_test_split X_train, X_test, y_train, y_test = train_test_split(X, y, train_size=0.80) ###Output _____no_output_____ ###Markdown [[7ac] Feature Scaling](https://franciszheng.com/dsper2020/tutorials/tutorial06/feature-scaling) ###Code from sklearn.preprocessing import StandardScaler scaler = StandardScaler() scaler.fit(X_train) X_train = scaler.transform(X_train) X_test = scaler.transform(X_test) ###Output _____no_output_____ ###Markdown [[7b] Logit](https://franciszheng.com/dsper2020/tutorials/tutorial06/logit-2) [[7ba] Fitting Our Model](https://franciszheng.com/dsper2020/tutorials/tutorial06/fitting-our-model) ###Code from sklearn.linear_model import LogisticRegression logit_model = LogisticRegression() logit_model.fit(X_train, y_train) ###Output _____no_output_____ ###Markdown [[7bb] Making Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/making-predictions) ###Code y_pred = logit_model.predict(X_test) ###Output _____no_output_____ ###Markdown [[7bc] Evaluating Our Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/evaluating-our-predictions) ###Code from sklearn.metrics import classification_report print(classification_report(y_test, y_pred)) from sklearn.metrics import confusion_matrix cm = confusion_matrix(y_test, y_pred) cm_df = pd.DataFrame(cm, index = ['setosa','versicolor','virginica'], columns = ['setosa','versicolor','virginica']) sns.heatmap(cm_df, annot=True) plt.ylabel('Actual') plt.xlabel('Predicted') plt.show() ###Output _____no_output_____ ###Markdown [[7c] Linear Discriminant Analysis](https://franciszheng.com/dsper2020/tutorials/tutorial06/linear-discriminant-analysis-1) [[7ca] Fitting Our Model](https://franciszheng.com/dsper2020/tutorials/tutorial06/fitting-our-model-1) ###Code lda_model = LinearDiscriminantAnalysis() lda_model.fit(X_train, y_train); ###Output _____no_output_____ ###Markdown [[7cb] Making Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/making-predictions-1) ###Code y_pred = lda_model.predict(X_test) ###Output _____no_output_____ ###Markdown [[7cd] Evaluating Our Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/evaluating-our-predictions-1) ###Code print(classification_report(y_test, y_pred)) cm = confusion_matrix(y_test, y_pred) cm_df = pd.DataFrame(cm, index = ['setosa','versicolor','virginica'], columns = ['setosa','versicolor','virginica']) sns.heatmap(cm_df, annot=True) plt.ylabel('Actual') plt.xlabel('Predicted') plt.show() ###Output _____no_output_____ ###Markdown [[7d] Quadratic Discriminant Analysis](https://franciszheng.com/dsper2020/tutorials/tutorial06/quadratic-discriminant-analysis-1) [[7da] Fitting Our Model](https://franciszheng.com/dsper2020/tutorials/tutorial06/fitting-our-model-2) ###Code qda_model = QuadraticDiscriminantAnalysis() qda_model.fit(X_train, y_train) ###Output _____no_output_____ ###Markdown [[7db] Making Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/making-predictions-2) ###Code y_pred = qda_model.predict(X_test) ###Output _____no_output_____ ###Markdown [[7dc] Evaluating Our Predictions](https://franciszheng.com/dsper2020/tutorials/tutorial06/evaluating-our-predictions-2) ###Code print(classification_report(y_test, y_pred)) cm = confusion_matrix(y_test, y_pred) cm_df = pd.DataFrame(cm, index = ['setosa','versicolor','virginica'], columns = ['setosa','versicolor','virginica']) sns.heatmap(cm_df, annot=True) plt.ylabel('Actual') plt.xlabel('Predicted') plt.show() ###Output _____no_output_____
RESEARCH/melanoma_classification.ipynb
###Markdown Inroduction Importing Libraries ###Code import warnings warnings.filterwarnings('ignore') import os import datetime import pandas as pd import numpy as np import matplotlib.pyplot as plt from matplotlib.pyplot import * %matplotlib inline import seaborn as sns sns.set(style='darkgrid', color_codes=True) import tensorflow as tf from keras.optimizers import Adam, RMSprop from keras.preprocessing.image import ImageDataGenerator, array_to_img, img_to_array, load_img from keras.layers import Conv2D, MaxPooling2D, Flatten, Dense, Dropout from keras.utils import plot_model from keras.models import Sequential from keras.applications.vgg16 import VGG16 from sklearn.metrics import classification_report, confusion_matrix, precision_score, recall_score, f1_score # set random seed to 42 which is used to replicate the same result everytime , the output will be same if i run the cell in future np.random.seed(42) tf.random.set_seed(42) ###Output _____no_output_____ ###Markdown ###Code ###Output _____no_output_____ ###Markdown Defining the Functions ###Code # Function to create a dictionary of model's train, validation and test results def store_results_to_dict(model, model_description): from sklearn.metrics import precision_score, recall_score, f1_score train_steps_per_epoch = np.math.ceil(train_generator.samples / train_generator.batch_size) val_steps_per_epoch = np.math.ceil(val_generator.samples / val_generator.batch_size) test_steps_per_epoch = np.math.ceil(test_generator.samples / test_generator.batch_size) train_loss, train_acc = model.evaluate_generator(train_generator, steps=train_steps_per_epoch) val_loss, val_acc = model.evaluate_generator(val_generator, steps=val_steps_per_epoch) test_loss, test_acc = model.evaluate_generator(test_generator, steps=test_steps_per_epoch) pred = model.predict_generator(test_generator, test_steps_per_epoch) pred_classes = np.round(pred) true_classes = test_generator.classes class_labels = list(test_generator.class_indices.keys()) precision_score = precision_score(true_classes, pred_classes) recall_score = recall_score(true_classes, pred_classes) f1_score = f1_score(true_classes, pred_classes) curr_dict = { 'Model':model_description ,'Train Accuracy': round(train_acc, 4) ,'Train Loss': round(train_loss, 4) ,'Validation Accuracy':round(val_acc, 4) ,'validation Loss':round(val_loss, 4) ,'Test Accuracy':round(test_acc, 4) ,'Test Loss':round(test_loss, 4) ,'Precision':round(precision_score, 4) ,'Recall':round(recall_score, 4) ,'f1':round(f1_score, 4) } return curr_dict # Function to plot the accuracy and loss of the model def plot_acc_and_loss(model_history): acc = model_history.history['acc'] val_acc = model_history.history['val_acc'] loss = model_history.history['loss'] val_loss = model_history.history['val_loss'] epochs = range(len(acc)) plt.figure(figsize=(16,7)) plt.plot(epochs, acc, 'bo', label='Training acc') plt.plot(epochs, val_acc, 'g', label='Validation acc') plt.title('Training and validation accuracy',fontsize=20 ) plt.legend() plt.figure(figsize=(16,7)) plt.plot(epochs, loss, 'bo', label='Training loss') plt.plot(epochs, val_loss, 'g', label='Validation loss') plt.title('Training and validation loss', fontsize=20) plt.legend() plt.show() ###Output _____no_output_____ ###Markdown About the data3297 total skin lesion images from ISIC-Archive(https://www.isic-archive.com/!/topWithHeader/wideContentTop/main) are used in the model. The files at https://www.kaggle.com/fanconic/skin-cancer-malignant-vs-benign are uploaded. A validation set is made from the images to train the model at each epoch. Directories ###Code # train, test and validation directories of melanoma images train_data_dir=('/content/drive/My Drive/melanoma_image_classification/data/train') val_data_dir=('/content/drive/My Drive/melanoma_image_classification/data/val') test_data_dir=('/content/drive/My Drive/melanoma_image_classification/data/test') ###Output _____no_output_____ ###Markdown Numbers of Train, Test and Validation Images ###Code # counting number of benign images and malignant images in each dataset of training, testing and validation path='/content/drive/My Drive/melanoma_image_classification/data/' # Numbers of benign and malignant images in each set for folder in ['train', 'val', 'test']: n_malignant = len(os.listdir(path + folder + '/malignant')) n_benign = len(os.listdir(path + folder + '/benign')) print("There are {} benign skin images and {} malignant skin images in {} set. ".format(n_benign, n_malignant, folder)) total_images_train_benign = os.listdir(path + 'train/benign') total_images_train_malignant = os.listdir(path + 'train/malignant') total_images_test_benign = os.listdir(path + 'test/benign') total_images_test_malignant = os.listdir(path + 'test/malignant') total_images_val_benign = os.listdir(path + 'val/benign') total_images_val_malignant = os.listdir(path + 'val/malignant') ###Output _____no_output_____ ###Markdown Visualization of distributions ###Code # plot shows the number of normal images versus pneumonia images in Train, Val and test set plt.figure(figsize=(20, 7)) plt.subplot(131) plt.title('Train Data Distribution', fontsize=14) sns.barplot(x=['benign','malignant'],y=[len(total_images_train_benign),len(total_images_train_malignant)]) plt.subplot(132) plt.title('Test Data Distribution', fontsize=14) sns.barplot(x=['benign','malignant'],y=[len(total_images_test_benign),len(total_images_test_malignant)]) plt.subplot(133) plt.title('Validation Data Distribution', fontsize=14) sns.barplot(x=['benign','malignant'],y=[len(total_images_val_benign),len(total_images_val_malignant)]) plt.show() ###Output _____no_output_____ ###Markdown Display Images ###Code # displaying normal x-ray images and pneumonia x-ray images in train, test and val using imshow,imread fig, ax = plt.subplots(2, 3, figsize=(18, 9)) ax = ax.ravel() fig.suptitle('Benign and Malignant Skin Images', fontsize=24) for i, _set in enumerate(['train', 'val', 'test']): set_path = path+_set ax[i].imshow(plt.imread(set_path+'/benign/'+os.listdir(set_path+'/benign')[0]), cmap='gray') ax[i].set_title('{} set: benign'.format(_set)) ax[i+3].imshow(plt.imread(set_path+'/malignant/'+os.listdir(set_path+'/malignant')[0]), cmap='gray') ax[i+3].set_title('{} set: malignant'.format(_set)) ###Output _____no_output_____ ###Markdown Define generators ###Code # Get all the data in the directory DATA/TRAIN, and reshape them print("Train data:") train_generator = ImageDataGenerator(rescale=1./255).flow_from_directory(train_data_dir, target_size=(150, 150), batch_size=32, class_mode='binary') # Get all the data in the directory DATA/TEST , and reshape them print("Test data:") test_generator = ImageDataGenerator(rescale=1./255).flow_from_directory(test_data_dir, target_size=(150, 150), batch_size=32, class_mode='binary', shuffle=False) # Get all the data in the directory DATA/VAL, and reshape them print("Validation data:") val_generator = ImageDataGenerator(rescale=1./255).flow_from_directory(val_data_dir, target_size=(150, 150), batch_size=32, class_mode='binary') ###Output Train data: Found 2141 images belonging to 2 classes. Test data: Found 660 images belonging to 2 classes. Validation data: Found 496 images belonging to 2 classes. ###Markdown Calculate the step size per epoch ###Code # counting the stepsize per epoch train_steps_per_epoch = np.math.ceil(train_generator.samples / train_generator.batch_size) val_steps_per_epoch = np.math.ceil(val_generator.samples / val_generator.batch_size) test_steps_per_epoch = np.math.ceil(test_generator.samples / test_generator.batch_size) ###Output _____no_output_____ ###Markdown Define the parameters ###Code # Define the parameters of image transformation train_datagen = ImageDataGenerator(rescale=1./255,rotation_range=40, # rotating from -40 to40 degree width_shift_range=0.2, # allowing image generator to shift our images left or right about 20% of total width height_shift_range=0.2,#allowing image generator to shift our images up or dowm about 20% of total height shear_range=0.2, zoom_range=0.2, horizontal_flip=True, fill_mode='nearest') test_datagen = ImageDataGenerator(rescale=1./255) train_generator = train_datagen.flow_from_directory( # This is the target directory train_data_dir, # All images will be resized to 150x150 target_size=(150, 150), batch_size=32, # Since we use binary_crossentropy loss, we need binary labels class_mode='binary') val_generator = test_datagen.flow_from_directory(val_data_dir, target_size=(150, 150), batch_size=32, class_mode='binary') ###Output Found 2141 images belonging to 2 classes. Found 496 images belonging to 2 classes. ###Markdown VGG16 Model ###Code conv_base = VGG16(weights='imagenet',#imagenet is the final weight of vgg16 include_top=False, #meaning we will import only the conv base not the whole thing input_shape=(150,150,3)) conv_base.summary() model1=Sequential() model1.add(conv_base) #adding conv_base that was imported model1.add(Flatten()) model1.add(Dense(256,activation='relu'))#adding fully connected dense layer with 256 neurons model1.add(Dense(1,activation='sigmoid')) # output layer with single neuron #structure of model1 plot_model(model1) # looking at the structure of my neural network ###Output _____no_output_____ ###Markdown Compile and train the model ###Code model1.compile(loss='binary_crossentropy', optimizer=RMSprop(lr=2e-5),# small learning rate used because weights already optimized metrics=['acc']) # set random seed to 42 to is used to replicate the same result everytime , the output will be same if i run the cell in future np.random.seed(42) tf.random.set_seed(42) # ⏰ This cell may take several minutes to run start = datetime.datetime.now() history = model1.fit_generator(train_generator, steps_per_epoch=train_steps_per_epoch, epochs=23, validation_data=val_generator, validation_steps=val_steps_per_epoch) end = datetime.datetime.now() elapsed = end - start print('---------Elapsed Time-----------') print('Time to fit the VGG16 model is:\n {}'.format(elapsed)) ###Output Epoch 1/23 67/67 [==============================] - 24s 359ms/step - loss: 0.4692 - acc: 0.7753 - val_loss: 0.3795 - val_acc: 0.8125 Epoch 2/23 67/67 [==============================] - 24s 360ms/step - loss: 0.3615 - acc: 0.8202 - val_loss: 0.4405 - val_acc: 0.7823 Epoch 3/23 67/67 [==============================] - 24s 358ms/step - loss: 0.3240 - acc: 0.8538 - val_loss: 0.3847 - val_acc: 0.8125 Epoch 4/23 67/67 [==============================] - 24s 356ms/step - loss: 0.3112 - acc: 0.8566 - val_loss: 0.3394 - val_acc: 0.8387 Epoch 5/23 67/67 [==============================] - 24s 356ms/step - loss: 0.2850 - acc: 0.8627 - val_loss: 0.4044 - val_acc: 0.8085 Epoch 6/23 67/67 [==============================] - 24s 355ms/step - loss: 0.2722 - acc: 0.8716 - val_loss: 0.4397 - val_acc: 0.7782 Epoch 7/23 67/67 [==============================] - 23s 350ms/step - loss: 0.2584 - acc: 0.8814 - val_loss: 0.5426 - val_acc: 0.7177 Epoch 8/23 67/67 [==============================] - 24s 356ms/step - loss: 0.2528 - acc: 0.8865 - val_loss: 0.5782 - val_acc: 0.7177 Epoch 9/23 67/67 [==============================] - 24s 351ms/step - loss: 0.2391 - acc: 0.8930 - val_loss: 0.6174 - val_acc: 0.7198 Epoch 10/23 67/67 [==============================] - 23s 349ms/step - loss: 0.2329 - acc: 0.9019 - val_loss: 0.5095 - val_acc: 0.7419 Epoch 11/23 67/67 [==============================] - 23s 349ms/step - loss: 0.2129 - acc: 0.9010 - val_loss: 0.3976 - val_acc: 0.8266 Epoch 12/23 67/67 [==============================] - 23s 347ms/step - loss: 0.2081 - acc: 0.9047 - val_loss: 0.5241 - val_acc: 0.7823 Epoch 13/23 67/67 [==============================] - 23s 345ms/step - loss: 0.1928 - acc: 0.9220 - val_loss: 0.5147 - val_acc: 0.7581 Epoch 14/23 67/67 [==============================] - 23s 345ms/step - loss: 0.1938 - acc: 0.9183 - val_loss: 0.3532 - val_acc: 0.8569 Epoch 15/23 67/67 [==============================] - 23s 346ms/step - loss: 0.1937 - acc: 0.9159 - val_loss: 0.8103 - val_acc: 0.7359 Epoch 16/23 67/67 [==============================] - 23s 342ms/step - loss: 0.1786 - acc: 0.9141 - val_loss: 0.4237 - val_acc: 0.7984 Epoch 17/23 67/67 [==============================] - 23s 342ms/step - loss: 0.1736 - acc: 0.9164 - val_loss: 0.5471 - val_acc: 0.8226 Epoch 18/23 67/67 [==============================] - 23s 342ms/step - loss: 0.1706 - acc: 0.9253 - val_loss: 0.5121 - val_acc: 0.7601 Epoch 19/23 67/67 [==============================] - 23s 340ms/step - loss: 0.1603 - acc: 0.9271 - val_loss: 0.5804 - val_acc: 0.7460 Epoch 20/23 67/67 [==============================] - 23s 341ms/step - loss: 0.1599 - acc: 0.9351 - val_loss: 0.5521 - val_acc: 0.7560 Epoch 21/23 67/67 [==============================] - 23s 347ms/step - loss: 0.1383 - acc: 0.9430 - val_loss: 0.4002 - val_acc: 0.8185 Epoch 22/23 67/67 [==============================] - 23s 339ms/step - loss: 0.1400 - acc: 0.9383 - val_loss: 0.8403 - val_acc: 0.7661 Epoch 23/23 67/67 [==============================] - 23s 340ms/step - loss: 0.1462 - acc: 0.9454 - val_loss: 0.4531 - val_acc: 0.8044 ---------Elapsed Time----------- Time to fit the VGG16 model is: 0:09:06.069182 ###Markdown Store the results ###Code results=store_results_to_dict(model1, 'Base Model(with 4 conv,pooling layers,and dense full conv layer)') results_final = [] # appending the results of the new model results_final.append(results) # putting the results in dataframe df_model_results = pd.DataFrame(results_final) df_model_results ###Output _____no_output_____ ###Markdown Plot the results ###Code plot_acc_and_loss(history) ###Output _____no_output_____ ###Markdown Prediction results on the test setWe got 87% accuracy in classifying benign and malignant melanoma images. ###Code pred = model1.predict_generator(test_generator, test_steps_per_epoch) pred_classes = np.round(pred) true_classes = test_generator.classes class_labels = list(test_generator.class_indices.keys()) print("Baseline Model:\n") print("Confusion Matrix:\n", confusion_matrix(true_classes, pred_classes)) print("----------------------------------------------------") print("Classification Report:\n", classification_report(true_classes, pred_classes, target_names=class_labels)) ###Output Baseline Model: Confusion Matrix: [[325 35] [ 54 246]] ---------------------------------------------------- Classification Report: precision recall f1-score support benign 0.86 0.90 0.88 360 malignant 0.88 0.82 0.85 300 accuracy 0.87 660 macro avg 0.87 0.86 0.86 660 weighted avg 0.87 0.87 0.86 660 ###Markdown Conclusion In this project, we classified the melanoma images as beningn and malignant. As for NF patients we are able to make a smaller classification model and classify, segment and track the skin lessions when we get the image data. ###Code ###Output _____no_output_____
w3/w3-day_1/Seaborn_working_with.ipynb
###Markdown Regression line with Confidence interval ###Code # a regression line with a confidence inetrval plot sns.lmplot(x='total_bill', y='tip', data = tips) plt.show() sns.lmplot(x='total_bill',y='tip', data = tips, hue = 'sex') plt.show() sns.lmplot(x='total_bill',y='tip', data = tips, col = 'sex') plt.show() ###Output _____no_output_____ ###Markdown Residual Plot ###Code sns.residplot(x='total_bill',y='tip', data = tips) plt.show() ###Output _____no_output_____ ###Markdown Strip plot ###Code sns.stripplot(x='day',y ='tip',data = tips, size =5, jitter = True) # overwrite the original y-label from 'tip' to 'tip ($)' plt.ylabel('tip ($)') plt.show() ###Output _____no_output_____ ###Markdown Swarm plot ###Code sns.swarmplot(x='day',y='tip',data = tips, hue = 'sex') # overwrite the original y-label from 'tip' to 'tip ($)' plt.ylabel('tip ($)') plt.show() ###Output /Users/louisrossi/opt/anaconda3/lib/python3.8/site-packages/seaborn/categorical.py:1296: UserWarning: 6.5% of the points cannot be placed; you may want to decrease the size of the markers or use stripplot. warnings.warn(msg, UserWarning) /Users/louisrossi/opt/anaconda3/lib/python3.8/site-packages/seaborn/categorical.py:1296: UserWarning: 5.7% of the points cannot be placed; you may want to decrease the size of the markers or use stripplot. warnings.warn(msg, UserWarning) ###Markdown Boxplot + Violin plot ###Code # create a grid with 1 row 2 columns plt.subplot(1,2,1) # the first plot sns.boxplot(x='day',y='tip', data = tips) plt.ylabel('tip ($)') plt.subplot(1,2,2) # the second plot sns.violinplot(x='day',y='tip', data = tips) plt.ylabel('tip ($)') plt.tight_layout() plt.show() ###Output _____no_output_____ ###Markdown Joint plot ###Code sns.jointplot(x='total_bill', y='tip', data = tips) plt.show() sns.jointplot(x='total_bill', y='tip', data = tips, kind = 'kde') plt.show() ###Output _____no_output_____ ###Markdown Pair plot ###Code sns.pairplot(tips) plt.show() sns.pairplot(tips, hue = 'sex') plt.show() ###Output _____no_output_____ ###Markdown Heatmap ###Code # compute correlations between features df_corr = tips.corr() # plot the correlations sns.heatmap(df_corr) plt.title('Correlation plot') plt.show() ###Output _____no_output_____
successful_runs/comparisons-to-default/flu-MARL-[MAC, Default Cmp].ipynb
###Markdown k-vs-(N-k) Flu ABM Env- k-vs-(N-k) experiment- Kicking tires on multiplayer instance of Flu ABM with RL learners - MADDPG/MAC RL algo ###Code import itertools, importlib, sys, warnings, os import numpy as np import pandas as pd import matplotlib.pyplot as plt import seaborn as sns # ML libs import tensorflow as tf print("Tensorflow version:", tf.__version__) # warnings.filterwarnings("ignore") log_path = './log/flu' #tensorboard --logdir=flugame_worker_1:'./log/train_rf_flugame_worker' ## suppress annoy verbose tf msgs warnings.filterwarnings("ignore") os.environ['TF_CPP_MIN_LOG_LEVEL'] = '2' # '3' to block all including error msgs tf.compat.v1.logging.set_verbosity(tf.compat.v1.logging.ERROR) sys.path.append('./embodied_arch') import embodied_arch.embodied_central_Qcritic as emac importlib.reload(emac) import flumodel_python.flu_env as Fenv from embodied_misc import ActionPolicyNetwork, ValueNetwork, SensoriumNetworkTemplate ###Output _____no_output_____ ###Markdown Env Setup ###Code # exos = [1,2,3,10] # (np.random.sample(33) < 0.3) exos = (np.random.sample(9223) < 0.004) exos = [j for j in range(len(exos)) if exos[j]==True] print(len(exos)) importlib.reload(Fenv); importlib.reload(emac); tf.reset_default_graph() flu_menv = Fenv.Flu_env( exo_idx=exos, model_path="./flumodel_python/" ) print(flu_menv.actor_count) print(flu_menv.state_space_size, flu_menv.action_space_size) ###Output 38 8 1 ###Markdown MARL Setup Demo ###Code actor = lambda s: ActionPolicyNetwork(s, hSeq=(8,), gamma_reg=1e-1) value = lambda s: ValueNetwork(s, hSeq=(8,), gamma_reg=1.) sensor = lambda st, out_dim: SensoriumNetworkTemplate(st, hSeq=(16,8,8,), out_dim=out_dim, gamma_reg=5.) # num_episodes, n_epochs, max_len = (5, 4, 5) # num_episodes, n_epochs, max_len = (100, 1501, 35) # num_episodes, max_len, n_epochs, evry = (25, 10, 121, 40) num_episodes, max_len, n_epochs, evry = (100, 35, 451, 50) flumac = emac.EmbodiedAgent_MAC( name="flu_MAC", env_=flu_menv, alpha_p=150., alpha_v=5., alpha_q=2.5, actorNN=actor, valueNN=value, sensorium=sensor,latentDim=4, max_episode_length=max_len, _every_=evry ) # ??flumac.play # (flumac.a_size, flumac.env.action_space_size) sess = tf.InteractiveSession() flumac.init_graph(sess) # note tboard log dir saver = tf.train.Saver(max_to_keep=1) ###Output Tensorboard logs in: ./log/train_flu_MAC ###Markdown Baseline Baseline for RL/Adaptive Behavioral Model ###Code print('Baselining untrained pnet...') rwds0 = [] acts_cov = np.zeros([flumac.actor_count,flumac.actor_count]) for k in range(num_episodes): flumac.play(sess, terminal_reward=0.); rwds0.append(flumac.last_total_returns) actions = np.array(flumac.episode_buffer['actions']).T acts_cov = acts_cov + (np.cov(actions)/num_episodes) print("\rEpisode {}/{}".format(k, num_episodes),end="") # Compute average rewards base_perf = 100.*np.mean(np.array(rwds0)/float(flumac.max_episode_length)) base_per_agent = 100.*np.mean(np.array(rwds0)/float(flumac.max_episode_length), axis=0) print("\nAgent is flu-free for an average of {}pct of seasons".format( 1.*base_perf)) acts_corr = acts_cov.copy() jm, km = acts_corr.shape for j in range(jm): for k in range(km): denom = np.sqrt((acts_corr[j,j])*(acts_corr[k,k])) acts_corr[j,k] = acts_corr[j,k]/denom print("Agent Action Correlations:") sns.heatmap(acts_corr, center=0) ###Output Agent Action Correlations: ###Markdown Train Agent Population ###Code obs = [] for ct in range(50): flumac.play(sess) tmp = flumac.train_eval_QC(sess) obs.append(np.mean(tmp, axis=0)) if ct%25==0: print('\r\tIteration {}: Value loss({})'.format( ct, np.mean(tmp)), end="") plt.plot(obs[1:]); # ### Train Agents print('Training...') hist = flumac.work(sess, num_epochs=n_epochs, saver=saver) ###Output Training... Starting agent flu_MAC Epoch no.: 0/451 Stats @Step 0: (Min, Mean, Max) Perf/Recent Rewards: (25.0, 32.55263157894737, 35.0) Losses/Policy LLs: (-2.2608888, -0.60087085, -0.11010261) Losses/Policy Entropies: (0.33433878, 0.6049257, 0.6931392) Values/Critic Scores: (-4.1316667, 0.23619446, 3.2716048) Values/Mean Q Scores: (13.148774, 17.42588, 19.983746) Saved Model Epoch no.: 50/451 Stats @Step 50: (Min, Mean, Max) Perf/Recent Rewards: (27.0, 32.973684210526315, 35.0) Losses/Policy LLs: (-2.8019228, -0.58850634, -0.06261318) Losses/Policy Entropies: (0.22887048, 0.59595674, 0.6928854) Values/Critic Scores: (-3.635611, 0.1409748, 3.0927284) Values/Mean Q Scores: (21.384138, 25.468422, 27.90224) Saved Model Epoch no.: 100/451 Stats @Step 100: (Min, Mean, Max) Perf/Recent Rewards: (26.0, 32.921052631578945, 35.0) Losses/Policy LLs: (-2.313823, -0.5986697, -0.104119614) Losses/Policy Entropies: (0.3226206, 0.6056469, 0.6927041) Values/Critic Scores: (-6.481694, 0.041140266, 2.7762804) Values/Mean Q Scores: (19.315922, 22.823996, 25.531332) Saved Model Epoch no.: 150/451 Stats @Step 150: (Min, Mean, Max) Perf/Recent Rewards: (27.0, 33.078947368421055, 35.0) Losses/Policy LLs: (-2.1723638, -0.6047279, -0.12093454) Losses/Policy Entropies: (0.3546087, 0.59984016, 0.69303197) Values/Critic Scores: (-4.421058, -0.036370203, 4.140975) Values/Mean Q Scores: (17.072886, 21.429373, 24.028694) Saved Model Epoch no.: 200/451 Stats @Step 200: (Min, Mean, Max) Perf/Recent Rewards: (30.0, 33.3421052631579, 35.0) Losses/Policy LLs: (-3.0960093, -0.5981115, -0.046284117) Losses/Policy Entropies: (0.18422109, 0.5904095, 0.69313556) Values/Critic Scores: (-5.4957156, 0.055095013, 4.2656927) Values/Mean Q Scores: (18.929699, 21.962545, 23.63159) Saved Model Epoch no.: 250/451 Stats @Step 250: (Min, Mean, Max) Perf/Recent Rewards: (25.0, 33.05263157894737, 35.0) Losses/Policy LLs: (-2.9169164, -0.59213865, -0.0556187) Losses/Policy Entropies: (0.21041551, 0.5970644, 0.69302166) Values/Critic Scores: (-3.5398364, 0.1870522, 4.429839) Values/Mean Q Scores: (21.011278, 25.063356, 26.982113) Saved Model Epoch no.: 300/451 Stats @Step 300: (Min, Mean, Max) Perf/Recent Rewards: (28.0, 33.26315789473684, 35.0) Losses/Policy LLs: (-2.35666, -0.5936334, -0.099528804) Losses/Policy Entropies: (0.3133606, 0.6017741, 0.6930785) Values/Critic Scores: (-2.9842036, 0.15928702, 3.469826) Values/Mean Q Scores: (19.607126, 24.719528, 27.827423) Saved Model Epoch no.: 350/451 Stats @Step 350: (Min, Mean, Max) Perf/Recent Rewards: (28.0, 32.921052631578945, 35.0) Losses/Policy LLs: (-2.349049, -0.5890703, -0.10032864) Losses/Policy Entropies: (0.3149911, 0.59216577, 0.6927165) Values/Critic Scores: (-4.4624715, 0.26040095, 3.1501908) Values/Mean Q Scores: (15.378132, 23.509571, 26.322552) Saved Model Epoch no.: 400/451 Stats @Step 400: (Min, Mean, Max) Perf/Recent Rewards: (27.0, 33.23684210526316, 35.0) Losses/Policy LLs: (-2.4175484, -0.5871441, -0.093365945) Losses/Policy Entropies: (0.30054313, 0.57564735, 0.69309914) Values/Critic Scores: (-5.593222, 0.39028728, 3.8277018) Values/Mean Q Scores: (18.914165, 23.568161, 26.058767) Saved Model Epoch no.: 450/451 Stats @Step 450: (Min, Mean, Max) Perf/Recent Rewards: (25.0, 32.89473684210526, 35.0) Losses/Policy LLs: (-2.6269681, -0.5628719, -0.075043984) Losses/Policy Entropies: (0.25954115, 0.56649894, 0.69313854) Values/Critic Scores: (-5.78232, 0.4455961, 4.4236746) Values/Mean Q Scores: (18.889389, 23.88217, 26.022322) Saved Model ###Markdown Test ###Code # Test pnet! print('Testing...') rwds = [] acts_cov_trained = np.zeros([flumac.actor_count,flumac.actor_count]) for k in range(num_episodes): flumac.play(sess) rwds.append(flumac.last_total_returns) actions = np.array(flumac.episode_buffer['actions']).T acts_cov_trained = acts_cov_trained + (np.cov(actions)/num_episodes) print("\rEpisode {}/{}".format(k, num_episodes),end="") trained_perf = 100.*np.mean(np.array(rwds)/float(flumac.max_episode_length)) trained_per_agent = 100.*np.mean(np.array(rwds)/float(flumac.max_episode_length), axis=0) print("\nAgent is flu-free for an average of {} pct compared to baseline of {} pct".format( 1.*trained_perf, 1.*base_perf) ) acts_corr_trained = acts_cov_trained.copy() jm, km = acts_corr_trained.shape for j in range(jm): for k in range(km): denom = np.sqrt((acts_cov_trained[j,j])*(acts_cov_trained[k,k])) acts_corr_trained[j,k] = acts_corr_trained[j,k]/denom mask = np.zeros_like(acts_corr_trained) mask[np.triu_indices_from(mask,k=0)] = True with sns.axes_style("darkgrid"): plt.rcParams['figure.figsize'] = (15, 12) ax = sns.heatmap(acts_corr_trained, mask=mask, vmax=0.125, center=0) ax.set_ylabel("Agent Index") ax.set_xlabel("Agent Index") ax.set_title("Action Correlations") ###Output _____no_output_____ ###Markdown Evaluate ###Code rwds0_df = pd.DataFrame(100.*(np.array(rwds0)/float(flumac.max_episode_length))) rwds_df = pd.DataFrame(100.*(np.array(rwds)/float(flumac.max_episode_length))) rwds0_df['Wave'] = "Baseline" rwds_df['Wave'] = "Trained" resDF = pd.concat([rwds0_df, rwds_df]) resDF.columns = ["Agent"+str(tc) if tc is not "Wave" else tc for tc in resDF.columns] # resDF['id'] = resDF.index print(resDF.shape) # resDF.head() resDF = resDF.melt( id_vars=['Wave'], #['id', 'Wave'], value_vars=[tc for tc in resDF.columns if "Agent" in tc] ) resDF = resDF.rename(columns={"variable": "Agent", "value": "Immune_pct"}) print(resDF.shape) res_tabs = resDF.groupby(['Agent','Wave']).aggregate(['mean','std']) # res_tabs # resDF.head() plt.rcParams['figure.figsize'] = (9, 35) sns.set(font_scale=1.25) fig = sns.violinplot(data=resDF, inner="box", cut=0, x="Immune_pct", y="Agent", hue="Wave", split=True); fig.set_title( 'Average Episode Rewards: Baseline vs Trained Agents.'); fig.legend(loc='upper left'); base_meanDF = resDF[resDF.Wave=="Baseline"].groupby(['Agent']).aggregate(['mean']) base_meanDF.sort_index(inplace=True) trained_meanDF = resDF[resDF.Wave=="Trained"].groupby(['Agent']).aggregate(['mean']) trained_meanDF.sort_index(inplace=True) mean_diffDF = (trained_meanDF - base_meanDF) mean_diffDF.columns = ['Mean_Immune_Pct_Change'] # mean_diffDF.head() plt.rcParams['figure.figsize'] = (9, 19) sns.set_color_codes("dark") fig, axs = plt.subplots(2,1, sharex=True, gridspec_kw={'height_ratios': [1,4]}) cmp = sns.violinplot(x='Mean_Immune_Pct_Change', cut=0, inner='quartile', data=mean_diffDF, ax=axs[0]) axs[0].set_ylabel('Agent Aggregate'); axs[0].set_title( 'Distribution of Changes in Flu Immunity Rates:\nIn Aggregate & Per-Agent.' ); sns.barplot(y=mean_diffDF.index, x="Mean_Immune_Pct_Change", data=mean_diffDF, color="r", label="Success Rate", ax=axs[1]); plt.subplots_adjust(wspace=0, hspace=0) axs[1].set_xlabel('Avg. Change in Immunity Rates'); ###Output _____no_output_____ ###Markdown Baseline for Default Behavioral Model ###Code import flumodel_python.flu_env_basic as FABM # ?Fenv.Flu_ABM importlib.reload(FABM); flu = FABM.Flu_ABM(model_path="./flumodel_python/") # Burn-in Flu ABM First... for _ in range(30): _ = flu.stepAll() flu_hist = np.zeros([num_episodes, len(exos)]) for k in range(num_episodes): tmp = np.zeros(len(exos)) for _ in range(max_len): tmp += (1.-np.array(flu.stepAll(), dtype=float))[exos] flu_hist[k,:] = tmp rwds_dbm = 100.*flu_hist/float(max_len) print(len(exos), rwds_dbm.shape, np.mean(rwds_dbm, axis=0).shape) print(np.mean(rwds_dbm), "\n", np.mean(rwds_dbm, axis=0) ) plt.rcParams['figure.figsize'] = (8,3) sns.boxplot(np.mean(rwds_dbm, axis=0) - base_per_agent) ###Output _____no_output_____ ###Markdown Compare to Default Behavioral Model ###Code trcmp = 100.*(np.array(rwds)/float(flumac.max_episode_length)) cmp = np.mean((trcmp-rwds_dbm), axis=0) bplot = sns.boxplot(cmp) bplot.set_title( 'Pct Improvement in Flu Outcomes\nRL Behaviors over Default Behavioral Model'); plt.rcParams['figure.figsize'] = (8,3) np.mean(trcmp - np.mean(rwds_dbm, axis=0)) ###Output _____no_output_____
Normal_Equation/Exploration of Data.ipynb
###Markdown Exploritory Analysis on SHPO OLI Dataset.By Kellen Bullock Framing the Problem and Big PictureThe OLI is a database that contains roughly 65,000 records. Roughly up to 20,000 of these records are dupliates. Several people have been cleaning the data by hand by indicating poss_dup or good in the duplicate_check field. The State Historical Presvation Office then makes the decision to roll those records off the main table into another. In order to make more informed deicsions and queries on the database duplicate records should be taken out. Duplicate create confusion for the users by providing redudent information that the user does not want. Problem StatementHow can automatic complex duplicate record detection be implemented to prevent more duplicate records from entering the dataset? Proposed solution:Training a neural network to make a propblistic descision on whether a record is a duplicate or not is far more effeicent than checking every record in the dataset. Since several thousand records have been classified already that data will be used for supervised learning. How to solve this problem manually:A field such as PROPNAME, ADDRESS or RESNAME are sorted alphabetically. First the PROPNAME fields are compared for similarity. Then ADDRESS, RESNAME, ROOF_TYP, and WINDOW_TYPE. If records matchup the records are marked as poss_dup in the duplicate_check column. Peformance measurment:A CSV file that contains the OBJECTID and probability score that the classifier assigns will be out. This will be joined with the main table dataset and checked by a person. Accuracy measrument: RSMERecall will also be measruedIf the Classifier can out perform at 85% or higher I will call it a success. Assumptions: The data has complete records, there are no misspelling in the data, There is a combination of numerical, text, and catagorical values. The people classifying the records as duplicates are very accurate. 92% accurate or higher. Overall projectTitle what attempt this is to differintate models.We will use brnaching in git to help us develop. Toolset Jupyter notebooks will be for data exploration and visziulation Tensorboard will be used for accuracy analysis and cost. Notes for Table of Contents:Establish sectionsEnable hyperlinks Put this into the README.md file: Objectives: Name attributes and describe characteristics. % of Nulls Type of data. ie String, int, float Noise present. Such as outliers, logistic, rounding errors What is useful what isn't and why Type of distribution. Identifying Label data Visualization of data Identify correlations between variables Propose how the problem would be solved manually Provide transformations if nessiary Anything else of interest Exploritory data StrategyThe nature of the dataset is complex. This is due to the descriptive attirbutes assocatied with properites and cemetaries. There are only a couple of real numerical datatypes such as lat and long. I intend to go through each type figuring out if it is catagorical, a string/text, or numerical. Once the catagroies are identified applying a numerical number scheme for them will be adopted. Questions:Can I even descrptive statstics on catagorical data? From Comer's class I remember there beng some very strang things that happened.What do I even do with the catagorical data? Data cleaning Strats:I could take all null values and fill them with No Data. This would allow for complete records and give vectors to data that isn't there. In turn allowing for the whole record to be proccessed.I really think this should be prsued, because of the amount of null records there are.Drawbacks: I am creating data and altering data that is in the database. I would have to make general enough preproccessing to catch all of the null or missing values too. Then the model could train but if I miss something it could crash horribily and not really tell me why. Or I could get a very poor level of accuracy. Implementation: Look at all columns and identify incompletes Do an fillna() where possible Some catagorical data has 00 or none and that would need to change Run info() again to see what pandas says. Visiually inspect data Models Turn everything into text and concatenate all attributes into one string, apply TF-IDF and cosine similarity, then run model. Notes I do not know how and if I need to do describptive stats on the vectors of the strings Exploritiory driven modeling Convert individual columns into vectors PCA (principle compoent analysis vectors and drop relivant columns Train model Find complete records. After discovering complete fields do prepoccessing and run model just on those. This strategy will probably not work on the whole dataset becuase there are so many null. Implement One Hot Encoder instead of TD-IDF. Problems I forsee with doing this method are there are a lot of nominal values that are misspelled. They create a whole new catagory which will add more complexity to the model. It might not generalize well to new data. For example if we OneHotEncode Oklahoma and Olahoma (county) we create two different catagories. (Contrasting with the TD-IDF method): We are creating vectors based on occurance of important words within the corpus. Okay it seems it may have the same effect. This continues to prove that a massive data cleaning proccess needs to be done on the dataset or I need to cherry pick good records out of the dataset. Importing modules ###Code import pandas as pd import matplotlib as plt import numpy as np import seaborn as sns import sys %matplotlib inline plt.rcParams["figure.figsize"] = [10,10] pd.set_option('display.max_columns', 70) pd.set_option('display.max_rows', 200) ###Output _____no_output_____ ###Markdown Loading dataset in: ###Code df = pd.read_excel('datasets/prepared_data/Oklahoma_Working.xls') df.head() ###Output _____no_output_____ ###Markdown What is each Field? ###Code df.info() ###Output <class 'pandas.core.frame.DataFrame'> Int64Index: 3304 entries, 1 to 3948 Data columns (total 6 columns): PROPNAME 3304 non-null object RESNAME 3304 non-null object ADDRESS 3304 non-null object Lat 3304 non-null float64 Long 3304 non-null float64 duplicate_check 3304 non-null int64 dtypes: float64(2), int64(1), object(3) memory usage: 180.7+ KB ###Markdown Unique Fields Below is a custom made function designed to display unique values in the dataset all in one table. ###Code from uniques import uniques uniques(df) df.PROPNAME.unique() ###Output _____no_output_____ ###Markdown The code above does exactly what the table does. As you can see there are multiple errors in the spelling. With no standard of naming was used. ###Code # Sort alphabetical len(df.PROPNAME.unique().tolist()) ###Output _____no_output_____
notebooks/mesonic-mbs.ipynb
###Markdown Model-Based Sonification using mesonicThis notebooks shows how a Model-Based Sonification can be implemented using mesonic. ###Code import mesonic import sc3nb as scn import numpy as np from scipy.spatial import ConvexHull from scipy.spatial.distance import cdist, euclidean import matplotlib.pyplot as plt import seaborn as sns ###Output _____no_output_____ ###Markdown Preparation of Synths Lets start by preparing our Context and Synths. ###Code context = mesonic.create_context() context.enable_realtime() ###Output _____no_output_____ ###Markdown The model we use allows to interact with it using the mouse.For this we create a additional SynthDef as a click marker. ###Code scn.SynthDef("noise", r""" { |out=0, freq=2000, rq=0.02, amp=0.3, dur=1, pos=0 | Out.ar(out, Pan2.ar( BPF.ar(WhiteNoise.ar(10), freq, rq) * Line.kr(1, 0, dur, doneAction: 2).pow(4), pos, amp)); }""").add() test = context.synths.create("noise", mutable=False) test.start() ###Output _____no_output_____ ###Markdown Data Preparation We also prepare the data for the sonification.We will use the [Palmer penguins dataset](https://allisonhorst.github.io/palmerpenguins/) for the examples. ###Code df = sns.load_dataset("penguins") df = df.dropna(subset=["bill_length_mm", "bill_depth_mm", "flipper_length_mm", "body_mass_g", "sex"]) df = df.reset_index(drop=True) df %matplotlib inline sns.pairplot(data=df, hue="species") ###Output _____no_output_____ ###Markdown Data SonogramThe `DataSonogram` implements a Data Sonogram- The model gets a dataset which is plotted in two dimensions.- Imagine that for each point in the provided dataset a spring is created.- The data label of the point defines the stiffness of the spring.- When the user clicks into the plot a shock wave (signaled by noise Synth) is created from the nearest data point.- The shock wave excites the springs as it is spreading.- The resulting sonification can reveal clusters.More details about the Data Sonogram and Model-Based Sonification in general can be found in the [corresponding Sonification Handbook Chapter](https://sonification.de/handbook/chapters/chapter16/) ###Code class DataSonogram: def __init__(self, context, df, x, y, label, max_duration=1.5, spring_synth="s1", trigger_synth="noise"): self.context = context #prepare synths self.trigger_synth = context.synths.create(trigger_synth, mutable=False) self.spring_synth = context.synths.create(spring_synth, mutable=False) # save dataframe self.df = df self.numeric_df = df.select_dtypes(include=[np.number]) # check if x and y are valid allowed_columns = self.numeric_df.columns assert x in allowed_columns, f"x must be in {allowed_columns}" assert y in allowed_columns, f"y must be in {allowed_columns}" # prepare data for model self.labels = self.df[label] self.unique_labels = self.labels.unique() label2id = {label: idx for idx, label in enumerate(self.unique_labels)} self.numeric_labels = [label2id[label] for label in self.labels] self.xy_data = self.numeric_df[[x,y]].values self.data = self.numeric_df.values # get the convex hull of the data hull = ConvexHull(self.data) hull_data = self.data[hull.vertices,:] # get distances of the data points in the hull hull_distances = cdist(hull_data, hull_data, metric='euclidean') self.max_distance = hull_distances.max() # set model parameter self.max_duration = max_duration # prepare plot self.fig = plt.figure(figsize=(5,5)) self.ax = plt.subplot(111) # plot data sns.scatterplot(x=x, y=y, hue=label, data=df, ax=self.ax) # set callback def onclick(event): if event.inaxes is None: # outside plot area return if event.button != 1: # ignore other than left click return click_xy = np.array([event.xdata, event.ydata]) self.create_shockwave(click_xy) self.fig.canvas.mpl_connect('button_press_event', onclick) def create_shockwave(self, click_xy): self.context.reset() with self.context.now() as start_time: self.trigger_synth.start() # find the point that is the nearest to the click location center_idx = np.argmin(np.linalg.norm(self.xy_data - click_xy, axis=1)) center = self.data[center_idx] # get the distances from the other points to this point distances_to_center = np.linalg.norm(self.data - center, axis=1) # get idx sorted by distances order_of_points = np.argsort(distances_to_center) # for each point create a sound using the spring synth for idx in order_of_points: distance = distances_to_center[idx] nlabel = self.numeric_labels[idx] n = len(self.unique_labels)-1 onset = (distance / self.max_distance) * self.max_duration with self.context.at(start_time + onset): self.spring_synth.start( freq = 2 * (400 + 100 * nlabel), amp = scn.dbamp(scn.linlin(distance, 0, self.max_distance, -10, -30)), pan = [-1,1][int(self.xy_data[idx, 0]-click_xy[0] > 0)], dur = 0.04, info = {"label": self.labels[idx]}, ) ###Output _____no_output_____ ###Markdown To interact with the plot we use qt as matplotlib backend ###Code %matplotlib qt ###Output _____no_output_____ ###Markdown The `x` and `y` value can be setted to one of the numeric columns of the data set ###Code numeric_columns = ['bill_length_mm', 'bill_depth_mm', 'flipper_length_mm', 'body_mass_g'] ###Output _____no_output_____ ###Markdown Create two views of the model with different `x` and `y` value ###Code dsg1 = DataSonogram(context, df, x="flipper_length_mm", y="body_mass_g", label="species") dsg2 = DataSonogram(context, df, x="bill_length_mm", y="bill_depth_mm", label="species") ###Output _____no_output_____ ###Markdown We can enable the `fast_mode` of the `Clock` as the onsets will be very close. - The `fast_mode` is a workaround that will make the Playback worker skip `time.sleep`- `time.sleep` sleeps too long on Windows for many tasks- The upcomming Python 3.11 will fix this: https://docs.python.org/3.11/whatsnew/3.11.htmltime ###Code context.realtime_playback.clock.fast_mode # default is False context.realtime_playback.clock.fast_mode = True context.realtime_playback.clock.fast_mode = False ###Output _____no_output_____ ###Markdown Disable the `fast_mode` of the `Clock` if you want to do something different like creating a new Data Sonogram as the `fast_mode` keeps the [GIL](https://wiki.python.org/moin/GlobalInterpreterLock) busy. Filtering the MBSThe MBS can be filtered using the `processor.event_filter` - Lets create a helper that provides us with a filter function for the labels ###Code def create_label_filter(allowed): def label_filter(event): label = event.info.get("label", None) if label: return event if label in allowed else None return event return label_filter ###Output _____no_output_____ ###Markdown Select a filter and click again on the plot to listen to the filtered result. ###Code context.processor.event_filter = create_label_filter(["Chinstrap", "Gentoo"]) context.processor.event_filter = create_label_filter(["Adelie", "Gentoo"]) context.processor.event_filter = create_label_filter(["Adelie", "Chinstrap"]) context.processor.event_filter = create_label_filter(["Chinstrap"]) context.processor.event_filter = create_label_filter(["Adelie"]) context.processor.event_filter = create_label_filter(["Gentoo"]) ###Output _____no_output_____ ###Markdown - We can also remove the panning from the Events - This can help to identify the same point in the two different views ###Code def pan_filter(event): pan = event.data.get("pan", None) if pan: event.data["pan"] = 0 return event context.processor.event_filter = pan_filter ###Output _____no_output_____ ###Markdown Setting the filter to `None` will reset it ###Code context.processor.event_filter = None context.close() ###Output Quitting SCServer... Done. Exiting sclang... Done.
analysis/6.2_LightGBM (sklearn API).ipynb
###Markdown LightGBM Models ###Code import matplotlib.pyplot as plt import seaborn as sns import numpy as np import pandas as pd from pathlib import Path import lightgbm as lgb from sklearn.metrics import mean_squared_error from sklearn.model_selection import train_test_split from sklearn.metrics import f1_score DATA_DIR = Path('/content/drive/MyDrive/Work/Delivery/Current/Earthquake_damage/data') SUBMISSIONS_DIR = Path('drive/MyDrive/Work/Delivery/Current/Earthquake_damage/submissions') from google.colab import drive drive.mount('/content/drive') train_values = pd.read_csv(DATA_DIR / 'train_values.csv', index_col='building_id') train_labels = pd.read_csv(DATA_DIR / 'train_labels.csv', index_col='building_id') import lightgbm as lgb from sklearn.metrics import mean_squared_error from sklearn.model_selection import train_test_split from sklearn.metrics import f1_score sns.set() ###Output _____no_output_____ ###Markdown Getting Set Up ###Code !git clone --recursive https://github.com/Microsoft/LightGBM %cd /content/LightGBM !mkdir build !cmake -DUSE_GPU=1 !make -j$(nproc) !sudo apt-get -y install python-pip !sudo -H pip install setuptools pandas numpy scipy scikit-learn seaborn matplotlib -U %cd /content/LightGBM/python-package !sudo python setup.py install --precompile train_values y_train.shape train_values.info() print('Loading data...') # load or create your dataset train_values = pd.read_csv(DATA_DIR / 'train_values.csv', index_col='building_id') train_labels = pd.read_csv(DATA_DIR / 'train_labels.csv', index_col='building_id') X_train, X_test, y_train, y_test = train_test_split(train_values, train_labels, test_size=0.3, random_state=123, stratify=train_labels) # create dataset for lightgbm lgb_train = lgb.Dataset(X_train, y_train) lgb_eval = lgb.Dataset(X_test, y_test, reference=lgb_train) # specify your configurations as a dict params = { 'boosting_type': 'gbdt', 'objective': 'classification', 'metric': {'l2', 'l1'}, 'num_leaves': 31, 'learning_rate': 0.05, 'feature_fraction': 0.9, 'bagging_fraction': 0.8, 'bagging_freq': 5, 'verbose': 0 } print('Starting training...') # train gbm = lgb.train(params, lgb_train, num_boost_round=20, valid_sets=lgb_eval, early_stopping_rounds=5) print('Saving model...') # save model to file gbm.save_model('model.txt') print('Starting predicting...') # predict y_pred = gbm.predict(X_test, num_iteration=gbm.best_iteration) # eval print('The rmse of prediction is:', mean_squared_error(y_test, y_pred) ** 0.5) ###Output Loading data... ###Markdown Using Sklearn API and fucking off GPU ###Code import matplotlib.pyplot as plt import seaborn as sns import numpy as np import pandas as pd from pathlib import Path import lightgbm as lgb from sklearn.metrics import mean_squared_error from sklearn.model_selection import train_test_split from sklearn.metrics import f1_score DATA_DIR = Path('/content/drive/MyDrive/Work/Delivery/Current/Earthquake_damage/data') SUBMISSIONS_DIR = Path('drive/MyDrive/Work/Delivery/Current/Earthquake_damage/submissions') from google.colab import drive drive.mount('/content/drive') X = pd.read_csv(DATA_DIR / 'train_values.csv', index_col='building_id') y = pd.read_csv(DATA_DIR / 'train_labels.csv', index_col='building_id') sns.set() categorical_columns = X.select_dtypes(include='object').columns X[categorical_columns] = X[categorical_columns].astype('category') bool_columns = [col for col in X.columns if col.startswith('has')] X[bool_columns] = X[bool_columns].astype('bool') X = pd.get_dummies(X) from sklearn.pipeline import Pipeline from sklearn.preprocessing import StandardScaler from sklearn.model_selection import GridSearchCV from lightgbm import LGBMClassifier steps = [('scaler', StandardScaler()), ('lgbm', LGBMClassifier(random_state=42))] pipe = Pipeline(steps) pipe param_grid = {'lgbm__n_estimators': [50, 100, 150], 'lgbm__num_leaves': [20, 31, 40]} gs = GridSearchCV(pipe, param_grid, cv=5, verbose=3, n_jobs=-1) X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=123, stratify=y) pipe.fit(X_train, y_train.values.ravel()) from sklearn.metrics import f1_score y_pred = pipe.predict(X_test) f1_score(y_test, y_pred, average='micro') def make_submission(pipeline, title): """ Given a trained pipeline object, use it to make predictions on the submission test set 'test_values.csv' and write them a csv in the submissions folder. """ # Read in test_values csv and apply data preprocessing # note: will create a data preprocessing pipeline or function in future test_values = pd.read_csv(DATA_DIR / 'test_values.csv', index_col='building_id') test_values[categorical_columns] = test_values[categorical_columns].astype('category') test_values[bool_columns] = test_values[bool_columns].astype('bool') test_values = pd.get_dummies(test_values) # Generate predictions using pipeline we pass in predictions = pipeline.predict(test_values) submission_format = pd.read_csv(DATA_DIR / 'submission_format.csv', index_col='building_id') my_submission = pd.DataFrame(data=predictions, columns=submission_format.columns, index=submission_format.index) my_submission.to_csv(SUBMISSIONS_DIR / f'{title}.csv') make_submission(pipe, 'LGBMClassifier defaults all features') ###Output _____no_output_____ ###Markdown Exploring Built Model ###Code data = {'features': X.columns, 'importances': pipe.named_steps['lgbm'].feature_importances_} feature_df = pd.DataFrame(data) fig, ax = plt.subplots(figsize=plt.figaspect(1/4)) sns.barplot(x='features', y='importances', data=feature_df) plt.xticks(rotation=90) plt.show() feature_df.sort_values('importances', ascending=False).tail(20) # Quite a lot are less than 10 feature_df.importances.value_counts() # 14 features with importance above 100, let's train a model using just these feature_df[feature_df.importances >= 100] ###Output _____no_output_____ ###Markdown Model with Most Important 14 Features ###Code most_important_features = feature_df[feature_df.importances >= 100].features.to_numpy() cv = KFold(n_splits=5, random_state=1, shuffle=True) steps = [('scaler', StandardScaler()), ('lgbm', LGBMClassifier(random_state=42))] model = Pipeline(steps) scores = cross_val_score(model, X[most_important_features], y, scoring='f1_micro', cv=cv, n_jobs=-1) print('Accuracy: %.3f (%.3f)' % (np.mean(scores), np.std(scores))) param_grid = {'lgbm__n_estimators': [50, 100, 150], 'lgbm__num_leaves': [20, 31, 40]} gs = GridSearchCV(pipe, param_grid, cv=5, verbose=10, n_jobs=-1, scoring='f1_micro') gs.fit(X[most_important_features], y) # Seems like more estimators and more leaves is better gs.best_params_ gs.best_score_ y_pred = gs.predict(X[most_important_features]) f1_score(y, y_pred, average='micro') def make_submission_top_14_features(pipeline, title): """ Given a trained pipeline object, use it to make predictions on the submission test set 'test_values.csv' and write them a csv in the submissions folder. """ # Read in test_values csv and apply data preprocessing # note: will create a data preprocessing pipeline or function in future test_values = pd.read_csv(DATA_DIR / 'test_values.csv', index_col='building_id') test_values[categorical_columns] = test_values[categorical_columns].astype('category') test_values[bool_columns] = test_values[bool_columns].astype('bool') test_values = pd.get_dummies(test_values) test_values = test_values[most_important_features] # Generate predictions using pipeline we pass in predictions = pipeline.predict(test_values) submission_format = pd.read_csv(DATA_DIR / 'submission_format.csv', index_col='building_id') my_submission = pd.DataFrame(data=predictions, columns=submission_format.columns, index=submission_format.index) my_submission.to_csv(SUBMISSIONS_DIR / f'{title}.csv') title = 'Top 14 most informative features - minor hyperparameter tuning' make_submission_top_14_features(gs, title) ###Output _____no_output_____ ###Markdown Intense Hyperparameter Tuning ###Code LGBMClassifier() param_grid = {'lgbm__n_estimators': [150, 175, 200], 'lgbm__num_leaves': [40, 50, 60], #'lgbm__bosting_type': ['gbdt', 'dart', 'goss'], 'lgbm__learning_rate': [0.01, 0.1, 1], 'lgbm__min_split_gain': [0., 0.5], 'lgbm__min_child_weight': [1e-3, 1e-4, 1e-2], 'lgbm__min_child_samples': [10, 20, 30]} gs = GridSearchCV(pipe, param_grid, cv=2, verbose=10, n_jobs=-1, scoring='f1_micro') gs.fit(X[most_important_features], y) gs.best_params_ gs.best_estimator_ # This gave me a score of 0.7264 on the submission placing 518 gs.best_score_ y_pred = gs.predict(X[most_important_features]) f1_score(y, y_pred, average='micro') make_submission_top_14_features(gs, 'mid-level hyperparameter tuning') ###Output _____no_output_____ ###Markdown More intense hypereparameter tuning ###Code from sklearn.model_selection import RandomizedSearchCV param_dist = {'lgbm__n_estimators': np.arange(200, 410, 10), 'lgbm__num_leaves': np.arange(60, 130, 10), 'lgbm__bosting_type': ['gbdt', 'dart', 'goss'], 'lgbm__learning_rate': [0.1, 0.2, 0.3], 'lgbm__min_child_samples': np.arange(30, 100, 10)} rs = RandomizedSearchCV(pipe, param_dist, n_iter=300, cv=2, verbose=10, n_jobs=-1, scoring='f1_micro', random_state=42) rs.fit(X[most_important_features], y) rs.best_params_ rs.best_estimator_ # Scoresd 0.7397 on submission - placed 331 (in the top 10%!!!!) rs.best_score_ y_pred = rs.predict(X[most_important_features]) f1_score(y, y_pred, average='micro') make_submission_top_14_features(rs, 'random search hyperparameter tuning LightGBM') import pickle MODEL_DIR = Path('/content/drive/MyDrive/Work/Delivery/Current/Earthquake_damage/models') pkl_filename = MODEL_DIR / 'random search LightGBM.pkl' with open(pkl_filename, 'wb') as f: pickle.dump(rs, f) ###Output _____no_output_____ ###Markdown Even more intense RandomizedSearch ###Code from sklearn.model_selection import RandomizedSearchCV import pickle param_dist = {'lgbm__n_estimators': np.arange(200, 410, 10), 'lgbm__num_leaves': np.arange(60, 130, 10), 'lgbm__bosting_type': ['goss'], 'lgbm__learning_rate': [0.1, 0.2, 0.25, 0.3], 'lgbm__min_child_samples': np.arange(30, 100, 10)} rs = RandomizedSearchCV(pipe, param_dist, n_iter=500, cv=2, verbose=10, n_jobs=-1, scoring='f1_micro', random_state=42) most_important_features = ['geo_level_1_id', 'geo_level_2_id', 'geo_level_3_id', 'count_floors_pre_eq', 'age' , 'area_percentage' , 'height_percentage', 'has_superstructure_mud_mortar_stone', 'has_superstructure_stone_flag', 'has_superstructure_mud_mortar_brick', 'has_superstructure_cement_mortar_brick', 'has_superstructure_timber', 'count_families', 'other_floor_type_q'] rs.fit(X[most_important_features], y) print('Best params') print(rs.best_params_) print('Best score') print(rs.best_score_) print('F1 score on entire dataset') y_pred = rs.predict(X[most_important_features]) f1_score(y, y_pred, average='micro') print('Creating submission csv...') make_submission_top_14_features(rs, '0102 GOSS random search tuning LightGBM') print('Writing model to hard drive...') MODEL_DIR = Path('/content/drive/MyDrive/Work/Delivery/Current/Earthquake_damage/models') pkl_filename = MODEL_DIR / 'GOSS random search LightGBM.pkl' with open(pkl_filename, 'wb') as f: pickle.dump(rs, f) print('Finished') ###Output Fitting 2 folds for each of 500 candidates, totalling 1000 fits ###Markdown Exploring KFold Cross Val ###Code from sklearn.model_selection import KFold from sklearn.model_selection import cross_val_score cv = KFold(n_splits=10, random_state=1, shuffle=True) model = LGBMClassifier() scores = cross_val_score(model, X, y, scoring='f1_micro', cv=cv, n_jobs=-1) print('Accuracy: %.3f (%.3f)' % (np.mean(scores), np.std(scores))) fig, ax = plt.subplots() sns.boxplot(scores) plt.xlim([0.6, 0.8]) plt.show() ###Output /usr/local/lib/python3.6/dist-packages/seaborn/_decorators.py:43: FutureWarning: Pass the following variable as a keyword arg: x. From version 0.12, the only valid positional argument will be `data`, and passing other arguments without an explicit keyword will result in an error or misinterpretation. FutureWarning
tutorials/wowPython.ipynb
###Markdown 0. Preparation and setup One Python library that makes GraphX support available to our Jupyter notebooks is not yet bound to the runtime by default. To get it added to the Spark context you have to use the `!pip` magic cell command `install` first to bind the library to the existing runtime.The `pixiedust` library is implemented and loaded from [https://github.com/ibm-cds-labs/pixiedust](https://github.com/ibm-cds-labs/pixiedust). See the project documentation for details. ###Code !pip install --user --upgrade --no-deps pixiedust ###Output _____no_output_____ ###Markdown Pixiedust provides a nice visualization plugin for d3 style plots. Have a look at [https://d3js.org/](https://d3js.org/) if you are not yet familiar with d3. Having non-ascii characters in some of your tweets requires the Python interpreter to be set to support UTF-8. Reload your Python sys settings with UTF-8 encoding. ###Code import sys reload(sys) sys.setdefaultencoding('utf-8') ###Output _____no_output_____ ###Markdown When the library has been loaded successfully you have access to the PackageManager. Use the PackageManager to install a package to supply GraphFrames. Those are needed later in the notebook to complete the instructions for Spark GraphX. ###Code from pixiedust.packageManager import PackageManager pkg=PackageManager() pkg.installPackage("graphframes:graphframes:0") ###Output _____no_output_____ ###Markdown At this point you are being asked to _Please restart Kernel to complete installation of the new package_. Use the Restart Kernel dialog from the menu to do that. Once completed, you can start the analysis and resume with the next section. ****************************************************Please restart your Kernel before you proceed!**************************************************** 1. Load data from Twitter to CloudantFollowing the lab instructions you should at this point have:- a Cloudant account- an empty database in your Cloudant account- an IBM Insights for Twitter service instance Provide the details for both into the global variables section below, including*Twitter*:- _restAPI_ - the API endpoint we use to query the Twitter API with. Use the URL for your IBM Insights for Twitter service and add `/api/v1/messages/search` as path (for example `https://cdeservice.stage1.mybluemix.net/api/v1/messages/search`)- _username_ - the username for your IBM Insights for Twitter service instance- _password_ - the password for your IBM Insights for Twitter service instance*Cloudant*:- _account_ - the fully qualified account https URL (for example `https://testy-dynamite-001.cloudant.com`)- _username_ - the Cloudant account username- _password_ - the Cloudant account password- _database_ - the database name you want your tweets to be loaded into (Note: the database will NOT get created by the script below. Please create the database manually into your Cloudant account first.) ###Code properties = { 'twitter': { 'restAPI': 'https://xxx:[email protected]/api/v1/messages/search', 'username': 'xxx', 'password': 'xxx' }, 'cloudant': { 'account':'https://xxx:[email protected]', 'username':'xxx', 'password':'xxx', 'database':'election2016' } } ###Output _____no_output_____ ###Markdown Import all required Python libraries. ###Code import requests import json from requests.auth import HTTPBasicAuth import http.client ###Output _____no_output_____ ###Markdown Define a class with helper functions to query the Twitter service API and load documents in the Cloudant database using the bulk load API. (Note: no code is being executed yet and you don't expect any output for these declarations.) ###Code class TwitterToCloudant: count = 100 def query_twitter(self, config, url, query, loop): loop = loop + 1 if loop > (int(self.count) / 100): return # QUERY TWITTER if url is None: url = config["twitter"]["restAPI"] print(url, query) tweets = self.get_tweets(config, url, query) else: print(url) tweets = self.get_tweets(config, url, query) # LOAD TO CLOUDANT self.load_cloudant(config, tweets) # CONTINUE TO PAGE THROUGH RESULTS .... if "related" in tweets: url = tweets["related"]["next"]["href"] #!! recursive call self.query_twitter(config, url, None, loop) def get_tweets(self, config, url, query): # GET tweets from twitter endpoint user = config["twitter"]["username"] password = config["twitter"]["password"] print ("GET: Tweets from {} ".format(url)) if query is None: payload = {'country_code' :' us', 'lang' : 'en'} else: payload = {'q': query, 'country_code' :' us', 'lang' : 'en'} response = requests.get(url, params=payload, auth=HTTPBasicAuth(user, password)) print ("Got {} response ".format(response.status_code)) tweets = json.loads(response.text) return tweets def load_cloudant(self, config, tweets): # POST tweets to Cloudant database url = config["cloudant"]["account"] + "/" + config["cloudant"]["database"] + "/_bulk_docs" user = config["cloudant"]["username"] password = config["cloudant"]["password"] headers = {"Content-Type": "application/json"} if "tweets" in tweets: docs = {} docs["docs"] = tweets["tweets"] print ("POST: Docs to {}".format(url)) response = requests.post(url, data=json.dumps(docs), headers=headers, auth=HTTPBasicAuth(user, password)) print ("Got {} response ".format(response.status_code)) ###Output _____no_output_____ ###Markdown Finally we make the call the load our Cloudant database with tweets. To do that, we require two parameters:- _query_ - the query string to pass to the Twitter API. Use **election2016** as default or experiment with your own.- _count_ - the number of tweets to process. Use **200** as a good start or scale up if you want. (Note: Execution time depends on ....) ###Code query = "#election2016" count = 300 TtC = TwitterToCloudant() TtC.count = count TtC.query_twitter(properties, None, query, 0) ###Output _____no_output_____ ###Markdown At this point you should see a number of debug messages with response codes 200 and 201. As a result your database is loaded with the number of tweets you provided in _count_ variable above.If there are response codes like 401 (unauthorized) or 404 (not found) please check your credentails and URLs provided in the _properties_ above. Changes you make to these settings are applied when you execute the cell again. There is no need to execute other cells (that have not been changed) and you can immediately come back here to re-run your TwitterToCloudant functions.Should there be any severe problems that can not be resolved, we made a database called `tweets` already avaialable in your Cloudant account. You can continue to work through the following instructions using the `tweets` database instead. 2. Analyze tweets with Spark SQLIn this section your are going to explore the tweets loaded into your Cloudant database using Spark SQL queries. The Cloudant Spark connector library available at [https://github.com/cloudant-labs/spark-cloudant](https://github.com/cloudant-labs/spark-cloudant) is already linked with the Spark deployment underneath this notebook. All you have to do at this point is to read your Cloudant documents into a DataFrame. First, this notebook runs on a shared Spark cluster but obtains a dedicated Spark context for isolated binding. The Spark context (sc) is made available automatically when the notebook is launched and should be started at this point. With a few statements you can inspect the Spark version and resources allocated for this context._Note: If there is ever a problem with the running Spark context, you can submit sc.stop() and sc.start() to recycle it_ ###Code sc.version sc._conf.getAll() ###Output _____no_output_____ ###Markdown Now you want to create a Spark SQL context object off the given Spark context. ###Code sqlContext = SQLContext(sc) ###Output _____no_output_____ ###Markdown The Spark SQL context (sqlContext) is used to read data from the Cloudant database. We use a schema sample size and specified number of partitions to load the data with. For details on these parameters check [https://github.com/cloudant-labs/spark-cloudantconfiguration-on-sparkconf](https://github.com/cloudant-labs/spark-cloudantconfiguration-on-sparkconf) ###Code tweetsDF = sqlContext.read.format("com.cloudant.spark").\ option("cloudant.host",properties['cloudant']['account'].replace('https://','')).\ option("cloudant.username", properties['cloudant']['username']).\ option("cloudant.password", properties['cloudant']['password']).\ option("schemaSampleSize", "-1").\ option("jsonstore.rdd.partitions", "5").\ load(properties['cloudant']['database']) tweetsDF.show(5) ###Output _____no_output_____ ###Markdown For performance reasons we will cache the Data Frame to prevent re-loading. ###Code tweetsDF.cache() ###Output _____no_output_____ ###Markdown The schema of a Data Frame reveals the structure of all JSON documents loaded from your Cloudant database. Depending on the setting for the parameter `schemaSampleSize` the created RDD contains attributes for the first document only, for the first N documents, or for all documents. Please have a look at [https://github.com/cloudant-labs/spark-cloudantschema-variance](https://github.com/cloudant-labs/spark-cloudantschema-variance) for details on schema computation. ###Code tweetsDF.printSchema() ###Output _____no_output_____ ###Markdown With the use of the IBM Insights for Twitter API all tweets are enriched with metadata. For example, the gender of the Twitter user or the state of his account location are added in clear text. Sentiment analysis is also done at the time the tweets are loaded from the original Twitter API. This allows us to group tweets according to their positive, neutral, or negative sentiment.In a first example you can extract the gender, state, and polarity details from the DataFrame (or use any other field available in the schema output above). _Note: To extract a nested field you have to use the full attribute path, for example cde.author.gender or cde.content.sentiment.polarity. The alias() function is available to simplify the name in the resulting DataFrame._ ###Code tweetsDF2 = tweetsDF.select(tweetsDF.cde.author.gender.alias("gender"), tweetsDF.cde.author.location.state.alias("state"), tweetsDF.cde.content.sentiment.polarity.alias("polarity")) ###Output _____no_output_____ ###Markdown The above statement executes extremely fast because no actual function or transformation was computed yet. Spark uses a lazy approach to compute functions only when they are actually needed. The following function is used to show the output of the Data Frame. At that point only do you see a longer runtime to compute `tweetsDF2`. ###Code tweetsDF2.count() tweetsDF2.printSchema() ###Output _____no_output_____ ###Markdown Work with other Spark SQL functions to do things like counting, grouping etc. ###Code # count tweets by state tweets_state = tweetsDF2.groupBy(tweetsDF2.state).count() tweets_state.show(100) # count by gender & polarity tweets_gp0 = tweetsDF2.groupBy(tweetsDF2.gender, tweetsDF2.polarity).count() tweets_gp0.show(100) tweets_gp= tweetsDF2.where(tweetsDF2.polarity.isNotNull()).groupBy("polarity").pivot("gender").count() tweets_gp.show(100) ###Output _____no_output_____ ###Markdown 2.1 Plot results using matplotlibIn Python you can use simple libraries to plot your DataFrames directly in diagrams. However, the use of matplotlib is not trivial and once the data is rendered in the diagram it is static. For more comprehensive graphing Spark provides the GraphX extension. Here the data is transformed into a directed multigraph model (similar to those used in GraphDBs) called GraphFrames. You will explore GraphFrames later in this lab. Let's first have a look at simply plotting your DataFrames using matplotlib. ###Code import pandas as pd %matplotlib inline import matplotlib.pyplot as plt import numpy as np ###Output _____no_output_____ ###Markdown Plot the number of tweets per state. Notice again how Spark computes the result lazily. In no previous output did we require the full DataFrame and it did not have to get fully computed until now. ###Code tweets_state_pd = tweets_state.toPandas() values = tweets_state_pd['count'] labels = tweets_state_pd['state'] plt.gcf().set_size_inches(16, 12, forward=True) plt.title('Number of tweets by state') plt.barh(range(len(values)), values) plt.yticks(range(len(values)), labels) plt.show() ###Output _____no_output_____ ###Markdown More plots to group data by gender and polarity. ###Code tweets_gp_pd = tweets_gp.toPandas() labels = tweets_gp_pd['polarity'] N = len(labels) male = tweets_gp_pd['male'] female = tweets_gp_pd['female'] unknown = tweets_gp_pd['unknown'] ind = np.arange(N) # the x locations for the groups width = 0.2 # the width of the bars fig, ax = plt.subplots() rects1 = ax.bar(ind-width, male, width, color='b', label='male') rects2 = ax.bar(ind, female, width, color='r', label='female') rects3 = ax.bar(ind + width, unknown, width, color='y', label='unknown') ax.set_ylabel('Count') ax.set_title('Tweets by polarity and gender') ax.set_xticks(ind + width) ax.set_xticklabels(labels) ax.legend((rects1[0], rects2[0], rects3[0]), ('male', 'female', 'unknown')) plt.show() ###Output _____no_output_____ ###Markdown 2.2 Create SQL temporary tables With Spark SQL you can create in-memory tables and query your Spark RDDs in tables using SQL syntax. This is just an alternative represenation of your RDD where SQL functions (like filters or projections) are converted into Spark functions. For the user it mostly provides a SQL wrapper over Spark and a familiar way to query data. ###Code tweetsDF.registerTempTable("tweets_DF") ###Output _____no_output_____ ###Markdown Run SQL statements using the sqlContext.sql() function and render output with show(). The result of a SQL function could again be mapped to a data frame. ###Code sqlContext.sql("SELECT count(*) AS cnt FROM tweets_DF").show() sqlContext.sql("SELECT message.actor.displayName AS author, count(*) as cnt FROM tweets_DF GROUP BY message.actor.displayName ORDER BY cnt DESC").show(10) ###Output _____no_output_____ ###Markdown With multiple temporary tables (potentially from different databases) you can execute JOIN and UNION queries to analyze the database in combination. In the next query we will return all hashtags used in our body of tweets. ###Code hashtags = sqlContext.sql("SELECT message.object.twitter_entities.hashtags.text as tags \ FROM tweets_DF \ WHERE message.object.twitter_entities.hashtags.text IS NOT NULL") ###Output _____no_output_____ ###Markdown The hashtags are in lists, one per tweet. We flat map this list to a large list and then store it back into a temporary table. The temporary table can be used to define a hashtag cloud to understand which hashtag has been used how many times. ###Code l = hashtags.map(lambda x: x.tags).collect() tagCloud = [item for sublist in l for item in sublist] ###Output _____no_output_____ ###Markdown Create a DataFrame from the Python dictionary we used to flatten our hashtags into. The DataFrame has a simple schema with just a single column called `hastag`. ###Code from pyspark.sql import Row tagCloudDF = sc.parallelize(tagCloud) row = Row("hashtag") hashtagsDF = tagCloudDF.map(row).toDF() ###Output _____no_output_____ ###Markdown Register a new temp table for hashtags. Group and count tags to get a sense of trending issues. ###Code hashtagsDF.registerTempTable("hashtags_DF") trending = sqlContext.sql("SELECT count(hashtag) as CNT, hashtag as TAG FROM hashtags_DF GROUP BY hashtag ORDER BY CNT DESC") trending.show(10) ###Output _____no_output_____ ###Markdown 2.3 Visualize tag cloud with Brunel Let's create some charts and diagrams with Brunel commands.The basic format of each call to Brunel is simple. Whether the command is a single line or a set of lines, the commands are concatenated together and the result interpreted as one command.Here are some of the rules for using Brunel that you'll need in this notebook:- _DataFrame_: Use the data command to specify the pandas DataFrame.- _Chart type_: Use commands like chord and treemap to specify a chart type. If you don't specify a type, the default chart type is a scatterplot.- _Chart definition_: Use the x and y commands to specify the data to include on the x-axis and the y-axis.- _Styling_: Use commands like color, tooltip, and label to control the styling of the graph.- _Size_: Use the width and height key-value pairs to specify the size of the graph. The key-value pairs must be preceded with two colons and separated with a comma, for example: :: width=800, height=300See detailed documentation on the Brunel Visualization language at [https://brunel.mybluemix.net/docs](https://brunel.mybluemix.net/docs). ###Code import brunel import sys reload(sys) sys.setdefaultencoding('utf-8') trending_pd = trending.toPandas() ###Output _____no_output_____ ###Markdown Brunel libraries are able to read data from CSV files only. We will export our Panda DataFrames to CSV first to be able to load them with the Brunel libraries below. ###Code trending_pd.to_csv('trending_pd.csv') tweets_state_pd.to_csv('tweets_state_pd.csv') tweets_gp_pd.to_csv('tweets_gp_pd.csv') ###Output _____no_output_____ ###Markdown Top 5 records in every Panda DF. ###Code trending_pd.head(5) ###Output _____no_output_____ ###Markdown The hast tag cloud is visualized using the Brunel cloud graph. ###Code %brunel data('trending_pd') cloud color(cnt) size(cnt) label(tag) :: width=900, height=600 ###Output _____no_output_____ ###Markdown State and location data can be plotted on a map or a bubble graph representing the number of tweets per state. We will exercise maps later using the GraphX framework. ###Code tweets_state_pd.head(5) %brunel data('tweets_state_pd') bubble label(state) x(state) color(count) size(count) ###Output _____no_output_____ ###Markdown Brunel graphs are D3 based and interactive. Try using your mouse on the graph for Gender polarity to hover over details and zoom in on the Y axis. ###Code tweets_gp_pd.head(5) %brunel data('tweets_gp_pd') bar x(polarity) y(male, female) color(male, female) tooltip(#all) legends(none) :: width=800, height=300 ###Output _____no_output_____ ###Markdown 2.4 Write analysis results back to Cloudant Next we are going to persist the hashtags_DF back into a Cloudant database. (Note: The database `hashtags` has to exist in Cloudant. Please create that database first.) ###Code hashtagsDF.write.format("com.cloudant.spark").\ option("cloudant.host",properties['cloudant']['account'].replace('https://','')).\ option("cloudant.username", properties['cloudant']['username']).\ option("cloudant.password", properties['cloudant']['password']).\ option("bulkSize", "2000").\ save("hashtags") ###Output _____no_output_____ ###Markdown 3. Analysis with Spark GraphX Import dependencies from the Pixiedust library loaded in the preperation section. See [https://github.com/ibm-cds-labs/pixiedust](https://github.com/ibm-cds-labs/pixiedust) for details. ###Code from pixiedust.display import * ###Output _____no_output_____ ###Markdown To render a chart you have options to select the columns to display or the aggregation function to apply. ###Code tweets_state_us = tweets_state.filter(tweets_state.state.isin("Alabama", "Alaska", "Arizona", "Arkansas", "California", "Colorado", "Connecticut", "Delaware", "Florida", "Georgia", "Hawaii", "Idaho", "Illinois Indiana", "Iowa", "Kansas", "Kentucky", "Louisiana", "Maine", "Maryland", "Massachusetts", "Michigan", "Minnesota", "Mississippi", "Missouri", "Montana Nebraska", "Nevada", "New Hampshire", "New Jersey", "New Mexico", "New York", "North Carolina", "North Dakota", "Ohio", "Oklahoma", "Oregon", "Pennsylvania Rhode Island", "South Carolina", "South Dakota", "Tennessee", "Texas","Utah", "Vermont", "Virginia", "Washington", "West Virginia", "Wisconsin", "Wyoming")) tweets_state_us.show(5) display(tweets_state_us) ###Output _____no_output_____ ###Markdown Use a data set with at least two numeric columns to create scatter plots. 4. Analysis with Spark MLlibHere we are going to use KMeans clustering algorithm from Spark MLlib.Clustering will let us cluster similar tweets together.We will then display clusters using Brunel library. ###Code # TRAINING by hashtag from pyspark.mllib.feature import HashingTF from pyspark.mllib.clustering import KMeans, KMeansModel # dataframe of tweets' messages and hashtags mhDF = sqlContext.sql("SELECT message.body as message, \ message.object.twitter_entities.hashtags.text as tags \ FROM tweets_DF \ WHERE message.object.twitter_entities.hashtags.text IS NOT NULL") mhDF.show() # create an RDD of hashtags hashtagsRDD = mhDF.rdd.map(lambda h: h.tags) # create Feature verctor for every tweet's hastags # each hashtag represents feature # a function calculates how many time hashtag is in a tweet htf = HashingTF(100) vectors = hashtagsRDD.map(lambda hs: htf.transform(hs)).cache() print(vectors.take(2)) # Build the model (cluster the data) numClusters = 10 # number of clusters model = KMeans.train(vectors, numClusters, maxIterations=10, initializationMode="random") from pyspark.sql.functions import udf from pyspark.sql.types import IntegerType, StringType def predict(tags): vector = htf.transform(tags) return model.predict(vector) # Creates a Column expression representing a user defined function udfPredict = udf(predict, IntegerType()) def formatstr(message): lines = message.splitlines() return " ".join(lines) udfFormatstr = udf(formatstr, StringType()) # transform mhDF into cmhDF, a dataframe containing formatted messages, # hashtabs and cluster mhDF2 = mhDF.withColumn("message", udfFormatstr(mhDF.message)) cmhDF = mhDF2.withColumn("cluster", udfPredict(mhDF2.tags)) cmhDF.show() import sys reload(sys) sys.setdefaultencoding('utf-8') # visualizing clusters import brunel cmh_pd = cmhDF.toPandas() cmh_pd.to_csv('cmh_pd.csv') %brunel data('cmh_pd') bubble x(cluster) color(#all) size(#count) tooltip(message, tags) legends(none) ###Output _____no_output_____
Scraping_websites.ipynb
###Markdown http://www.gregreda.com/2016/10/16/asynchronous-scraping-with-python/ ###Code # import libraries from urllib.request import urlopen from bs4 import BeautifulSoup # specify the url quote_page = 'http://www.bloomberg.com/quote/SPX:IND' # query the website and return the html to the variable ‘page’ page = urlopen(quote_page) # parse the html using beautiful soup and store in variable `soup` soup = BeautifulSoup(page, 'html.parser') # Take out the <div> of name and get its value name_box = soup.find('h1', attrs={'class': 'name'}) name = name_box.text.strip() # strip() is used to remove starting and trailing print(name) # get the index price price_box = soup.find('div', attrs={'class':'price'}) price = price_box.text print(price) import csv from datetime import datetime # open a csv file with append, so old data will not be erased with open('index.csv', 'a') as csv_file: writer = csv.writer(csv_file) writer.writerow([name, price, datetime.now()]) ###Output _____no_output_____ ###Markdown multiple pages example ###Code from concurrent.futures import ProcessPoolExecutor import concurrent.futures from concurrent import futures URLS = ['http://www.bloomberg.com/quote/SPX:IND', 'http://www.bloomberg.com/quote/CCMP:IND'] # list of urls data =[] def parse(url): page = urlopen(url) # parse the html using beautiful soap and store in variable `soup` soup = BeautifulSoup(page, 'html.parser') # Take out the <div> of name and get its value name_box = soup.find('h1', attrs={'class': 'name'}) name = name_box.text.strip() # strip() is used to remove starting and trailing # get the index price price_box = soup.find('div', attrs={'class':'price'}) price = price_box.text data.extend([name, price]) return data for url in URLS: parse(url) print(data) def main(): with ProcessPoolExecutor(max_workers=4) as executor: future_results = {executor.submit(parse, url): url for url in URLS} print(name, price) # for future in concurrent.futures.as_completed(future_results): # print(future.result()) # if name == 'main': # for future in concurrent.futures.as_completed(future_results): # results.append(future.result()) if __name__ == '__main__': main() print(results) print(results) ###Output []
notebooks/collect_data_from_db.ipynb
###Markdown Pobranie danych z bazyW tym skrypcie pobrałem wszystkie potrzebne tabele z lokalnego zrzutu bazy danych "MSR 2014 Mining Challenge Dataset" do plików `.pkl`, aby później móc na tych danych operować już bez połączenia z bazą, a jedynie przy pomocy lokalnych plików. Wszystkie przygotowane w tym miejscu pliki można znaleźć spakowane w archiwum pod adresem https://www-users.mat.umk.pl/~maciejdudek/, aby można było pominąć etap podłączania bazy danych.Szczegółowy opis znajduje się z rozdziale 4.2.2 pracy. ###Code import pandas as pd from sqlalchemy import create_engine data_from_db = '../data/from_db/' db_connection = create_engine( 'mysql+pymysql://msr14:haslo@localhost:3306/msr14') def from_db_to_pickle(db_con, entity, path): """ Zapisuje tabelę z bazy danych bezpośrednio do pliku pickle. Parameters ---------- db_con : sqlalchemy.engine.Engine obiekt silnika pozwalający na nawiązanie połączenia z bazą danych entity : str nazwa tabeli z bazy danych do pobrania path : str ścieżka do katalogu, w którym zapisać plik """ pd.read_sql(entity, con=db_con).to_pickle(path + entity + '.pkl') entities = ['projects', 'commits', 'commit_comments', 'issues', 'issue_comments', 'pull_requests', 'pull_request_history', 'pull_request_comments', 'watchers'] for entity in entities: from_db_to_pickle(db_connection, entity, data_from_db) ###Output _____no_output_____
code/chap08-mky.ipynb
###Markdown Modeling and Simulation in PythonChapter 8Copyright 2017 Allen DowneyLicense: [Creative Commons Attribution 4.0 International](https://creativecommons.org/licenses/by/4.0) ###Code # Configure Jupyter so figures appear in the notebook %matplotlib inline # Configure Jupyter to display the assigned value after an assignment %config InteractiveShell.ast_node_interactivity='last_expr_or_assign' # import functions from the modsim.py module from modsim import * from pandas import read_html ###Output _____no_output_____ ###Markdown Functions from the previous chapter ###Code def plot_results(census, un, timeseries, title): """Plot the estimates and the model. census: TimeSeries of population estimates un: TimeSeries of population estimates timeseries: TimeSeries of simulation results title: string """ plot(census, ':', label='US Census') plot(un, '--', label='UN DESA') plot(timeseries, color='gray', label='model') decorate(xlabel='Year', ylabel='World population (billion)', title=title) def run_simulation(system, update_func): """Simulate the system using any update function. system: System object update_func: function that computes the population next year returns: TimeSeries """ results = TimeSeries() results[system.t_0] = system.p_0 for t in linrange(system.t_0, system.t_end): results[t+1] = update_func(results[t], t, system) return results ###Output _____no_output_____ ###Markdown Reading the data ###Code filename = 'data/World_population_estimates.html' tables = read_html(filename, header=0, index_col=0, decimal='M') table2 = tables[2] table2.columns = ['census', 'prb', 'un', 'maddison', 'hyde', 'tanton', 'biraben', 'mj', 'thomlinson', 'durand', 'clark'] un = table2.un / 1e9 census = table2.census / 1e9 plot(census, ':', label='US Census') plot(un, '--', label='UN DESA') decorate(xlabel='Year', ylabel='World population (billion)', title='Estimated world population') ###Output _____no_output_____ ###Markdown Running the quadratic model Here's the update function for the quadratic growth model with parameters `alpha` and `beta`. ###Code def update_func_quad(pop, t, system): """Update population based on a quadratic model. pop: current population in billions t: what year it is system: system object with model parameters """ net_growth = system.alpha * pop + system.beta * pop**2 return pop + net_growth ###Output _____no_output_____ ###Markdown Extract the starting time and population. ###Code t_0 = get_first_label(census) t_end = get_last_label(census) p_0 = get_first_value(census) ###Output _____no_output_____ ###Markdown Initialize the system object. ###Code system = System(t_0=t_0, t_end=t_end, p_0=p_0, alpha=0.025, beta=-0.0018) ###Output _____no_output_____ ###Markdown Run the model and plot results. ###Code results= run_simulation(system, update_func_quad) plot_results(census, un, results, 'quadratic') ###Output _____no_output_____ ###Markdown Generating projections To generate projections, all we have to do is change `t_end` ###Code system.t_end = 2250 results = run_simulation(system, update_func_quad) plot_results(census, un, results, 'World population projection') savefig('figs/chap04-fig01.pdf') ###Output Saving figure to file figs/chap04-fig01.pdf ###Markdown The population in the model converges on the equilibrium population, `-alpha/beta` ###Code results[system.t_end] -system.alpha / system.beta ###Output _____no_output_____ ###Markdown **Exercise:** What happens if we start with an initial population above the carrying capacity, like 20 billion? Run the model with initial populations between 1 and 20 billion, and plot the results on the same axes. ###Code p0_array = linspace(1,20,20) for system.p_0 in p0_array: results = run_simulation(system, update_func_quad) plot(results) ###Output _____no_output_____ ###Markdown Comparing projections We can compare the projection from our model with projections produced by people who know what they are doing. ###Code table3 = tables[3] table3.head() ###Output _____no_output_____ ###Markdown `NaN` is a special value that represents missing data, in this case because some agencies did not publish projections for some years. ###Code table3.columns = ['census', 'prb', 'un'] ###Output _____no_output_____ ###Markdown This function plots projections from the UN DESA and U.S. Census. It uses `dropna` to remove the `NaN` values from each series before plotting it. ###Code def plot_projections(table): """Plot world population projections. table: DataFrame with columns 'un' and 'census' """ census_proj = table.census / 1e9 un_proj = table.un / 1e9 plot(census_proj.dropna(), 'b:', label='US Census') plot(un_proj.dropna(), 'g--', label='UN DESA') ###Output _____no_output_____ ###Markdown Run the model until 2100, which is as far as the other projections go. ###Code system = System(t_0=t_0, t_end=2100, p_0=p_0, alpha=0.025, beta=-0.0018) results = run_simulation(system, update_func_quad) plot_results(census, un, results, 'World population projections') plot_projections(table3) savefig('figs/chap04-fig02.pdf') ###Output Saving figure to file figs/chap04-fig02.pdf ###Markdown People who know what they are doing expect the growth rate to decline more sharply than our model projects. Exercises**Optional exercise:** The net growth rate of world population has been declining for several decades. That observation suggests one more way to generate projections, by extrapolating observed changes in growth rate.The `modsim` library provides a function, `compute_rel_diff`, that computes relative differences of the elements in a sequence. It is a wrapper for the NumPy function `ediff1d`: ###Code %psource compute_rel_diff ###Output _____no_output_____ ###Markdown Here's how we can use it to compute the relative differences in the `census` and `un` estimates: ###Code alpha_census = compute_rel_diff(census) plot(alpha_census) alpha_un = compute_rel_diff(un) plot(alpha_un) decorate(xlabel='Year', label='Net growth rate') ###Output _____no_output_____ ###Markdown Other than a bump around 1990, net growth rate has been declining roughly linearly since 1965. As an exercise, you can use this data to make a projection of world population until 2100.1. Define a function, `alpha_func`, that takes `t` as a parameter and returns an estimate of the net growth rate at time `t`, based on a linear function `alpha = intercept + slope * t`. Choose values of `slope` and `intercept` to fit the observed net growth rates since 1965.2. Call your function with a range of `ts` from 1960 to 2020 and plot the results.3. Create a `System` object that includes `alpha_func` as a system variable.4. Define an update function that uses `alpha_func` to compute the net growth rate at the given time `t`.5. Test your update function with `t_0 = 1960` and `p_0 = census[t_0]`.6. Run a simulation from 1960 to 2100 with your update function, and plot the results.7. Compare your projections with those from the US Census and UN. ###Code def update_alpha_func(population, t, system): population= 2 + 0.4 * t p_0=2 return alpha system = System(t_0=t_0, t_end=2020, p_0=p_0 intercept = 2) results= run_simulation(system, update_alpha_func) plot_results(census, un, results, 'quadratic') # Solution goes here # Solution goes here # Solution goes here # Solution goes here # Solution goes here ###Output _____no_output_____
Classify_events/classify_events.ipynb
###Markdown Download events related to each convocation ###Code df_events = None for file in data_files: sep_symbol = ';' if '9' in file else ',' df = pd.read_csv(os.path.join(data_folder, file), sep=sep_symbol) print(file, df.shape) if df_events is None: df_events = df else: df_events = pd.concat([df_events, df]) print(f'\nCombined dataset: {df_events.shape}') # data sample df_events.head().T ###Output identifiers_9skl.csv (3857, 18) identifiers_7skl.csv (4135, 18) identifiers_8skl.csv (24103, 18) identifiers_3skl.csv (7943, 18) identifiers_4skl.csv (10929, 18) identifiers_6skl.csv (23282, 18) identifiers_5skl.csv (3424, 18) Combined dataset: (77673, 18) ###Markdown Data exploration what is the unique ID of an event? ###Code print(f'Number of events: {len(df_events):,.0f}') print(f'Number of unique event IDs: {len(df_events.id_event.unique()):,.0f}') print(f'Number of unique IDs: {len(df_events.id.unique()):,.0f}') print(f'Number of unique question IDs: {len(df_events.id_question.unique()):,.0f}') print(f'Number of unique question number|event: {len(df_events["number|event"].unique()):,.0f}') print('\n"number|event" column is a unique event identifier') print('\nExample of events events with identical "id_event" values:') display(df_events[df_events.id_event.isin([5184, 1305, 29505])].sort_values('id_event').T) ###Output Number of events: 77,673 Number of unique event IDs: 73,232 Number of unique IDs: 15,804 Number of unique question IDs: 21,457 Number of unique question number|event: 77,673 "number|event" column is a unique event identifier Example of events events with identical "id_event" values: ###Markdown Main categories of documents in the dataset ###Code print('Column "rubric"') df_events.loc[:, 'rubric'] = np.where(df_events.rubric.isna(), 'Undefined', df_events.rubric) display(df_events.groupby('rubric')['number|event'].count().sort_values(ascending=False)) print('\nColumn "type"') df_events.loc[:, 'type'] = np.where(df_events.type.isna(), 'Undefined', df_events.type) display(df_events.groupby('type')['number|event'].count().sort_values(ascending=False)) print('\nColumns "type" and "registrationConvocation"') df_events.loc[:, 'registrationConvocation'] = np.where(df_events.registrationConvocation.isna(), 'Undefined', df_events.registrationConvocation) df_type_convocation = pd.pivot_table(data = df_events, index = "type", columns = "registrationConvocation", values = "number|event", aggfunc = "count", fill_value = 0) df_type_convocation = df_type_convocation[['Undefined', 'III скликання', 'IV скликання', 'V скликання', 'VI скликання', 'VII скликання', 'VIII скликання', 'IX скликання']] display(df_type_convocation) ###Output Column "rubric" ###Markdown Classify events ###Code events_number = df_events.shape[0] df_remainer, df_combined = separate_meta_types(df_events, types_dict) events_classified = df_combined.shape[0] print(f'Number of events in the database: {events_number:,.0f}') print(f'Number of classified events: {events_classified:,.0f} ({events_classified/events_number*100:,.2f})') # combine classified and not classified df_remainer.loc[:, 'type_name'] = 'not classified' df_remainer.loc[:, 'meta_type_name'] = 'not classified' df_combined = pd.concat([df_combined, df_remainer]) df_meta_convocation = get_type_pivot(df_combined, "registrationConvocation") df_meta_type = get_type_pivot(df_combined, "type") df_classification_count = df_combined.groupby(['meta_type_name', 'type_name'])[['number|event']].count() display(df_meta_convocation) print('\n\n') display(df_meta_type) print('\n\n') display(df_classification_count) df_meta_convocation.to_csv('summary_convocation.csv', index=False, sep="\t") df_meta_type.to_csv('summary_document_type.csv', index=False, sep="\t") df_classification_count.to_csv('summary_classification.csv', index=True, sep="\t") df_combined.to_csv(os.path.join(data_folder, 'identifiers_combined.csv'), index=False, sep="\t") ###Output _____no_output_____
code/HIV MODEL.ipynb
###Markdown HIV MODELChris Lee ###Code # Configure Jupyter so figures appear in the notebook %matplotlib inline # Configure Jupyter to display the assigned value after an assignment %config InteractiveShell.ast_node_interactivity='last_expr_or_assign' # import functions from the modsim.py module from modsim import * def make_system(alpha, beta, delta, gamma, mu, pi, rho, sigma, tau): """Make a system object for the HIV model. alpha: rate at which lymphocytes switch from latent to active beta: rate of infection of lymphocytes per virion delta: rate of removal of infected cells gamma: rate at which new, uninfected CD4 lymphocytes arise mu: death rate of lymphocytes pi: free virion production rate sigma: rate of removal of free virions tau: proportion of lymphocytes that are activated returns: System object """ init = State(R=1000, L=0, E=0, V=100) #R = lymphocyts #L = Latent Infected #E = Actively Infected #V = Free Virions t0 = 0 t_end = 120 return System(init=init, t0=t0, t_end=t_end, alpha=alpha, beta=beta, delta=delta, gamma=gamma, mu=mu, pi=pi, rho = rho, sigma=sigma, tau=tau) sys = make_system(3.6*10e-2, .00027, .33, 1.36, 1.36e-3, 100, .1, 2, .2) def update_func(state, t, system): """Update the SIR model. state: State (R, L, E, V) t: time system: System object returns: State (rlev) """ r, l, e, v = state unpack(system) dt = 1/(t_end/120) BRV = beta*v*r delta_lymphocytes = ((gamma*tau) - (mu*r) - BRV) * dt delta_latent = ((rho*BRV) - (mu*l) - (alpha*l)) * dt delta_active = (((1-rho)*BRV) + (alpha*l) - (delta*e)) * dt delta_virions = ((pi*e) - (sigma*v)) * dt r += delta_lymphocytes l += delta_latent e += delta_active v += delta_virions return State(R=r, L=l, E=e, V=v) def run_simulation(system, update_func): """Runs a simulation of the system. system: System object update_func: function that updates state returns: TimeFrame """ unpack(system) frame = TimeFrame(columns=init.index) frame.row[t0] = init for t in linrange(t0, t_end): frame.row[t+1] = update_func(frame.row[t], t, system) return frame def plot_results(r, l, e, v): """Plot the results of a SIR model. r, l, e, v: timeseries """ plot(r, 'r-', label='Lymphocytes') plot(l, 'b-', label='Latent Infected') plot(e, 'g-', label='Activated Infected') plot(v, 'y-', label='Free Virions') decorate(xlabel='Time (days)', ylabel='Fraction of population') results = run_simulation(sys, update_func) plot_results(results.R, results.L, results.E, results.V) def slope_func(state, t, system): """ Outputs slope of each type per time step system: System object t: time step state: State (R, L, E, V) """ unpack(system) r, l, e, v = state drdt = gamma*tau - mu*r - beta*v*r dldt = rho*beta*r*v - mu*gamma - alpha*l dedt = (1-rho)*beta*r*v + alpha*l - delta*e dvdt = pi*e - sigma*v return drdt, dldt, dedt, dvdt results, details = run_ode_solver(sys, slope_func) details plot_results(results.R, results.L, results.E, results.V) ###Output _____no_output_____
.ipynb_checkpoints/Setting Up-checkpoint.ipynb
###Markdown U1L1 - Setting Up---Easiest way to start coding in Python is to use a web service from [repl.it](https://repl.it)If you'd like to use native software or other set ups please follow the guides below:- For Windows: [Guide Link from Visual Studio](https://code.visualstudio.com/docs/python/python-tutorial)- For macOS: [Guide Link from Digital Ocean](https://www.digitalocean.com/community/tutorials/how-to-install-python-3-and-set-up-a-local-programming-environment-on-macos)**It will be assumed that you will be coding on repl.it. You may require additional setting configurations if you are using a different setup.** Starting with repl.it![](figures/01_f1.png)**To Do List:**1. Create an account. No need to select the _I'm a teacher_ option.2. Click on the blue button with a plus sign to create a new python file.3. Call your python file: "My First Python Program" My First Python Program![](figures/01_f2.png)Examine the figure above, there are three vertical columns of importance:1. File Directory2. Code Editor3. Python Interpreter 1. File DirectoryAs your python projects get more advanced, you will be creating more files here. The basics course will mainly deal with just editing our _"main.py"_ file. 2. Code EditorThis is where you write your python code.- To left we have numbers letting us know which line we are at- Python is a Top-Down Intrepreted Language; therefore, Python will execute the lines of code starting at Line 0 all the way until it reaches the end of the file- Currently we've wrote some lines of code that are executable. 3. Python InterpreterThis is where we get to see our code in action.- It will start from Line 0 from our editor- It will start to read line by line- If there is an _output_ to be made on the interpreter, it will be outputted on the interpreter._At the moment, I have some code written down. Please feel free to copy it._To see all of this in action, please press **Run**. ###Code # My First Python Program # By: Mr. Park print('Hello, World!') ###Output Hello, World!
module3/DS7_assignment_kaggle_challenge_3.ipynb
###Markdown Lambda School Data Science, Unit 2: Predictive Modeling Kaggle Challenge, Module 3 Assignment- [X] [Review requirements for your portfolio project](https://lambdaschool.github.io/ds/unit2), then submit your dataset.- [X] Continue to participate in our Kaggle challenge. - [X] Use scikit-learn for hyperparameter optimization with RandomizedSearchCV.- [X] Submit your predictions to our Kaggle competition. (Go to our Kaggle InClass competition webpage. Use the blue **Submit Predictions** button to upload your CSV file. Or you can use the Kaggle API to submit your predictions.)- [X] Commit your notebook to your fork of the GitHub repo. Stretch Goals Reading- Jake VanderPlas, [Python Data Science Handbook, Chapter 5.3](https://jakevdp.github.io/PythonDataScienceHandbook/05.03-hyperparameters-and-model-validation.html), Hyperparameters and Model Validation- Jake VanderPlas, [Statistics for Hackers](https://speakerdeck.com/jakevdp/statistics-for-hackers?slide=107)- Ron Zacharski, [A Programmer's Guide to Data Mining, Chapter 5](http://guidetodatamining.com/chapter5/), 10-fold cross validation- Sebastian Raschka, [A Basic Pipeline and Grid Search Setup](https://github.com/rasbt/python-machine-learning-book/blob/master/code/bonus/svm_iris_pipeline_and_gridsearch.ipynb)- Peter Worcester, [A Comparison of Grid Search and Randomized Search Using Scikit Learn](https://blog.usejournal.com/a-comparison-of-grid-search-and-randomized-search-using-scikit-learn-29823179bc85) Doing- In additon to `RandomizedSearchCV`, scikit-learn has [`GridSearchCV`](https://scikit-learn.org/stable/modules/generated/sklearn.model_selection.GridSearchCV.html). Another library called scikit-optimize has [`BayesSearchCV`](https://scikit-optimize.github.io/notebooks/sklearn-gridsearchcv-replacement.html). Experiment with these alternatives.- _[Introduction to Machine Learning with Python](http://shop.oreilly.com/product/0636920030515.do)_ discusses options for "Grid-Searching Which Model To Use" in Chapter 6:> You can even go further in combining GridSearchCV and Pipeline: it is also possible to search over the actual steps being performed in the pipeline (say whether to use StandardScaler or MinMaxScaler). This leads to an even bigger search space and should be considered carefully. Trying all possible solutions is usually not a viable machine learning strategy. However, here is an example comparing a RandomForestClassifier and an SVC ...The example is shown in [the accompanying notebook](https://github.com/amueller/introduction_to_ml_with_python/blob/master/06-algorithm-chains-and-pipelines.ipynb), code cells 35-37. Could you apply this concept to your own pipelines? More Categorical Encodings**1.** The article **[Categorical Features and Encoding in Decision Trees](https://medium.com/data-design/visiting-categorical-features-and-encoding-in-decision-trees-53400fa65931)** mentions 4 encodings:- **"Categorical Encoding":** This means using the raw categorical values as-is, not encoded. Scikit-learn doesn't support this, but some tree algorithm implementations do. For example, [Catboost](https://catboost.ai/), or R's [rpart](https://cran.r-project.org/web/packages/rpart/index.html) package.- **Numeric Encoding:** Synonymous with Label Encoding, or "Ordinal" Encoding with random order. We can use [category_encoders.OrdinalEncoder](https://contrib.scikit-learn.org/categorical-encoding/ordinal.html).- **One-Hot Encoding:** We can use [category_encoders.OneHotEncoder](http://contrib.scikit-learn.org/categorical-encoding/onehot.html).- **Binary Encoding:** We can use [category_encoders.BinaryEncoder](http://contrib.scikit-learn.org/categorical-encoding/binary.html).**2.** The short video **[Coursera — How to Win a Data Science Competition: Learn from Top Kagglers — Concept of mean encoding](https://www.coursera.org/lecture/competitive-data-science/concept-of-mean-encoding-b5Gxv)** introduces an interesting idea: use both X _and_ y to encode categoricals.Category Encoders has multiple implementations of this general concept:- [CatBoost Encoder](http://contrib.scikit-learn.org/categorical-encoding/catboost.html)- [James-Stein Encoder](http://contrib.scikit-learn.org/categorical-encoding/jamesstein.html)- [Leave One Out](http://contrib.scikit-learn.org/categorical-encoding/leaveoneout.html)- [M-estimate](http://contrib.scikit-learn.org/categorical-encoding/mestimate.html)- [Target Encoder](http://contrib.scikit-learn.org/categorical-encoding/targetencoder.html)- [Weight of Evidence](http://contrib.scikit-learn.org/categorical-encoding/woe.html)Category Encoder's mean encoding implementations work for regression problems or binary classification problems. For multi-class classification problems, you will need to temporarily reformulate it as binary classification. For example:```pythonencoder = ce.TargetEncoder(min_samples_leaf=..., smoothing=...) Both parameters > 1 to avoid overfittingX_train_encoded = encoder.fit_transform(X_train, y_train=='functional')X_val_encoded = encoder.transform(X_train, y_val=='functional')```**3.** The **[dirty_cat](https://dirty-cat.github.io/stable/)** library has a Target Encoder implementation that works with multi-class classification.```python dirty_cat.TargetEncoder(clf_type='multiclass-clf')```It also implements an interesting idea called ["Similarity Encoder" for dirty categories](https://www.slideshare.net/GaelVaroquaux/machine-learning-on-non-curated-data-154905090).However, it seems like dirty_cat doesn't handle missing values or unknown categories as well as category_encoders does. And you may need to use it with one column at a time, instead of with your whole dataframe.**4. [Embeddings](https://www.kaggle.com/learn/embeddings)** can work well with sparse / high cardinality categoricals._**I hope it’s not too frustrating or confusing that there’s not one “canonical” way to encode categorcals. It’s an active area of research and experimentation! Maybe you can make your own contributions!**_ BONUS: Stacking!Here's some code you can use to "stack" multiple submissions, which is another form of ensembling:```pythonimport pandas as pd Filenames of your submissions you want to ensemblefiles = ['submission-01.csv', 'submission-02.csv', 'submission-03.csv']target = 'status_group'submissions = (pd.read_csv(file)[[target]] for file in files)ensemble = pd.concat(submissions, axis='columns')majority_vote = ensemble.mode(axis='columns')[0]sample_submission = pd.read_csv('sample_submission.csv')submission = sample_submission.copy()submission[target] = majority_votesubmission.to_csv('my-ultimate-ensemble-submission.csv', index=False)``` ###Code import os, sys in_colab = 'google.colab' in sys.modules # If you're in Colab... if in_colab: # Pull files from Github repo os.chdir('/content') !git init . !git remote add origin https://github.com/LambdaSchool/DS-Unit-2-Kaggle-Challenge.git !git pull origin master # Install required python packages !pip install -r requirements.txt # Change into directory for module os.chdir('module3') import pandas as pd # Merge train_features.csv & train_labels.csv train = pd.merge(pd.read_csv('../data/tanzania/train_features.csv'), pd.read_csv('../data/tanzania/train_labels.csv')) # Read test_features.csv & sample_submission.csv test = pd.read_csv('../data/tanzania/test_features.csv') sample_submission = pd.read_csv('../data/tanzania/sample_submission.csv') import pandas as pd import numpy as np from scipy.stats import randint, uniform import random as ran import matplotlib.pyplot as plt import seaborn as sns from sklearn.base import BaseEstimator, TransformerMixin from sklearn.impute import SimpleImputer from sklearn.cluster import KMeans from sklearn.model_selection import train_test_split from sklearn.neighbors import KNeighborsClassifier from sklearn.linear_model import LogisticRegressionCV, LogisticRegression from sklearn.preprocessing import StandardScaler from sklearn.decomposition import PCA from sklearn.ensemble import GradientBoostingClassifier, RandomForestClassifier from sklearn.feature_selection import f_classif, chi2, SelectKBest, SelectPercentile, SelectFpr, SelectFromModel from sklearn.pipeline import make_pipeline, Pipeline, FeatureUnion from sklearn.model_selection import GridSearchCV, RandomizedSearchCV import category_encoders as ce import warnings warnings.filterwarnings(action='ignore', category=FutureWarning, module='sklearn') import numpy as np import pandas as pd from sklearn.model_selection import train_test_split # Merge train_features.csv & train_labels.csv train = pd.merge(pd.read_csv('train_features.csv'), pd.read_csv('train_labels.csv')) # Read test_features.csv & sample_submission.csv test = pd.read_csv('test_features.csv') sample_submission = pd.read_csv('sample_submission.csv') # Split train into train & val train, val = train_test_split(train, train_size=0.80, test_size=0.20, stratify=train['status_group'], random_state=42) def wrangle(X): """Wrangle train, validate, and test sets in the same way""" # Prevent SettingWithCopyWarning X = X.copy() # About 3% of the time, latitude has small values near zero, # outside Tanzania, so we'll treat these values like zero. X['latitude'] = X['latitude'].replace(-2e-08, 0) # When columns have zeros and shouldn't, they are like null values. # So we will replace the zeros with nulls, and impute missing values later. # Also create a "missing indicator" column, because the fact that # values are missing may be a predictive signal. cols_with_zeros = ['longitude', 'latitude', 'construction_year', 'gps_height', 'population'] for col in cols_with_zeros: X[col] = X[col].replace(0, np.nan) X[col+'_MISSING'] = X[col].isnull() # Drop duplicate columns duplicates = ['quantity_group', 'payment_type'] X = X.drop(columns=duplicates) # Drop recorded_by (never varies) and id (always varies, random) unusable_variance = ['recorded_by', 'id'] X = X.drop(columns=unusable_variance) # Convert date_recorded to datetime X['date_recorded'] = pd.to_datetime(X['date_recorded'], infer_datetime_format=True) # Extract components from date_recorded, then drop the original column X['year_recorded'] = X['date_recorded'].dt.year X['month_recorded'] = X['date_recorded'].dt.month X['day_recorded'] = X['date_recorded'].dt.day X = X.drop(columns='date_recorded') # Engineer feature: how many years from construction_year to date_recorded X['years'] = X['year_recorded'] - X['construction_year'] X['years_MISSING'] = X['years'].isnull() # return the wrangled dataframe return X train = wrangle(train) val = wrangle(val) test = wrangle(test) target = 'status_group' X_train = train.drop(columns=target) y_train = train[target] pipeline = Pipeline( [('ordinalencoder',ce.OrdinalEncoder()), ('simpleimputer',SimpleImputer()), ('randomforestregressor',RandomForestClassifier()) ] ) param_distributions = { 'randomforestregressor__n_estimators': randint(50, 500), 'randomforestregressor__max_depth': [5, 10, 15, 20, None], 'randomforestregressor__max_features': uniform(0, 1), } search = RandomizedSearchCV( pipeline, param_distributions=param_distributions, n_iter=10, cv=3, scoring='accuracy', verbose=10, return_train_score=True, n_jobs=-1 ) search.fit(X_train, y_train); print('Best hyperparameters', search.best_params_) print('Cross-validation MAE', search.best_score_) ###Output Best hyperparameters {'randomforestregressor__max_depth': 20, 'randomforestregressor__max_features': 0.10721042087254384, 'randomforestregressor__n_estimators': 342} Cross-validation MAE 0.8033670033670034
nb/GetHansard.ipynb
###Markdown Gets the URLS for Downloading the Hansard for a particular yearMost of the work was done here, Thanks Tim Sherratt, you're amazing! (https://timsherratt.org)https://github.com/GLAM-Workbench/australian-commonwealth-hansard ###Code import re import os import time import math import requests import arrow import csv import pandas as pd from tqdm import tqdm from bs4 import BeautifulSoup from requests.adapters import HTTPAdapter from requests.packages.urllib3.util.retry import Retry s = requests.Session() retries = Retry(total=5, backoff_factor=1, status_forcelist=[ 502, 503, 504 ]) s.mount('https://', HTTPAdapter(max_retries=retries)) s.mount('http://', HTTPAdapter(max_retries=retries)) URLS = { 'hofreps': ( 'http://parlinfo.aph.gov.au/parlInfo/search/summary/summary.w3p;' 'adv=yes;orderBy=date-eLast;page={page};' 'query={query}%20Dataset%3Ahansardr,hansardr80;resCount=100'), 'senate': ( 'http://parlinfo.aph.gov.au/parlInfo/search/summary/summary.w3p;' 'adv=yes;orderBy=date-eLast;page={page};' 'query={query}%20Dataset%3Ahansards,hansards80;resCount=100') } # write dictionary to csv # https://stackoverflow.com/questions/3086973/how-do-i-convert-this-list-of-dictionaries-to-a-csv-file def dict_to_csv(input_dict : dict, output_file : str,): with open(output_file, 'w', newline='') as of: dict_writer = csv.DictWriter(of, input_dict[0].keys()) dict_writer.writeheader() dict_writer.writerows(input_dict) def get_total_results(house, query): ''' Get the total number of results in the search. ''' # Insert query and page values into the ParlInfo url url = URLS[house].format(query=query, page=0) # Get the results page response = s.get(url) # Parse the HTML soup = BeautifulSoup(response.text) try: # Find where the total results are given in the HTML summary = soup.find('div', 'resultsSummary').contents[1].string # Extract the number of results from the string total = re.search(r'of (\d+)', summary).group(1) except AttributeError: total = 0 return int(total) def get_xml_url(url): ''' Extract the XML file url from an individual result. ''' # Load the page for an individual result response = s.get(url) # Parse the HTML soup = BeautifulSoup(response.text) # Find the XML url by looking for a pattern in the href xml_url = soup.find('a', href=re.compile('toc_unixml'))['href'] return xml_url def number_of_results(house, year): ''' Loop through a search by house and year, finding all the urls for XML files. ''' # Format the start and end dates start_date = '01%2F01%2F{}'.format(year) end_date = '31%2F12%2F{}'.format(year) # Prepare the query value using the start and end dates query = 'Date%3A{}%20>>%20{}'.format(start_date, end_date) # Get the total results total_results = get_total_results(house, query) xml_urls = [] dates = [] return total_results def harvest_year(house, year): ''' Loop through a search by house and year, finding all the urls for XML files. ''' # Format the start and end dates start_date = '01%2F01%2F{}'.format(year) end_date = '31%2F12%2F{}'.format(year) # Prepare the query value using the start and end dates query = 'Date%3A{}%20>>%20{}'.format(start_date, end_date) # Get the total results total_results = get_total_results(house, query) xml_urls = [] dates = [] if total_results > 0: # Calculate the number of pages in the results set num_pages = int(math.ceil(total_results / 100)) # Loop through the page range for page in range(0, num_pages + 1): # Get the next page of results url = URLS[house].format(query=query, page=page) response = s.get(url) # Parse the HTML soup = BeautifulSoup(response.text) # Find the list of results and loop through them for result in (soup.find_all('div', 'resultContent')): # Try to identify the date try: date = re.search(r'Date: (\d{2}\/\d{2}\/\d{4})', result.find('div', 'sumMeta').get_text()).group(1) date = arrow.get(date, 'DD/MM/YYYY').format('YYYY-MM-DD') except AttributeError: #There are some dodgy dates -- we'll just ignore them date = None # If there's a date, and we haven't seen it already, we'll grab the details if date and date not in dates: dates.append(date) # Get the link to the individual result page # This is where the XML file links live result_link = result.find('div', 'sumLink').a['href'] # Get the XML file link from the individual record page xml_url = get_xml_url(result_link) # Save dates and links xml_urls.append({'date': date, 'url': 'https://parlinfo.aph.gov.au{}'.format(xml_url)}) time.sleep(1) time.sleep(1) return xml_urls YEAR = 2018 def get_and_write_urls(year : int): print("Processing Year {}".format(year)) try: senate_urls = harvest_year('senate', year) except: senate_urls = [] try: horeps_urls = harvest_year('hofreps', year) except: horeps_urls = [] print("Number of results: Senate => {} House of Reps => {}".format(len(senate_urls), len(horeps_urls))) senate_output_file = data_output_path+"/{}-au-hansard-senate.csv".format(year) horeps_output_file = data_output_path+"/{}-au-hansard-hofreps.csv".format(year) clean_url = lambda x: x['url'].split(";")[0]+'\n' with open(senate_output_file, "w") as output: output.writelines(map(clean_url, senate_urls)) with open(horeps_output_file, "w") as output: output.writelines(map(clean_url, horeps_urls)) for y in reversed(range(1979, 2009)): get_and_write_urls(y) ###Output Processing Year 2008 Number of results: Senate => 0 House of Reps => 69 Processing Year 2007 Number of results: Senate => 41 House of Reps => 0 Processing Year 2006 Number of results: Senate => 56 House of Reps => 68 Processing Year 2005 Number of results: Senate => 55 House of Reps => 67 Processing Year 2004 Number of results: Senate => 0 House of Reps => 0 Processing Year 2003 Number of results: Senate => 64 House of Reps => 74 Processing Year 2002 Number of results: Senate => 60 House of Reps => 69 Processing Year 2001 Number of results: Senate => 51 House of Reps => 56 Processing Year 2000 Number of results: Senate => 0 House of Reps => 73 Processing Year 1999 Number of results: Senate => 79 House of Reps => 73 Processing Year 1998 Number of results: Senate => 59 House of Reps => 56 Processing Year 1997 Number of results: Senate => 0 House of Reps => 0 Processing Year 1996 Number of results: Senate => 0 House of Reps => 0 Processing Year 1995 Number of results: Senate => 0 House of Reps => 0 Processing Year 1994 Number of results: Senate => 0 House of Reps => 0 Processing Year 1993 Number of results: Senate => 0 House of Reps => 0 Processing Year 1992 Number of results: Senate => 0 House of Reps => 0 Processing Year 1991 Number of results: Senate => 0 House of Reps => 0 Processing Year 1990 Number of results: Senate => 0 House of Reps => 0 Processing Year 1989 Number of results: Senate => 0 House of Reps => 0 Processing Year 1988 Number of results: Senate => 0 House of Reps => 0 Processing Year 1987 Number of results: Senate => 0 House of Reps => 0 Processing Year 1986 Number of results: Senate => 0 House of Reps => 0 Processing Year 1985 Number of results: Senate => 0 House of Reps => 0 Processing Year 1984 Number of results: Senate => 0 House of Reps => 0 Processing Year 1983 Number of results: Senate => 0 House of Reps => 0 Processing Year 1982 Number of results: Senate => 0 House of Reps => 0 Processing Year 1981 Number of results: Senate => 0 House of Reps => 0 Processing Year 1980
Ch13_CV/kaggle_cifar10.ipynb
###Markdown The following additional libraries are needed to run thisnotebook. Note that running on Colab is experimental, please report a Githubissue if you have any problem. ###Code !pip install d2l==0.14.3 !pip install -U mxnet-cu101mkl==1.6.0.post0 # updating mxnet to at least v1.6 ###Output Requirement already satisfied: d2l==0.14.3 in /usr/local/lib/python3.6/dist-packages (0.14.3) Requirement already satisfied: matplotlib in /usr/local/lib/python3.6/dist-packages (from d2l==0.14.3) (3.2.2) Requirement already satisfied: jupyter in /usr/local/lib/python3.6/dist-packages (from d2l==0.14.3) (1.0.0) Requirement already satisfied: pandas in /usr/local/lib/python3.6/dist-packages (from d2l==0.14.3) (1.0.5) Requirement already satisfied: numpy in /usr/local/lib/python3.6/dist-packages (from d2l==0.14.3) (1.18.5) Requirement already satisfied: kiwisolver>=1.0.1 in /usr/local/lib/python3.6/dist-packages (from matplotlib->d2l==0.14.3) (1.2.0) Requirement already satisfied: pyparsing!=2.0.4,!=2.1.2,!=2.1.6,>=2.0.1 in /usr/local/lib/python3.6/dist-packages (from matplotlib->d2l==0.14.3) (2.4.7) Requirement already satisfied: python-dateutil>=2.1 in /usr/local/lib/python3.6/dist-packages (from matplotlib->d2l==0.14.3) (2.8.1) Requirement already satisfied: cycler>=0.10 in /usr/local/lib/python3.6/dist-packages (from matplotlib->d2l==0.14.3) (0.10.0) Requirement already satisfied: nbconvert in /usr/local/lib/python3.6/dist-packages (from jupyter->d2l==0.14.3) (5.6.1) Requirement already satisfied: qtconsole in /usr/local/lib/python3.6/dist-packages (from jupyter->d2l==0.14.3) (4.7.6) Requirement already satisfied: jupyter-console in /usr/local/lib/python3.6/dist-packages (from jupyter->d2l==0.14.3) (5.2.0) Requirement already satisfied: ipywidgets in /usr/local/lib/python3.6/dist-packages (from jupyter->d2l==0.14.3) (7.5.1) Requirement already satisfied: notebook in /usr/local/lib/python3.6/dist-packages (from jupyter->d2l==0.14.3) (5.3.1) Requirement already satisfied: ipykernel in /usr/local/lib/python3.6/dist-packages (from jupyter->d2l==0.14.3) (4.10.1) Requirement already satisfied: pytz>=2017.2 in /usr/local/lib/python3.6/dist-packages (from pandas->d2l==0.14.3) (2018.9) Requirement already satisfied: six>=1.5 in /usr/local/lib/python3.6/dist-packages (from python-dateutil>=2.1->matplotlib->d2l==0.14.3) (1.15.0) Requirement already satisfied: testpath in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (0.4.4) Requirement already satisfied: traitlets>=4.2 in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (4.3.3) Requirement already satisfied: entrypoints>=0.2.2 in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (0.3) Requirement already satisfied: pandocfilters>=1.4.1 in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (1.4.2) Requirement already satisfied: jinja2>=2.4 in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (2.11.2) Requirement already satisfied: nbformat>=4.4 in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (5.0.7) Requirement already satisfied: defusedxml in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (0.6.0) Requirement already satisfied: jupyter-core in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (4.6.3) Requirement already satisfied: bleach in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (3.1.5) Requirement already satisfied: pygments in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (2.1.3) Requirement already satisfied: mistune<2,>=0.8.1 in /usr/local/lib/python3.6/dist-packages (from nbconvert->jupyter->d2l==0.14.3) (0.8.4) Requirement already satisfied: jupyter-client>=4.1 in /usr/local/lib/python3.6/dist-packages (from qtconsole->jupyter->d2l==0.14.3) (5.3.5) Requirement already satisfied: qtpy in /usr/local/lib/python3.6/dist-packages (from qtconsole->jupyter->d2l==0.14.3) (1.9.0) Requirement already satisfied: ipython-genutils in /usr/local/lib/python3.6/dist-packages (from qtconsole->jupyter->d2l==0.14.3) (0.2.0) Requirement already satisfied: pyzmq>=17.1 in /usr/local/lib/python3.6/dist-packages (from qtconsole->jupyter->d2l==0.14.3) (19.0.2) Requirement already satisfied: prompt-toolkit<2.0.0,>=1.0.0 in /usr/local/lib/python3.6/dist-packages (from jupyter-console->jupyter->d2l==0.14.3) (1.0.18) Requirement already satisfied: ipython in /usr/local/lib/python3.6/dist-packages (from jupyter-console->jupyter->d2l==0.14.3) (5.5.0) Requirement already satisfied: widgetsnbextension~=3.5.0 in /usr/local/lib/python3.6/dist-packages (from ipywidgets->jupyter->d2l==0.14.3) (3.5.1) Requirement already satisfied: tornado>=4 in /usr/local/lib/python3.6/dist-packages (from notebook->jupyter->d2l==0.14.3) (5.1.1) Requirement already satisfied: terminado>=0.8.1 in /usr/local/lib/python3.6/dist-packages (from notebook->jupyter->d2l==0.14.3) (0.8.3) Requirement already satisfied: Send2Trash in /usr/local/lib/python3.6/dist-packages (from notebook->jupyter->d2l==0.14.3) (1.5.0) Requirement already satisfied: decorator in /usr/local/lib/python3.6/dist-packages (from traitlets>=4.2->nbconvert->jupyter->d2l==0.14.3) (4.4.2) Requirement already satisfied: MarkupSafe>=0.23 in /usr/local/lib/python3.6/dist-packages (from jinja2>=2.4->nbconvert->jupyter->d2l==0.14.3) (1.1.1) Requirement already satisfied: jsonschema!=2.5.0,>=2.4 in /usr/local/lib/python3.6/dist-packages (from nbformat>=4.4->nbconvert->jupyter->d2l==0.14.3) (2.6.0) Requirement already satisfied: packaging in /usr/local/lib/python3.6/dist-packages (from bleach->nbconvert->jupyter->d2l==0.14.3) (20.4) Requirement already satisfied: webencodings in /usr/local/lib/python3.6/dist-packages (from bleach->nbconvert->jupyter->d2l==0.14.3) (0.5.1) Requirement already satisfied: wcwidth in /usr/local/lib/python3.6/dist-packages (from prompt-toolkit<2.0.0,>=1.0.0->jupyter-console->jupyter->d2l==0.14.3) (0.2.5) Requirement already satisfied: setuptools>=18.5 in /usr/local/lib/python3.6/dist-packages (from ipython->jupyter-console->jupyter->d2l==0.14.3) (49.6.0) Requirement already satisfied: pickleshare in /usr/local/lib/python3.6/dist-packages (from ipython->jupyter-console->jupyter->d2l==0.14.3) (0.7.5) Requirement already satisfied: simplegeneric>0.8 in /usr/local/lib/python3.6/dist-packages (from ipython->jupyter-console->jupyter->d2l==0.14.3) (0.8.1) Requirement already satisfied: pexpect; sys_platform != "win32" in /usr/local/lib/python3.6/dist-packages (from ipython->jupyter-console->jupyter->d2l==0.14.3) (4.8.0) Requirement already satisfied: ptyprocess; os_name != "nt" in /usr/local/lib/python3.6/dist-packages (from terminado>=0.8.1->notebook->jupyter->d2l==0.14.3) (0.6.0) Collecting mxnet-cu101mkl==1.6.0.post0 [?25l Downloading https://files.pythonhosted.org/packages/45/3f/e33e3f92110fa5caba5e9eb052008208a33c1d5faccc7fe5312532e9aa42/mxnet_cu101mkl-1.6.0.post0-py2.py3-none-manylinux1_x86_64.whl (712.3MB)  |████████████████████████████████| 712.3MB 27kB/s [?25hCollecting graphviz<0.9.0,>=0.8.1 Downloading https://files.pythonhosted.org/packages/53/39/4ab213673844e0c004bed8a0781a0721a3f6bb23eb8854ee75c236428892/graphviz-0.8.4-py2.py3-none-any.whl Requirement already satisfied, skipping upgrade: requests<3,>=2.20.0 in /usr/local/lib/python3.6/dist-packages (from mxnet-cu101mkl==1.6.0.post0) (2.23.0) Requirement already satisfied, skipping upgrade: numpy<2.0.0,>1.16.0 in /usr/local/lib/python3.6/dist-packages (from mxnet-cu101mkl==1.6.0.post0) (1.18.5) Requirement already satisfied, skipping upgrade: idna<3,>=2.5 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.20.0->mxnet-cu101mkl==1.6.0.post0) (2.10) Requirement already satisfied, skipping upgrade: chardet<4,>=3.0.2 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.20.0->mxnet-cu101mkl==1.6.0.post0) (3.0.4) Requirement already satisfied, skipping upgrade: certifi>=2017.4.17 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.20.0->mxnet-cu101mkl==1.6.0.post0) (2020.6.20) Requirement already satisfied, skipping upgrade: urllib3!=1.25.0,!=1.25.1,<1.26,>=1.21.1 in /usr/local/lib/python3.6/dist-packages (from requests<3,>=2.20.0->mxnet-cu101mkl==1.6.0.post0) (1.24.3) Installing collected packages: graphviz, mxnet-cu101mkl Found existing installation: graphviz 0.10.1 Uninstalling graphviz-0.10.1: Successfully uninstalled graphviz-0.10.1 Successfully installed graphviz-0.8.4 mxnet-cu101mkl-1.6.0.post0 ###Markdown Image Classification (CIFAR-10) on Kaggle:label:`sec_kaggle_cifar10`So far, we have been using Gluon's `data` package to directly obtain image datasets in the tensor format. In practice, however, image datasets often exist in the format of image files. In this section, we will start with the original image files and organize, read, and convert the files to the tensor format step by step.We performed an experiment on the CIFAR-10 dataset in :numref:`sec_image_augmentation`.This is an important dataset in the computer vision field. Now, we will apply the knowledge we learned inthe previous sections in order to participate in the Kaggle competition, whichaddresses CIFAR-10 image classification problems. The competition's web addressis> https://www.kaggle.com/c/cifar-10:numref:`fig_kaggle_cifar10` shows the information on the competition's webpage. In order to submit the results, please register an account on the Kaggle website first.![CIFAR-10 image classification competition webpage information. The dataset for the competition can be accessed by clicking the "Data" tab.](https://github.com/d2l-ai/d2l-en-colab/blob/master/img/kaggle_cifar10.png?raw=1):width:`600px`:label:`fig_kaggle_cifar10`First, import the packages or modules required for the competition. ###Code import collections from d2l import mxnet as d2l import math from mxnet import autograd, gluon, init, npx from mxnet.gluon import nn import os import pandas as pd import shutil import time npx.set_np() ###Output _____no_output_____ ###Markdown Obtaining and Organizing the DatasetThe competition data is divided into a training set and testing set. The training set contains $50,000$ images. The testing set contains $300,000$ images, of which $10,000$ images are used for scoring, while the other $290,000$ non-scoring images are included to prevent the manual labeling of the testing set and the submission of labeling results. The image formats in both datasets are PNG, with heights and widths of 32 pixels and three color channels (RGB). The images cover $10$ categories: planes, cars, birds, cats, deer, dogs, frogs, horses, boats, and trucks. The upper-left corner of Figure 9.16 shows some images of planes, cars, and birds in the dataset. Downloading the DatasetAfter logging in to Kaggle, we can click on the "Data" tab on the CIFAR-10 image classification competition webpage shown in :numref:`fig_kaggle_cifar10` and download the dataset by clicking the "Download All" button. After unzipping the downloaded file in `../data`, and unzipping `train.7z` and `test.7z` inside it, you will find the entire dataset in the following paths:* ../data/cifar-10/train/[1-50000].png* ../data/cifar-10/test/[1-300000].png* ../data/cifar-10/trainLabels.csv* ../data/cifar-10/sampleSubmission.csvHere folders `train` and `test` contain the training and testing images respectively, `trainLabels.csv` has labels for the training images, and `sample_submission.csv` is a sample of submission. To make it easier to get started, we provide a small-scale sample of the dataset: it contains the first $1000$ training images and $5$ random testing images.To use the full dataset of the Kaggle competition, you need to set the following `demo` variable to `False`. ###Code #@save d2l.DATA_HUB['cifar10_tiny'] = (d2l.DATA_URL + 'kaggle_cifar10_tiny.zip', '2068874e4b9a9f0fb07ebe0ad2b29754449ccacd') # If you use the full dataset downloaded for the Kaggle competition, set # `demo` to False demo = True if demo: data_dir = d2l.download_extract('cifar10_tiny') else: data_dir = '../data/cifar-10/' ###Output Downloading ../data/kaggle_cifar10_tiny.zip from http://d2l-data.s3-accelerate.amazonaws.com/kaggle_cifar10_tiny.zip... ###Markdown Organizing the DatasetWe need to organize datasets to facilitate model training and testing. Let us first read the labels from the csv file. The following function returns a dictionary that maps the filename without extension to its label. ###Code #@save def read_csv_labels(fname): """Read fname to return a name to label dictionary.""" with open(fname, 'r') as f: # Skip the file header line (column name) lines = f.readlines()[1:] tokens = [l.rstrip().split(',') for l in lines] return dict(((name, label) for name, label in tokens)) labels = read_csv_labels(os.path.join(data_dir, 'trainLabels.csv')) print('# training examples:', len(labels)) print('# classes:', len(set(labels.values()))) ###Output # training examples: 1000 # classes: 10 ###Markdown Next, we define the `reorg_train_valid` function to segment the validation set from the original training set. The argument `valid_ratio` in this function is the ratio of the number of examples in the validation set to the number of examples in the original training set. In particular, let $n$ be the number of images of the class with the least examples, and $r$ be the ratio, then we will use $\max(\lfloor nr\rfloor,1)$ images for each class as the validation set. Let us use `valid_ratio=0.1` as an example. Since the original training set has $50,000$ images, there will be $45,000$ images used for training and stored in the path "`train_valid_test/train`" when tuning hyperparameters, while the other $5,000$ images will be stored as validation set in the path "`train_valid_test/valid`". After organizing the data, images of the same class will be placed under the same folder so that we can read them later. ###Code #@save def copyfile(filename, target_dir): """Copy a file into a target directory.""" d2l.mkdir_if_not_exist(target_dir) shutil.copy(filename, target_dir) #@save def reorg_train_valid(data_dir, labels, valid_ratio): # The number of examples of the class with the least examples in the # training dataset n = collections.Counter(labels.values()).most_common()[-1][1] # The number of examples per class for the validation set n_valid_per_label = max(1, math.floor(n * valid_ratio)) label_count = {} for train_file in os.listdir(os.path.join(data_dir, 'train')): label = labels[train_file.split('.')[0]] fname = os.path.join(data_dir, 'train', train_file) # Copy to train_valid_test/train_valid with a subfolder per class copyfile(fname, os.path.join(data_dir, 'train_valid_test', 'train_valid', label)) if label not in label_count or label_count[label] < n_valid_per_label: # Copy to train_valid_test/valid copyfile(fname, os.path.join(data_dir, 'train_valid_test', 'valid', label)) label_count[label] = label_count.get(label, 0) + 1 else: # Copy to train_valid_test/train copyfile(fname, os.path.join(data_dir, 'train_valid_test', 'train', label)) return n_valid_per_label ###Output _____no_output_____ ###Markdown The `reorg_test` function below is used to organize the testing set to facilitate the reading during prediction. ###Code #@save def reorg_test(data_dir): for test_file in os.listdir(os.path.join(data_dir, 'test')): copyfile(os.path.join(data_dir, 'test', test_file), os.path.join(data_dir, 'train_valid_test', 'test', 'unknown')) ###Output _____no_output_____ ###Markdown Finally, we use a function to call the previously defined `read_csv_labels`, `reorg_train_valid`, and `reorg_test` functions. ###Code def reorg_cifar10_data(data_dir, valid_ratio): labels = read_csv_labels(os.path.join(data_dir, 'trainLabels.csv')) reorg_train_valid(data_dir, labels, valid_ratio) reorg_test(data_dir) ###Output _____no_output_____ ###Markdown We only set the batch size to $4$ for the demo dataset. During actual training and testing, the complete dataset of the Kaggle competition should be used and `batch_size` should be set to a larger integer, such as $128$. We use $10\%$ of the training examples as the validation set for tuning hyperparameters. ###Code batch_size = 4 if demo else 128 valid_ratio = 0.1 reorg_cifar10_data(data_dir, valid_ratio) ###Output _____no_output_____ ###Markdown Image AugmentationTo cope with overfitting, we use image augmentation. For example, by adding `transforms.RandomFlipLeftRight()`, the images can be flipped at random. We can also perform normalization for the three RGB channels of color images using `transforms.Normalize()`. Below, we list some of these operations that you can choose to use or modify depending on requirements. ###Code transform_train = gluon.data.vision.transforms.Compose([ # Magnify the image to a square of 40 pixels in both height and width gluon.data.vision.transforms.Resize(40), # Randomly crop a square image of 40 pixels in both height and width to # produce a small square of 0.64 to 1 times the area of the original # image, and then shrink it to a square of 32 pixels in both height and # width gluon.data.vision.transforms.RandomResizedCrop(32, scale=(0.64, 1.0), ratio=(1.0, 1.0)), gluon.data.vision.transforms.RandomFlipLeftRight(), gluon.data.vision.transforms.ToTensor(), # Normalize each channel of the image gluon.data.vision.transforms.Normalize([0.4914, 0.4822, 0.4465], [0.2023, 0.1994, 0.2010])]) ###Output _____no_output_____ ###Markdown In order to ensure the certainty of the output during testing, we only perform normalization on the image. ###Code transform_test = gluon.data.vision.transforms.Compose([ gluon.data.vision.transforms.ToTensor(), gluon.data.vision.transforms.Normalize([0.4914, 0.4822, 0.4465], [0.2023, 0.1994, 0.2010])]) ###Output _____no_output_____ ###Markdown Reading the DatasetNext, we can create the `ImageFolderDataset` instance to read the organized dataset containing the original image files, where each example includes the image and label. ###Code train_ds, valid_ds, train_valid_ds, test_ds = [ gluon.data.vision.ImageFolderDataset( os.path.join(data_dir, 'train_valid_test', folder)) for folder in ['train', 'valid', 'train_valid', 'test']] ###Output _____no_output_____ ###Markdown We specify the defined image augmentation operation in `DataLoader`. During training, we only use the validation set to evaluate the model, so we need to ensure the certainty of the output. During prediction, we will train the model on the combined training set and validation set to make full use of all labelled data. ###Code train_iter, train_valid_iter = [gluon.data.DataLoader( dataset.transform_first(transform_train), batch_size, shuffle=True, last_batch='discard') for dataset in (train_ds, train_valid_ds)] valid_iter = gluon.data.DataLoader( valid_ds.transform_first(transform_test), batch_size, shuffle=False, last_batch='discard') test_iter = gluon.data.DataLoader( test_ds.transform_first(transform_test), batch_size, shuffle=False, last_batch='keep') ###Output _____no_output_____ ###Markdown Defining the ModelHere, we build the residual blocks based on the `HybridBlock` class, which isslightly different than the implementation described in:numref:`sec_resnet`. This is done to improve execution efficiency. ###Code class Residual(nn.HybridBlock): def __init__(self, num_channels, use_1x1conv=False, strides=1, **kwargs): super(Residual, self).__init__(**kwargs) self.conv1 = nn.Conv2D(num_channels, kernel_size=3, padding=1, strides=strides) self.conv2 = nn.Conv2D(num_channels, kernel_size=3, padding=1) if use_1x1conv: self.conv3 = nn.Conv2D(num_channels, kernel_size=1, strides=strides) else: self.conv3 = None self.bn1 = nn.BatchNorm() self.bn2 = nn.BatchNorm() def hybrid_forward(self, F, X): Y = F.npx.relu(self.bn1(self.conv1(X))) Y = self.bn2(self.conv2(Y)) if self.conv3: X = self.conv3(X) return F.npx.relu(Y + X) ###Output _____no_output_____ ###Markdown Next, we define the ResNet-18 model. ###Code def resnet18(num_classes): net = nn.HybridSequential() net.add(nn.Conv2D(64, kernel_size=3, strides=1, padding=1), nn.BatchNorm(), nn.Activation('relu')) def resnet_block(num_channels, num_residuals, first_block=False): blk = nn.HybridSequential() for i in range(num_residuals): if i == 0 and not first_block: blk.add(Residual(num_channels, use_1x1conv=True, strides=2)) else: blk.add(Residual(num_channels)) return blk net.add(resnet_block(64, 2, first_block=True), resnet_block(128, 2), resnet_block(256, 2), resnet_block(512, 2)) net.add(nn.GlobalAvgPool2D(), nn.Dense(num_classes)) return net ###Output _____no_output_____ ###Markdown The CIFAR-10 image classification challenge uses 10 categories. We will perform Xavier random initialization on the model before training begins. ###Code def get_net(devices): num_classes = 10 net = resnet18(num_classes) net.initialize(ctx=devices, init=init.Xavier()) return net loss = gluon.loss.SoftmaxCrossEntropyLoss() ###Output _____no_output_____ ###Markdown Defining the Training FunctionsWe will select the model and tune hyperparameters according to the model's performance on the validation set. Next, we define the model training function `train`. We record the training time of each epoch, which helps us compare the time costs of different models. ###Code def train(net, train_iter, valid_iter, num_epochs, lr, wd, devices, lr_period, lr_decay): trainer = gluon.Trainer(net.collect_params(), 'sgd', {'learning_rate': lr, 'momentum': 0.9, 'wd': wd}) num_batches, timer = len(train_iter), d2l.Timer() animator = d2l.Animator(xlabel='epoch', xlim=[0, num_epochs], legend=['train loss', 'train acc', 'valid acc']) for epoch in range(num_epochs): metric = d2l.Accumulator(3) if epoch > 0 and epoch % lr_period == 0: trainer.set_learning_rate(trainer.learning_rate * lr_decay) for i, (features, labels) in enumerate(train_iter): timer.start() l, acc = d2l.train_batch_ch13( net, features, labels.astype('float32'), loss, trainer, devices, d2l.split_batch) metric.add(l, acc, labels.shape[0]) timer.stop() if (i + 1) % (num_batches // 5) == 0: animator.add(epoch + i / num_batches, (metric[0] / metric[2], metric[1] / metric[2], None)) if valid_iter is not None: valid_acc = d2l.evaluate_accuracy_gpus(net, valid_iter, d2l.split_batch) animator.add(epoch + 1, (None, None, valid_acc)) if valid_iter is not None: print(f'loss {metric[0] / metric[2]:.3f}, ' f'train acc {metric[1] / metric[2]:.3f}, ' f'valid acc {valid_acc:.3f}') else: print(f'loss {metric[0] / metric[2]:.3f}, ' f'train acc {metric[1] / metric[2]:.3f}') print(f'{metric[2] * num_epochs / timer.sum():.1f} examples/sec ' f'on {str(devices)}') ###Output _____no_output_____ ###Markdown Training and Validating the ModelNow, we can train and validate the model. The following hyperparameters can be tuned. For example, we can increase the number of epochs. Because `lr_period` and `lr_decay` are set to 80 and 0.1 respectively, the learning rate of the optimization algorithm will be multiplied by 0.1 after every 80 epochs. For simplicity, we only train one epoch here. ###Code devices, num_epochs, lr, wd = d2l.try_all_gpus(), 5, 0.1, 5e-4 lr_period, lr_decay, net = 50, 0.1, get_net(devices) net.hybridize() train(net, train_iter, valid_iter, num_epochs, lr, wd, devices, lr_period, lr_decay) ###Output loss nan, train acc 0.102, valid acc 0.100 163.4 examples/sec on [gpu(0)] ###Markdown Classifying the Testing Set and Submitting Results on KaggleAfter obtaining a satisfactory model design and hyperparameters, we use all training datasets (including validation sets) to retrain the model and classify the testing set. ###Code net, preds = get_net(devices), [] net.hybridize() train(net, train_valid_iter, None, num_epochs, lr, wd, devices, lr_period, lr_decay) for X, _ in test_iter: y_hat = net(X.as_in_ctx(devices[0])) preds.extend(y_hat.argmax(axis=1).astype(int).asnumpy()) sorted_ids = list(range(1, len(test_ds) + 1)) sorted_ids.sort(key=lambda x: str(x)) df = pd.DataFrame({'id': sorted_ids, 'label': preds}) df['label'] = df['label'].apply(lambda x: train_valid_ds.synsets[x]) df.to_csv('submission.csv', index=False) ###Output loss nan, train acc 0.102 174.4 examples/sec on [gpu(0)]
Python/requests.ipynb
###Markdown Get Request ###Code requests.get('http://api.github.com') response = requests.get('http://api.github.com') response.status_code if response.status_code: print('status code between 200 and 400: status_code={}'.format(response.status_code)) else: print('false, status_code={}'.format(response.status_code)) from requests.exceptions import HTTPError for url in ['https://api.github.com', 'https://api.github.com/invalid']: try: response = requests.get(url) # If the response was successful, no Exception will be raised response.raise_for_status() except HTTPError as http_err: print(f'HTTP error occurred: {http_err}') # Python 3.6 except Exception as err: print(f'Other error occurred: {err}') # Python 3.6 else: print('Success!') response.content response.text response.encoding = 'utf-8' # Optional: requests infers this internally response.text response = requests.get('http://api.github.com') response.json() response.headers ###Output _____no_output_____ ###Markdown Request Download ImageAdvanced Usage: https://docs.python-requests.org/en/master/user/advanced/prepared-requests ###Code image_url = "https://www.apple.com/ac/structured-data/images/open_graph_logo.png?201810271035" resp = requests.get(image_url, stream=True) print(f"resp.status_code: {resp.status_code}") resp.content ###Output _____no_output_____
_build/html/_sources/curriculum-notebooks/Science/SpecificAndLatentHeat/specific-and-latent-heat.ipynb
###Markdown ![Callysto.ca Banner](https://github.com/callysto/curriculum-notebooks/blob/master/callysto-notebook-banner-top.jpg?raw=true) (Click **Cell** > **Run All** before proceeding.) ###Code %matplotlib inline #---------- #Import modules and packages import ipywidgets as widgets import random import math import matplotlib.pyplot as plt from ipywidgets import Output, IntSlider, VBox, HBox, Layout from IPython.display import clear_output, display, HTML, Javascript, SVG #---------- #import ipywidgets as widgets #import random #This function produces a multiple choice form with four options def multiple_choice(option_1, option_2, option_3, option_4): option_list = [option_1, option_2, option_3, option_4] answer = option_list[0] letters = ["(A) ", "(B) ", "(C) ", "(D) "] #Boldface letters at the beginning of each option start_bold = "\033[1m"; end_bold = "\033[0;0m" #Randomly shuffle the options random.shuffle(option_list) #Print the letters (A) to (D) in sequence with randomly chosen options for i in range(4): option_text = option_list.pop() print(start_bold + letters[i] + end_bold + option_text) #Store the correct answer if option_text == answer: letter_answer = letters[i] button1 = widgets.Button(description="(A)"); button2 = widgets.Button(description="(B)") button3 = widgets.Button(description="(C)"); button4 = widgets.Button(description="(D)") button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Whitesmoke' container = widgets.HBox(children=[button1,button2,button3,button4]) display(container) print(" ", end='\r') def on_button1_clicked(b): if "(A) " == letter_answer: print("Correct! ", end='\r') button1.style.button_color = 'Moccasin'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Whitesmoke' else: print("Try again.", end='\r') button1.style.button_color = 'Lightgray'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Whitesmoke' def on_button2_clicked(b): if "(B) " == letter_answer: print("Correct! ", end='\r') button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Moccasin' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Whitesmoke' else: print("Try again.", end='\r') button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Lightgray' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Whitesmoke' def on_button3_clicked(b): if "(C) " == letter_answer: print("Correct! ", end='\r') button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Moccasin'; button4.style.button_color = 'Whitesmoke' else: print("Try again.", end='\r') button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Lightgray'; button4.style.button_color = 'Whitesmoke' def on_button4_clicked(b): if "(D) " == letter_answer: print("Correct! ", end='\r') button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Moccasin' else: print("Try again.", end='\r') button1.style.button_color = 'Whitesmoke'; button2.style.button_color = 'Whitesmoke' button3.style.button_color = 'Whitesmoke'; button4.style.button_color = 'Lightgray' button1.on_click(on_button1_clicked); button2.on_click(on_button2_clicked) button3.on_click(on_button3_clicked); button4.on_click(on_button4_clicked) ###Output _____no_output_____ ###Markdown Specific and Latent Heat Introduction**Heat** is defined as the *transfer of energy* from one object to another due to a difference in their relative temperatures. As heat flows from one object into another, the temperature of either one or both objects changes. Specific Heat CapacityThe amount of heat required to change the temperature of a given material is given by the following equation:$$Q = m C \Delta T$$where $Q$ represents heat in joules (J), $m$ represents mass kilograms (kg), and $\Delta T$ represents the change in temperature in Celsius (°C) or kelvin (K). The parameter $C$ is an experimentally determined value characteristic of a particular material. This parameter is called the **specific heat** or **specific heat capacity** (J/kg$\cdot$°C). The specific heat capacity of a material is determined by measuring the amount of heat required to raise the temperature of 1 kg of the material by 1°C. For ordinary temperatures and pressures, the value of $C$ is considered constant. Values for the specific heat capacity of common materials are shown in the table below: Material | Specific Heat Capacity (J/kg$\cdot$°C) --- | --- Aluminum | 903 Brass | 376 Carbon | 710 Copper | 385 Glass | 664 Ice | 2060 Iron | 450 Lead | 130 Methanol | 2450 Silver | 235 Stainless Steal | 460 Steam | 2020 Tin | 217 Water | 4180 Zinc | 388 Use the slider below to observe the relationship between the specific heat capacity and the amount of heat required to raise the temperature of a 5 kg mass by 50 °C. ###Code #import ipywidgets as widgets #from ipywidgets import Output, VBox, HBox mass_1 = 5 delta_temperature = 50 specific_heat_capacity = widgets.IntSlider(description="C (J/kg⋅°C)",min=100,max=1000) #Boldface text between these strings start_bold = "\033[1m"; end_bold = "\033[0;0m" def f(specific_heat_capacity): heat_J = int((mass_1 * specific_heat_capacity * delta_temperature)) heat_kJ = int(heat_J/1000) print(start_bold + "Heat = (mass) X (specific heat capacity) X (change in temperature)" + end_bold) print("Heat = ({} X {} X {}) J = {} J or {} kJ".format(mass_1, specific_heat_capacity, delta_temperature, heat_J, heat_kJ)) out1 = widgets.interactive_output(f,{'specific_heat_capacity': specific_heat_capacity,}) HBox([VBox([specific_heat_capacity]), out1]) ###Output _____no_output_____ ###Markdown **Question:** *As the specific heat increases, the amount of heat required to cause the temperature change:* ###Code #import ipywidgets as widgets #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = "Increases" option_2 = "Decreases" option_3 = "Remains constant" option_4 = "Equals zero" multiple_choice(option_1, option_2, option_3, option_4) ###Output (A) Remains constant (B) Decreases (C) Increases (D) Equals zero ###Markdown ExampleHow many kilojoules (kJ) of heat are needed to raise the temperature of a 3.0 kg piece of aluminum from 10°C to 50°C? Round the answer to 2 significant figures. ###Code #import ipywidgets as widgets #from ipywidgets import Output, VBox #from IPython.display import clear_output, display, HTML out2 = Output() button_step1 = widgets.Button(description="Step One", layout=Layout(width='20%', height='100%'), button_style='primary') count1 = 1 text1_1 = widgets.HTMLMath(value="The first step is to identify all known and unknown variables required to solve the problem. In this case, three variables are known ($m$, $C$, $\Delta T$), and one variable is unknown ($Q$):") text1_2 = widgets.HTMLMath(value="$m$ = 3.0 kg") text1_3 = widgets.HTMLMath(value="$\Delta T$ = 50°C $-$ 10°C = 40°C") text1_4 = widgets.HTMLMath(value="$C$ = 903 J/kg$\cdot$°C (The specific heat capacity for aluminum may be found in the table above.)") text1_5 = widgets.HTMLMath(value="$Q$ = ?") def on_button_step1_clicked(b): global count1 count1 += 1 with out2: clear_output() if count1 % 2 == 0: display(text1_1, text1_2, text1_3, text1_4, text1_5) display(VBox([button_step1, out2])) button_step1.on_click(on_button_step1_clicked) #import ipywidgets as widgets #from ipywidgets import Output, VBox #from IPython.display import clear_output, display, HTML out3 = Output() button_step2 = widgets.Button(description="Step Two", layout=Layout(width='20%', height='100%'), button_style='primary') count2 = 1 text2_1 = widgets.HTMLMath(value="Substitute each known variable into the formula to solve for the unknown variable:") text2_2 = widgets.HTMLMath(value="$Q = mC\Delta T$") text2_3 = widgets.HTMLMath(value="$Q$ = (3.0 kg) (903 J/kg$\cdot$°C) (40°C) = 108,360 J") text2_4 = widgets.HTMLMath(value="$Q$ = 108,360 J") def on_button_step2_clicked(b): global count2 count2 += 1 with out3: clear_output() if count2 % 2 == 0: display(text2_1, text2_2, text2_3, text2_4) display(VBox([button_step2, out3])) button_step2.on_click(on_button_step2_clicked) #import ipywidgets as widgets #from ipywidgets import Output, VBox #from IPython.display import clear_output, display, HTML out4 = Output() button_step3 = widgets.Button(description="Step Three", layout=Layout(width='20%', height='100%'), button_style='primary') count3 = 1 text3_1 = widgets.HTMLMath(value="Round the answer to the correct number of significant figures and convert to the correct units (if needed):") text3_2 = widgets.HTMLMath(value="$Q$ = 108,360 J = 110,000 J or 110 kJ") text3_3 = widgets.HTMLMath(value="The amount of heat required to increase the temperature of a 3.0 kg piece of aluminum from 10°C to 50°C is 110,000 J or 110 kJ.") def on_button_step3_clicked(b): global count3 count3 += 1 with out4: clear_output() if count3 % 2 == 0: display(text3_1, text3_2, text3_3) display(VBox([button_step3, out4])) button_step3.on_click(on_button_step3_clicked) ###Output _____no_output_____ ###Markdown PracticeThe heat transfer equation shown above may be rearranged to solve for each variable in the equation. These rearrangements are shown below:$$Q = mC\Delta T \qquad m = \dfrac{Q}{C \Delta T} \qquad C = \dfrac{Q}{m \Delta T} \qquad \Delta T = \dfrac{Q}{mC}$$Try the four different practice problems below. Each question will require the use of one or more formula above. Use the *Generate New Question* button to generate additional practice problems. ###Code #from IPython.display import Javascript, display #from ipywidgets import widgets def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Randomize variables mass = round(random.uniform(25.0, 50.0), 1) temperature_initial = round(random.uniform(15.0, 25.0), 1) temperature_final = round(random.uniform(55.0, 65.0), 1) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "silver": 235, "stainless steal": 460, "tin": 217, "zinc": 388} material = random.choice(list(materials.keys())) #Print question question = "How much heat is required to raise the temperature of a {} g sample of {} from {}°C to {}°C?".format(mass, material, temperature_initial, temperature_final) print(question) #Answer and option calculations answer = (mass/1000) * materials[material] * (temperature_final - temperature_initial) #Define range of values for random multiple choices mini = 100 maxa = 2300 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Round options to the specified number of significant figures def round_sf(number, significant): return round(number, significant - len(str(number))) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(round_sf(int(answer),3)) + " J" option_2 = str(round_sf(int(choice_list[0]),3)) + " J" option_3 = str(round_sf(int(choice_list[1]),3)) + " J" option_4 = str(round_sf(int(choice_list[2]),3)) + " J" multiple_choice(option_1, option_2, option_3, option_4) #import math #from IPython.display import Javascript, display #from ipywidgets import widgets def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Randomize variables heat = random.randint(10, 250) temperature_initial = round(random.uniform(10.0, 35.0), 1) temperature_final = round(random.uniform(45.0, 100.0), 1) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "silver": 235, "stainless steal": 460, "tin": 217, "zinc": 388} material = random.choice(list(materials.keys())) #Print question question = "Suppose some {} lost {} kJ of heat as it cooled from {}°C to {}°C. Find the mass. Note: you will need to make the sign of Q negative because heat is flowing out of the material as it cools.".format(material, heat, temperature_final, temperature_initial) print(question) #Answer calculation answer = (-heat*1000) / (materials[material] * (temperature_initial - temperature_final)) #Define range of values for random multiple choices mini = 100 maxa = 2000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str('{:.2f}'.format(round(answer,2))) + " kg" option_2 = str(round(choice_list[0],2)/100) + " kg" option_3 = str(round(choice_list[1],2)/100) + " kg" option_4 = str(round(choice_list[2],2)/100) + " kg" multiple_choice(option_1, option_2, option_3, option_4) #from IPython.display import Javascript, display #from ipywidgets import widgets def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Randomize variables heat = round(random.uniform(23.00, 26.00),1) mass = round(random.uniform(1.00, 3.00), 2) temperature_initial = round(random.uniform(24.0, 25.0), 1) temperature_final = round(random.uniform(35.0, 36.0), 1) #Print question question = "A newly made synthetic material weighing {} kg requires {} kJ to go from {}°C to {}°C (without changing state). What is the specific heat capacity of this new material?".format(mass, heat, temperature_initial, temperature_final) print(question) #Answer calculation answer = (heat*1000) / (mass * (temperature_final - temperature_initial)) #Define range of values for random multiple choices mini = 990 maxa = 2510 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Round options to the specified number of significant figures def round_sf(number, significant): return round(number, significant - len(str(number))) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(round_sf(int(answer),3)) + " J/(kg°C)" option_2 = str(round_sf(int(choice_list[0]),3)) + " J/(kg°C)" option_3 = str(round_sf(int(choice_list[1]),3)) + " J/(kg°C)" option_4 = str(round_sf(int(choice_list[2]),3)) + " J/(kg°C)" multiple_choice(option_1, option_2, option_3, option_4) #import math #from IPython.display import Javascript, display #from ipywidgets import widgets def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "silver": 235, "stainless steal": 460, "tin": 217, "zinc": 388} material = random.choice(list(materials.keys())) #Randomize Variables heat = random.randint(100, 150) mass = round(random.uniform(1.0, 5.0), 1) temperature_initial = round(random.uniform(10.0, 30.0), 1) temperature_final = round(random.uniform(40.0, 60.0), 1) #Determine question type question_type = random.randint(1,3) if question_type == 1: #Type 1: Finding change in temperature question = "If {} kg of {} receives {} kJ of heat, determine its change in temperature to one decimal place.".format(mass, material, heat) print(question) answer = (heat*1000) / (materials[material] * mass) elif question_type == 2: #Type 2: Finding final temperature question = "If {} kg of {} receives {} kJ of heat, and if the {}'s initial temperature is {}°C, determine its final temperature to one decimal place. Hint: ΔT = final temperature - initial temperature.".format(mass, material, heat, material, temperature_initial) print(question) answer = ((heat*1000) / (materials[material] * mass)) + temperature_initial elif question_type == 3: #Type 3: Finding initial temperature question = "If {} kg of {} receives {} kJ of heat, and if the {}'s final temperature is {}°C, determine its initial temperature to one decimal place. Hint: ΔT = final temperature - initial temperature.".format(mass, material, heat, material, temperature_final) print(question) answer = temperature_final - ((heat*1000) / (materials[material] * mass)) #Define range of values for random multiple choices mini = int(answer*100 - 1000) maxa = int(answer*100 + 1000) #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str((round(answer,1))) + " °C" option_2 = str(round(choice_list[0]/100,1)) + " °C" option_3 = str(round(choice_list[1]/100,1)) + " °C" option_4 = str(round(choice_list[2]/100,1)) + " °C" multiple_choice(option_1, option_2, option_3, option_4) ###Output _____no_output_____ ###Markdown Change of PhaseIn the previous examples and exercises, the material remained in a constant state while heat was added or taken away. However, the addition or subtraction of heat is often accompanied by a **phase change**. The three most common phases are solid, liquid, and gas: **Problem:** *Determine the amount of heat required to raise the temperature of a 100 g block of ice from -20°C to steam at 200°C.***Attempt:** There are two phase changes in this problem: (1) the melting of ice into water, and (2) the boiling of water into steam. To determine $Q$, let's utilize the heat formula: $$Q=mC\Delta T$$ To solve this problem, we can split it up into steps that are simple to calculate. For example, we can start by calculating the heat required to warm ice from -20°C to 0°C. Then, we can calculate the heat required to warm water from 0°C to 100°C. Finally, we can calculate the heat required to warm steam from 100°C to 200°C:$Q_{ice}$ = (0.100 kg) (2060 J/kg$\cdot$°C) (0°C - (-20°C)) = 4120 J$Q_{water}$ = (0.100 kg) (4180 J/kg$\cdot$°C) (100°C - 0°C) = 41800 J$Q_{steam}$ = (0.100 kg) (2020 J/kg$\cdot$°C) (200°C - 100°C) = 20200 JThen, by adding up the heat calculated in each step, the original problem can be solved: $Q$ = (4120 + 41800 + 20200) J = 66120 J, or 66.1 kJ. ExperimentLet's conduct an experiment to check the above calculation. We will start with a 100 g sample of ice at -20°C, and then add a constant amount of heat until the entire sample is converted to steam at 200°C. Every minute, we will take the temperature of the sample.The data from this experiment is shown in the interactive graphs below. The temperature of the material versus time is shown on left. The heat added to the material versus time is shown on the right. ###Code #import ipywidgets as widgets #import matplotlib.pyplot as plt #from ipywidgets import HBox, Output, VBox #from IPython.display import clear_output out5 = Output() play = widgets.Play(interval=500, value=0, min=0, max=25, step=1, description="Press play", disabled=False) time_slider = widgets.IntSlider(description='Time (min)', value=0, min=0, max=25, continuous_update = False) widgets.jslink((play, 'value'), (time_slider, 'value')) #Make lists of x and y values x_values = list(range(26)) y_values = [-20, -10, 0, 0, 10, 40, 80, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 100, 120, 140, 160, 180, 200] heat_y = [] increment = 0 for i in range(26): heat_y.append(increment) increment += 13.021 #Plot graphs def plot_graphs(change): x = change['new'] with out5: clear_output(wait=True) temp_x_values = [] temp_y_values = [] graph2y = [] for i in range(x+1): temp_x_values.append(x_values[i]) temp_y_values.append(y_values[i]) graph2y.append(heat_y[i]) plt.figure(figsize=(15,5)) plt.style.use('seaborn') plt.rcParams["axes.edgecolor"] = "black" plt.rcParams["axes.linewidth"] = 0.5 plt.subplot(1,2,1) plt.ylim(-30, 210) plt.xlim(-0.5,26) plt.scatter(temp_x_values, temp_y_values) plt.ylabel('Temperature (°C)') plt.xlabel('Time (min)') plt.subplot(1,2,2) plt.ylim(-25, 350) plt.xlim(-2,26) plt.scatter(temp_x_values, graph2y, color='red') plt.ylabel('Heat (kJ)') plt.xlabel('Time (min)') plt.show() #Get slider value time_slider.observe(plot_graphs, 'value') plot_graphs({'new': time_slider.value}) #Display widget display(HBox([play, time_slider])) display(out5) ###Output _____no_output_____ ###Markdown **Question**: *Examine the graph on the left. It shows the temperature of the material at each minute. At what temperature(s) does the temperature remain constant for some time?* ###Code #import ipywidgets as widgets #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = "0°C and 100°C. We have horizontal lines at those temperatures." option_2 = "-20°C, 0°C, 100°C, and 200°C." option_3 = "100°C." option_4 = "The temperature is never constant." multiple_choice(option_1, option_2, option_3, option_4) ###Output (A) The temperature is never constant. (B) 0°C and 100°C. We have horizontal lines at those temperatures. (C) 100°C. (D) -20°C, 0°C, 100°C, and 200°C. ###Markdown **Question:** *Examine the graph on the right. It shows how much heat was required to turn a block of ice at -20°C into steam at 200°C. Does this agree with the value we arrived at from our above calculation (66.1 kJ)?* ###Code #import ipywidgets as widgets #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = "Based on the graph, the amount of heat required is around 325 kJ. It does not agree with our calculation." option_2 = "Based on the graph, the amount of heat required is close enough to our calculation; hence, it does agree." option_3 = "Both values match perfectly." option_4 = "The values are close and it is impossible to say if they match perfectly or not." multiple_choice(option_1, option_2, option_3, option_4) ###Output (A) The values are close and it is impossible to say if they match perfectly or not. (B) Based on the graph, the amount of heat required is close enough to our calculation; hence, it does agree. (C) Based on the graph, the amount of heat required is around 325 kJ. It does not agree with our calculation. (D) Both values match perfectly. ###Markdown **Question**: *Examine the graph on the right. Observe that the slope of the line is constant. What does this imply?* ###Code #import ipywidgets as widgets #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = "The amount of heat added to the system is constant for the entire 25 min period." option_2 = "The amount of heat added to the system is not constant, the rate increases throughout the 25 min period." option_3 = "No heat is added at the start, but around 325 kJ of heat is added at the very end." option_4 = "As time increases, the amount of heat required decreases." multiple_choice(option_1, option_2, option_3, option_4) ###Output (A) No heat is added at the start, but around 325 kJ of heat is added at the very end. (B) As time increases, the amount of heat required decreases. (C) The amount of heat added to the system is not constant, the rate increases throughout the 25 min period. (D) The amount of heat added to the system is constant for the entire 25 min period. ###Markdown Experimental ResultsOur experimental data indicates that our calculation of 66.1 kJ is incorrect and that it in fact takes around 325 kJ to heat ice from -20°C to steam at 200°C. *So what did we miss?***Answer:** The *phase changes*.The graph on the right shows us that the rate heat was added to the system over the 25 minute period was constant, yet the temperature remained constant at two points for some time (0°C and 100°C). How is this possible? That is, *how can we add heat to a material while its temperature remains constant?***Answer:** Every material has two common "critical temperature points". These are the points at which the *state* of the material *changes*. For water, these points are at 0°C and 100°C. If heat is coming into a material *during a phase change*, then this energy is used to overcome the intermolecular forces between the molecules of the material.Let's consider when ice melts into water at 0°C. Immediately after the molecular bonds in the ice are broken, the molecules are moving (vibrating) at the same average speed as before, and so their average kinetic energy remains the same. *Temperature* is precisely a measure of the average kinetic energy of the particles in a material. Hence, during a phase change, the temperature remains constant. Latent Heat of Fusion and Vaporization The **latent heat of fusion ($H_f$)** is the quantity of heat needed to melt 1 kg of a solid to a liquid without a change in temperature.The **latent heat of vaporization ($H_v$)** is the quantity of heat needed to vaporise 1 kg of a liquid to a gas without a change in temperature.The latent heats of fusion and vaporization are empirical characteristics of a particular material. As such, they must be experimentally determined. Values for the latent heats of fusion and vaporization of common materials are shown in the table below:Materials | Heat of Fusion (J/kg) |Heat of Vaporization (J/kg) --- | --- | --- Copper | $2.05 \times 10^5$ | $5.07 \times 10^6$ Gold | $6.03 \times 10^4$ | $1.64 \times 10^6$ Iron | $2.66 \times 10^5$ | $6.29 \times 10^6$ Lead | $2.04 \times 10^4$ | $8.64 \times 10^5$ Mercury | $1.15 \times 10^4$ | $2.72 \times 10^5$ Methanol | $1.09 \times 10^5$ | $8.78 \times 10^5$ Silver | $1.04 \times 10^4$ | $2.36 \times 10^6$ Water (ice) | $3.34 \times 10^5$ | $2.26 \times 10^6$ The following formulae are used to calculate the amount of heat needed to change a material from a solid to a liquid (fusion), or from a liquid to a gas (vaporization):$$Q_f = mH_f \qquad Q_v = mH_v$$ Example (revisited)Recall our previous problem:**Problem:** *Determine the amount of heat required to raise the temperature of a 100 g block of ice from -20°C to steam at 200°C.***Solution:** Previously, we split the problem into three steps. It turns out that those steps correctly calculated the heat required to warm ice from -20°C to 0°C, water from 0°C to 100°C. and steam from 100°C to 200°C. What was absent was the latent heat required to complete the phase changes at 0°C and 100°C. Therefore, we need to **add two more steps**, which incorporate the above formulae. For completion, the previous steps are restated and the entire calculation is shown in **five steps** below (plus a final step to sum up the heats calculated in the previous steps): ###Code #import ipywidgets as widgets #from ipywidgets import Output, VBox, HBox #from IPython.display import clear_output, SVG, HTML, display out6 = Output() frame_1 = 1 #Toggle images def show_steps_1(): global frame_1 I11 = widgets.HTMLMath(value="Step 1: Calculate the heat required to change ice from -20°C to 0°C. Since the temperature changes, we use $Q = mCΔT$.") Q11 = widgets.HTMLMath(value="$Q_{1}$ = (0.1 kg) (2060 J/kg°C) (0°C - (-20°C)) = 4120 J") I12 = widgets.HTMLMath(value="Step 2: Calculate the heat required to change ice at 0°C to water at 0°C. Since the temperature does not change as we are at the melting point of water, we use $Q = mH_{f}$.") Q12 = widgets.HTMLMath(value="$Q_{2}$ = (0.1 kg) (334000 J/kg) = 33400 J") I13 = widgets.HTMLMath(value="Step 3: Calculate the heat required to change water from 0°C to 100°C. Since the temperature changes, we use $Q = mCΔT$.") Q13 = widgets.HTMLMath(value="$Q_{3}$ = (0.1 kg) (4180 J/kg°C) (100°C - 0°C) = 41800 J") I14 = widgets.HTMLMath(value="Step 4: Calculate the heat required to change water at 100°C to steam at 100°C. Since the temperature does not change at we are at the boiling point of water, we use $Q = mH_{v}$.") Q14 = widgets.HTMLMath(value="$Q_{4}$ = (0.1 kg) (2260000 J/kg) = 226000 J") I15 = widgets.HTMLMath(value="Step 5: Calculate the heat required to change steam from 100°C to 200°C. Since the temperature changes, we use $Q = mCΔT$.") Q15 = widgets.HTMLMath(value="$Q_{5}$ = (0.1 kg) (2020 J/kg°C) (200°C - 100°C) = 20200 J") I16 = widgets.HTMLMath(value="Summary: Calculate total heat by adding up the values calculated in the previous steps. $Q$ = $Q_1$ + $Q_2$ + $Q_3$ + $Q_4$ + $Q_5$") Q16 = widgets.HTMLMath(value="$Q$ = (4120 + 33400 + 41800 + 226000 + 20200) J = 325520 J or 326 kJ") if frame_1 == 0: display(SVG("Images/phase_diagram_1_0.svg")) frame_1 = 1 elif frame_1 == 1: display(SVG("Images/phase_diagram_1_1.svg")) display(I11, Q11) frame_1 = 2 elif frame_1 == 2: display(SVG("Images/phase_diagram_1_2.svg")) display(I11, Q11, I12, Q12) frame_1 = 3 elif frame_1 == 3: display(SVG("Images/phase_diagram_1_3.svg")) display(I11, Q11, I12, Q12, I13, Q13) frame_1 = 4 elif frame_1 == 4: display(SVG("Images/phase_diagram_1_4.svg")) display(I11, Q11, I12, Q12, I13, Q13, I14, Q14) frame_1 = 5 elif frame_1 == 5: display(SVG("Images/phase_diagram_1_5.svg")) display(I11, Q11, I12, Q12, I13, Q13, I14, Q14, I15, Q15) frame_1 = 6 elif frame_1 == 6: display(SVG("Images/phase_diagram_1_6.svg")) display(I11, Q11, I12, Q12, I13, Q13, I14, Q14, I15, Q15, I16, Q16) frame_1 = 0 button_phase_diagram_1 = widgets.Button(description="Show Next Step", button_style = 'primary') display(button_phase_diagram_1) def on_submit_button_phase_diagram_1_clicked(b): with out6: clear_output(wait=True) show_steps_1() with out6: display(SVG("Images/phase_diagram_1_0.svg")) button_phase_diagram_1.on_click(on_submit_button_phase_diagram_1_clicked) display(out6) ###Output _____no_output_____ ###Markdown **Note:** that the *state* of a material can include more than one *phase*. For example, at 0°C, the state of water includes both solid (ice) and liquid (water) phases. At 100°C, the state of water includes both liquid (water) and gas (steam) phases.It is common to cool down a material (as opposed to heating it up). In this scenario, heat must be taken away. By convention, a negative $Q$ is used to represent heat being taken away from a material (cooling), while a positive $Q$ is used to represent heat being added to a material (warming). Be aware of the sign of $Q$ as it indicates the direction the heat is flowing. For $Q=mH_f$ and $Q=mH_v$, you must be aware of whether heat is being added to or taken away from the material. If heat is being taken away, then a negative sign must be placed in front of $H_f$ and $H_v$. It is not necessary for each problem to be five steps. A problem could have 1-5 steps depending on the situation. Let's do another example together. An interactive graph is provided to help determine the number of steps required. ExampleHow much heat must be removed to change 10.0 g of steam at 120.0°C to water at 50°C? Round to two significant figures. ###Code #import ipywidgets as widgets #import matplotlib.pyplot as plt #from ipywidgets import HBox, Output, VBox #from IPython.display import clear_output out7 = Output() play2 = widgets.Play(interval=500, value=0, min=0, max=25, step=1, description="Press play", disabled=False) time_slider2 = widgets.IntSlider(description='Time', value=0, min=0, max=20, continuous_update = False) widgets.jslink((play2, 'value'), (time_slider2, 'value')) #Make lists of x and y values x_values2 = list(range(21)) y_values2 = [120, 110, 100, 100, 100, 100, 100, 100, 100, 100, 100, 95, 90, 85, 80, 75, 70, 65, 60, 55, 50] heat_y2 = [] increment2 = 0 for i in range(26): heat_y2.append(increment2) increment2 += 13021 #Plot graph def time_temp(change): x = change['new'] with out7: clear_output(wait=True) temp_x_values2 = [] temp_y_values2 = [] graph2y2 = [] for i in range(x+1): temp_x_values2.append(x_values2[i]) temp_y_values2.append(y_values2[i]) graph2y2.append(heat_y2[i]) plt.figure(figsize=(7,5)) plt.style.use('seaborn') plt.rcParams["axes.edgecolor"] = "black" plt.rcParams["axes.linewidth"] = 0.5 plt.ylim(0, 150) plt.xlim(-0.5,26) plt.xticks([]) plt.scatter(temp_x_values2, temp_y_values2) plt.ylabel('Temperature (°C)') plt.xlabel('Time') plt.figtext(0.5, 0.01, "This graph consists of three line-segments. This indicates that we require three steps.", wrap=True, horizontalalignment='center', fontsize=12) plt.show() #Get slider value time_temp({'new': time_slider2.value}) time_slider2.observe(time_temp, 'value') #Display widget display(HBox([play2, time_slider2])) display(out7) #import ipywidgets as widgets #from IPython.display import clear_output, SVG out8 = widgets.Output() frame_2 = 1 #Toggle images def show_steps_2(): global frame_2 I21 = widgets.HTMLMath(value="Step 1: Calculate the heat loss required to change steam from 120°C to 100°C. Since there is no phase change taking place, we use $Q = mCΔT$.") Q21 = widgets.HTMLMath(value="$Q_{1}$ = (0.01 kg) (2020 J/kg°C) (100°C - 120°C) = -404 J") I22 = widgets.HTMLMath(value="Step 2: Calculate the heat loss required to change steam at 100°C to water at 100°C. Since a phase change is taking place (condensation), we use $Q = -mH_{v}$.") Q22 = widgets.HTMLMath(value="$Q_{2}$ = - (0.01 kg) (2260000 J/kg) = -22600 J") I23 = widgets.HTMLMath(value="Step 3: Calculate the heat loss required to change water from 100°C to 50°C. Since there is no phase change taking place, we use $Q = mCΔT$.") Q23 = widgets.HTMLMath(value="$Q_{3}$ = (0.01 kg) (4180 J/kg°C) (50°C - 100°C) = -2090 J") I24 = widgets.HTMLMath(value="Summary: Calculate the total heat loss by adding up the values calculated in the previous steps. $Q$ = $Q_1$ + $Q_2$ + $Q_3$") Q24 = widgets.HTMLMath(value="$Q$ = (-404 + -22600 + -2090) J = -25000 J or -25 kJ") if frame_2 == 0: display(SVG("Images/phase_diagram_2_0.svg")) frame_2 = 1 elif frame_2 == 1: display(SVG("Images/phase_diagram_2_1.svg")) display(I21, Q21) frame_2 = 2 elif frame_2 == 2: display(SVG("Images/phase_diagram_2_2.svg")) display(I21, Q21, I22, Q22) frame_2 = 3 elif frame_2 == 3: display(SVG("Images/phase_diagram_2_3.svg")) display(I21, Q21, I22, Q22, I23, Q23) frame_2 = 4 elif frame_2 == 4: display(SVG("Images/phase_diagram_2_4.svg")) display(I21, Q21, I22, Q22, I23, Q23, I24, Q24) frame_2 = 0 button_phase_diagram_2 = widgets.Button(description="Show Next Step", button_style = 'primary') display(button_phase_diagram_2) def on_submit_button_phase_diagram_2_clicked(b): with out8: clear_output(wait=True) show_steps_2() with out8: display(SVG("Images/phase_diagram_2_0.svg")) button_phase_diagram_2.on_click(on_submit_button_phase_diagram_2_clicked) display(out8) ###Output _____no_output_____ ###Markdown PracticeThere are many variations that are possible with specific heat and latent heat questions. Use the *Generate New Question* button to generate additional practice problems. These practice problems will vary from one to five steps. **One Step Problem** ###Code #import math #import random #from IPython.display import Javascript, display #from ipywidgets import widgets, Layout def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "lead": 130, "silver": 235, "stainless Steal": 460, "tin": 217, "water": 4180, "zinc": 388} material = random.choice(list(materials.keys())) #Randomize variables mass = round(random.uniform(100.0, 1000.0), 1) temperature_initial, temperature_final = 0,0 variety1 = random.randint(1,5) if variety1 == 1: #Makes certain that initial and final temps are different while temperature_initial == temperature_final: temperature_initial = round(random.uniform(-50.0, 0.0), 1) temperature_final = round(random.uniform(-50.0, 0.0), 1) question = "How much heat is needed for a {} g block of ice at {}°C to change temperature to {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 2060 * (temperature_final - temperature_initial) elif variety1 == 2: while temperature_initial == temperature_final: temperature_initial = round(random.uniform(0.0, 100.0), 1) temperature_final = round(random.uniform(0.0, 100.0), 1) question = "How much heat is needed for {} g of water at {}°C to change temperature to {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 4180 * (temperature_final - temperature_initial) elif variety1 == 3: while temperature_initial == temperature_final: temperature_initial = round(random.uniform(100.0, 150.0), 1) temperature_final = round(random.uniform(100.0, 150.0), 1) question = "How much heat is needed for {} g of steam at {}°C to change temperature to {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 2020 * (temperature_final - temperature_initial) elif variety1 == 4: temperature_initial = 0 temperature_final = 0 direction_variety = random.randint(1,2) if direction_variety == 1: question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 334000 elif direction_variety == 2: question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * -334000 elif variety1 == 5: temperature_initial = 100 temperature_final = 100 direction_variety = random.randint(1,2) if direction_variety == 1: question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 2260000 elif direction_variety == 2: question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * -2260000 #Define range of values for random multiple choices mini = -1000 maxa = 1000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(int(round(answer/1000))) + " kJ" option_2 = str(round(choice_list[0],1)) + " kJ" option_3 = str(round(choice_list[1],1)) + " kJ" option_4 = str(round(-1*choice_list[2],1)) + " kJ" multiple_choice(option_1, option_2, option_3, option_4) ###Output _____no_output_____ ###Markdown **Two Step Problem** ###Code #import math #import random #from IPython.display import Javascript, display #from ipywidgets import widgets, Layout def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "lead": 130, "silver": 235, "stainless Steal": 460, "tin": 217, "water": 4180, "zinc": 388} material = random.choice(list(materials.keys())) #Randomize variables mass = round(random.uniform(100.0, 1000.0), 1) temperature_initial, temperature_final = 0,0 variety2 = random.randint(1,4) if variety2 == 1: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = 0 question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final - temperature_initial)) + ((mass/1000) * 334000) elif direction_variety == 2: temperature_initial = 0 temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final - temperature_initial)) + ((mass/1000) * -334000) elif variety2 == 2: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = 0 temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * 334000) elif direction_variety == 2: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = 0 question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * -334000) elif variety2 == 3: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = 100 question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = 100 temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * -2260000) elif variety2 == 4: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = 100 temperature_final = round(random.uniform(101.0, 150.0), 1) question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(temperature_final - temperature_initial)) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(101.0, 150.0), 1) temperature_final = 100 question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(temperature_final - temperature_initial)) + ((mass/1000) * -2260000) #Define range of values for random multiple choices mini = -1000 maxa = 1000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(int(round(answer/1000))) + " kJ" option_2 = str(round(choice_list[0],1)) + " kJ" option_3 = str(round(choice_list[1],1)) + " kJ" option_4 = str(round(-1*choice_list[2],1)) + " kJ" multiple_choice(option_1, option_2, option_3, option_4) ###Output _____no_output_____ ###Markdown **Three Step Problem** ###Code #import math #import random #from IPython.display import Javascript, display #from ipywidgets import widgets, Layout def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "lead": 130, "silver": 235, "stainless Steal": 460, "tin": 217, "water": 4180, "zinc": 388} material = random.choice(list(materials.keys())) #Randomize variables mass = round(random.uniform(100.0, 1000.0), 1) temperature_initial, temperature_final = 0,0 variety3 = random.randint(1,2) if variety3 == 1: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(0 - temperature_initial)) + ((mass/1000)*4180*(temperature_final - 0)) + ((mass/1000) * 334000) elif direction_variety == 2: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final-0)) + ((mass/1000)*4180*(0 - temperature_initial)) + ((mass/1000) * -334000) elif variety3 == 2: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = round(random.uniform(101.0, 150.0), 1) question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(100 - temperature_initial)) + ((mass/1000)*2020*(temperature_final - 100)) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(101.0, 150.0), 1) temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final-100)) + ((mass/1000)*2020*(100 - temperature_initial)) + ((mass/1000) * -2260000) #Define range of values for random multiple choices mini = -1000 maxa = 1000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(int(round(answer/1000))) + " kJ" option_2 = str(round(choice_list[0],1)) + " kJ" option_3 = str(round(choice_list[1],1)) + " kJ" option_4 = str(round(-1*choice_list[2],1)) + " kJ" multiple_choice(option_1, option_2, option_3, option_4) ###Output _____no_output_____ ###Markdown **Four Step Problem** ###Code #import math #import random #from IPython.display import Javascript, display #from ipywidgets import widgets, Layout def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "lead": 130, "silver": 235, "stainless Steal": 460, "tin": 217, "water": 4180, "zinc": 388} material = random.choice(list(materials.keys())) #Randomize variables mass = round(random.uniform(100.0, 1000.0), 1) temperature_initial, temperature_final = 0,0 variety4 = random.randint(1,2) if variety4 == 1: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = 100 question = "How much heat is needed to change {} g of ice at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(0 - temperature_initial)) + ((mass/1000)*4180*(temperature_final - 0)) + ((mass/1000) * 334000) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = 100 temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final-0)) + ((mass/1000)*4180*(0 - temperature_initial)) + ((mass/1000) * -334000) + ((mass/1000) * -2260000) elif variety4 == 2: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = 0 temperature_final = round(random.uniform(100.0, 150.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(temperature_final - 100)) + ((mass/1000)*4180*(100 - temperature_initial)) + ((mass/1000) * 334000) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(100.0, 150.0), 1) temperature_final = 0 question = "How much heat is needed to change {} g of steam at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(100-temperature_initial)) + ((mass/1000)*4180*(temperature_final-100)) + ((mass/1000) * -334000) + ((mass/1000) * -2260000) #Define range of values for random multiple choices mini = -1000 maxa = 1000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(int(round(answer/1000))) + " kJ" option_2 = str(round(choice_list[0],1)) + " kJ" option_3 = str(round(choice_list[1],1)) + " kJ" option_4 = str(round(-1*choice_list[2],1)) + " kJ" multiple_choice(option_1, option_2, option_3, option_4) ###Output _____no_output_____ ###Markdown **Five Step Problem** ###Code #import math #import random #from IPython.display import Javascript, display #from ipywidgets import widgets, Layout def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Dictionary of materials materials = {"aluminum": 903, "brass": 376, "carbon": 710, "copper": 385, "glass": 664, "iron": 450, "lead": 130, "silver": 235, "stainless Steal": 460, "tin": 217, "water": 4180, "zinc": 388} chosen_material = random.choice(list(materials.keys())) #Randomize variables mass = round(random.uniform(100.0, 1000.0), 1) temperature_initial, temperature_final = 0,0 direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = round(random.uniform(101.0, 150.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(0 - temperature_initial)) + ((mass/1000)*4180*(100 - 0)) + ((mass/1000)*2020*(temperature_final - 100)) + ((mass/1000) * 334000) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(101.0, 150.0), 1) temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(100 - temperature_initial)) + ((mass/1000)*4180*(0 - 100)) + ((mass/1000)*2060*(temperature_final - 0)) + ((mass/1000) * -334000) + ((mass/1000) * -2260000) #Define range of values for random multiple choices mini = -1000 maxa = 1000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(int(round(answer/1000))) + " kJ" option_2 = str(round(choice_list[0],1)) + " kJ" option_3 = str(round(choice_list[1],1)) + " kJ" option_4 = str(round(-1*choice_list[2],1)) + " kJ" multiple_choice(option_1, option_2, option_3, option_4) ###Output _____no_output_____ ###Markdown **Mixed Step Problems**In the dropdown menus below, select how many steps are required and select the correct amount of heat required for each question.**Hint:** Have some scrap-paper nearby for the calculations and be sure to sketch a diagram of each scenario to determine how many steps are required. ###Code #import math #import random #from IPython.display import Javascript, display #from ipywidgets import widgets, Layout def generate_new_question(ev): display(Javascript('IPython.notebook.execute_cell()')) button_generate_question = widgets.Button(description="Generate New Question", layout=Layout(width='20%', height='100%'), button_style='success') button_generate_question.on_click(generate_new_question) display(button_generate_question) #Randomize variables mass = round(random.uniform(100.0, 1000.0), 1) temperature_initial, temperature_final = 0,0 #Determine question type question_type = random.randint(1,5) if question_type == 1: #Type 1: One Step steps = "One Step" type1_variety = random.randint(1,5) if type1_variety == 1: #Makes certain that initial and final temps are different while temperature_initial == temperature_final: temperature_initial = round(random.uniform(-50.0, 0.0), 1) temperature_final = round(random.uniform(-50.0, 0.0), 1) question = "How much heat is needed for a {} g block of ice at {}°C to change temperature to {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 2060 * (temperature_final - temperature_initial) elif type1_variety == 2: while temperature_initial == temperature_final: temperature_initial = round(random.uniform(0.0, 100.0), 1) temperature_final = round(random.uniform(0.0, 100.0), 1) question = "How much heat is needed for {} g of water at {}°C to change temperature to {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 4180 * (temperature_final - temperature_initial) elif type1_variety == 3: while temperature_initial == temperature_final: temperature_initial = round(random.uniform(100.0, 150.0), 1) temperature_final = round(random.uniform(100.0, 150.0), 1) question = "How much heat is needed for {} g of steam at {}°C to change temperature to {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 2020 * (temperature_final - temperature_initial) elif type1_variety == 4: temperature_initial = 0 temperature_final = 0 direction_variety = random.randint(1,2) if direction_variety == 1: question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 334000 elif direction_variety == 2: question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * -334000 elif type1_variety == 5: temperature_initial = 100 temperature_final = 100 direction_variety = random.randint(1,2) if direction_variety == 1: question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * 2260000 elif direction_variety == 2: question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = (mass/1000) * -2260000 elif question_type == 2: #Type 2: Two Steps steps = "Two Steps" type2_variety = random.randint(1,4) if type2_variety == 1: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = 0 question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final - temperature_initial)) + ((mass/1000) * 334000) elif direction_variety == 2: temperature_initial = 0 temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final - temperature_initial)) + ((mass/1000) * -334000) elif type2_variety == 2: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = 0 temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * 334000) elif direction_variety == 2: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = 0 question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * -334000) elif type2_variety == 3: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = 100 question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = 100 temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final - temperature_initial)) + ((mass/1000) * -2260000) elif type2_variety == 4: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = 100 temperature_final = round(random.uniform(101.0, 150.0), 1) question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(temperature_final - temperature_initial)) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(101.0, 150.0), 1) temperature_final = 100 question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(temperature_final - temperature_initial)) + ((mass/1000) * -2260000) elif question_type == 3: #Type 3: Three Steps steps = "Three Steps" type3_variety = random.randint(1,2) if type3_variety == 1: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(0 - temperature_initial)) + ((mass/1000)*4180*(temperature_final - 0)) + ((mass/1000) * 334000) elif direction_variety == 2: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of water at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final-0)) + ((mass/1000)*4180*(0 - temperature_initial)) + ((mass/1000) * -334000) elif type3_variety == 2: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(1.0, 99.0), 1) temperature_final = round(random.uniform(101.0, 150.0), 1) question = "How much heat is needed to change {} g of water at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(100 - temperature_initial)) + ((mass/1000)*2020*(temperature_final - 100)) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(101.0, 150.0), 1) temperature_final = round(random.uniform(1.0, 99.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to water at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*4180*(temperature_final-100)) + ((mass/1000)*2020*(100 - temperature_initial)) + ((mass/1000) * -2260000) elif question_type == 4: #Type 4: Four Steps steps = "Four Steps" type4_variety = random.randint(1,2) if type4_variety == 1: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = 100 question = "How much heat is needed to change {} g of ice at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(0 - temperature_initial)) + ((mass/1000)*4180*(temperature_final - 0)) + ((mass/1000) * 334000) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = 100 temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(temperature_final-0)) + ((mass/1000)*4180*(0 - temperature_initial)) + ((mass/1000) * -334000) + ((mass/1000) * -2260000) elif type4_variety == 2: direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = 0 temperature_final = round(random.uniform(100.0, 150.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(temperature_final - 100)) + ((mass/1000)*4180*(100 - temperature_initial)) + ((mass/1000) * 334000) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(100.0, 150.0), 1) temperature_final = 0 question = "How much heat is needed to change {} g of steam at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(100-temperature_initial)) + ((mass/1000)*4180*(temperature_final-100)) + ((mass/1000) * -334000) + ((mass/1000) * -2260000) elif question_type == 5: #Type 5: Five Steps steps = "Five Steps" direction_variety = random.randint(1,2) if direction_variety == 1: temperature_initial = round(random.uniform(-50.0, -1.0), 1) temperature_final = round(random.uniform(101.0, 150.0), 1) question = "How much heat is needed to change {} g of ice at {}°C to steam at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2060*(0 - temperature_initial)) + ((mass/1000)*4180*(100 - 0)) + ((mass/1000)*2020*(temperature_final - 100)) + ((mass/1000) * 334000) + ((mass/1000) * 2260000) elif direction_variety == 2: temperature_initial = round(random.uniform(101.0, 150.0), 1) temperature_final = round(random.uniform(-50.0, -1.0), 1) question = "How much heat is needed to change {} g of steam at {}°C to ice at {}°C?".format(mass, temperature_initial, temperature_final) print(question) answer = ((mass/1000)*2020*(100 - temperature_initial)) + ((mass/1000)*4180*(0 - 100)) + ((mass/1000)*2060*(temperature_final - 0)) + ((mass/1000) * -334000) + ((mass/1000) * -2260000) #Define range of values for random multiple choices mini = -1000 maxa = 1000 #Create three choices that are unique (and not equal to the answer) choice_list = random.sample(range(mini,maxa),3) while choice_list.count(int(answer)) >= 1: choice_list = random.sample(range(mini,maxa),3) #Assign each multiple choice to these four variables #Option_1 contains the answer option_1 = str(int(round(answer/1000))) + " kJ" option_2 = str(round(choice_list[0],1)) + " kJ" option_3 = str(round(choice_list[1],1)) + " kJ" option_4 = str(round(-1*choice_list[2],1)) + " kJ" option_list = [option_1, option_2, option_3, option_4] correct_answer = option_list[0] #Randomly shuffle the options random.shuffle(option_list) #Create dropdown menus dropdown1_1 = widgets.Dropdown(options={' ':0,'One Step': 1, 'Two Steps': 2, 'Three Steps': 3, 'Four Steps': 4, 'Five Steps': 5}, value=0, description='Steps',) dropdown1_2 = widgets.Dropdown(options={' ':0,option_list[0]: 1, option_list[1]: 2, option_list[2]: 3, option_list[3]: 4}, value=0, description='Answer',) #Display menus as 1x2 table container1_1 = widgets.HBox(children=[dropdown1_1, dropdown1_2]) display(container1_1), print(" ", end='\r') #Evaluate input def check_answer_dropdown(b): answer1_1 = dropdown1_1.label answer1_2 = dropdown1_2.label if answer1_1 == steps and answer1_2 == correct_answer: print("Correct! ", end='\r') elif answer1_1 != ' ' and answer1_2 != ' ': print("Try again.", end='\r') else: print(" ", end='\r') dropdown1_1.observe(check_answer_dropdown, names='value') dropdown1_2.observe(check_answer_dropdown, names='value') ###Output _____no_output_____ ###Markdown Conclusions* The **specific heat capacity** of a material is an empirically determined value characteristic of a particular material. It is defined as the amount of heat needed to raise the temperature of 1 kg of the material by 1°C.* We use the formula $Q=mc\Delta T$ to calculate the amount of heat required to change the temperature of a material in which there is no change of phase.* The **latent heat of fusion** ($H_f$) is defined as the amount of heat needed to melt 1 kg of a solid without a change in temperature.* The **latent heat of vaporization** ($H_v$) is define as the amount of heat needed to vaporise 1 kg of a liquid without a change in temperature.* We use the formula $Q=mH_f$ to calculate the heat required to change a material from a solid to a liquid, or from a liquid to a solid.* We use the formula $Q=mH_v$ to calculate the heat required to change a material from a liquid to a gas, or from a gas to a liquid.* If heat is being taken away, then a negative sign must be placed in front of $H_f$ and $H_v$.* We use a combination of the above formulae to compute the heat required to change a material from an initial temperature to a final temperature when one (or more) phase changes occur across a range of temperatures.Images in this notebook represent original artwork. ###Code %%html <script> function code_toggle() { if (code_shown){ $('div.input').hide('500'); $('#toggleButton').val('Show Code') } else { $('div.input').show('500'); $('#toggleButton').val('Hide Code') } code_shown = !code_shown } $( document ).ready(function(){ code_shown=false; $('div.input').hide() }); </script> <form action="javascript:code_toggle()"><input type="submit" id="toggleButton" value="Show Code"></form> ###Output _____no_output_____
2017/tutorials/Tutorial2-Exercises.ipynb
###Markdown CS375 - Tutorial 2 Welcome to tutorial 2! This tutorial will introduce you to unsupervised learning methods, specifically how to train a sparse autoencoder and variational autoencoder on MNIST, and how to evaluate the trained model on neural data. As before everything will be implemented using TFUtils. We will start with a sparse autoencoder and then move on to variational autoencoders. 1.) Training and evaluating a sparse autoencoder on MNIST 1.1.) Define a simple sparse autoencoder consisting of one fully connected layer in the encoder and one fully connected layer in the decoder. The input dimension is 784. Use xavier initialization and l2-regularization on the weights, initialize all biases to 0 and use a tanh activation function. Regularize the hidden layer activations with a l1-regularization: ###Code %matplotlib inline from __future__ import division from tfutils import base, data, optimizer, utils import numpy as np import tensorflow as tf import os import pymongo as pm import matplotlib.pyplot as plt import scipy.signal as signal # connect to database dbname = 'mnist' collname = 'autoencoder' port = 24444 conn = pm.MongoClient(port = port) coll = conn[dbname][collname + '.files'] def sparse_autoencoder(inputs, train=True, beta = 5e-4, n_hidden = 100, **kwargs): ''' Implements a simple autoencoder consisting of two fully connected layers ''' # flatten the input images inp = tf.reshape(inputs['images'], [inputs['images'].get_shape().as_list()[0], -1]) ### YOUR CODE HERE return output, {} ###Output _____no_output_____ ###Markdown 1.2.) Define the l2 loss function for the sparse autoencoder: ###Code def sparse_autoencoder_loss(inputs, outputs, **kwargs): ''' Defines the loss = l2(inputs - outputs) + l1(weights) ''' # flatten the input images inputs = tf.reshape(inputs, [inputs.get_shape().as_list()[0], -1]) ### YOUR CODE HERE return loss def sparse_autoencoder_validation(inputs, outputs, **kwargs): ''' Wrapper for using the loss function as a validation target ''' return {'l2_loss': sparse_autoencoder_loss(inputs['images'], outputs), 'pred': outputs, 'gt': inputs['images']} ###Output _____no_output_____ ###Markdown Now let's define and run our sparse autoencoder experiment on MNIST in TFUtils. We will use the Adam optimizer and an exponentially decaying learning rate: ###Code def online_agg_mean(agg_res, res, step): """ Appends the mean value for each key """ if agg_res is None: agg_res = {k: [] for k in res} for k, v in res.items(): if k in ['pred', 'gt']: value = v else: value = np.mean(v) agg_res[k].append(value) return agg_res def agg_mean(results): for k in results: if k in ['pred', 'gt']: results[k] = results[k][0] elif k is 'l2_loss': results[k] = np.mean(results[k]) else: raise KeyError('Unknown target') return results # number of hidden neurons n_hidden = 100 # scaling of l1 regularization beta = 1e-2 # 1e-4 = no regularization params = {} params['load_params'] = { 'do_restore': False, } params['save_params'] = { 'host': 'localhost', 'port': 24444, 'dbname': 'mnist', 'collname': 'autoencoder', 'exp_id': 'exp1', 'save_metrics_freq': 200, 'save_valid_freq': 200, 'save_filters_freq': 1000, 'cache_filters_freq': 1000, } params['train_params'] = { 'validate_first': False, 'data_params': {'func': data.MNIST, 'batch_size': 256, 'group': 'train', 'n_threads': 1}, 'queue_params': {'queue_type': 'random', 'batch_size': 256}, 'num_steps': 4000, 'thres_loss': float("inf"), } params['validation_params'] = {'valid0': { 'data_params': {'func': data.MNIST, 'batch_size': 100, 'group': 'test', 'n_threads': 1}, 'queue_params': {'queue_type': 'fifo', 'batch_size': 100}, 'num_steps': 100, 'targets': {'func': sparse_autoencoder_validation}, 'online_agg_func': online_agg_mean, 'agg_func': agg_mean, }} params['model_params'] = { 'func': sparse_autoencoder, 'beta': beta, 'n_hidden': n_hidden, } params['learning_rate_params'] = { 'learning_rate': 5e-3, 'decay_steps': 2000, 'decay_rate': 0.95, 'staircase': True, } params['optimizer_params'] = { 'func': optimizer.ClipOptimizer, 'optimizer_class': tf.train.AdamOptimizer, 'clip': False, } params['loss_params'] = { 'targets': ['images'], 'loss_per_case_func': sparse_autoencoder_loss, 'loss_per_case_func_params' : {'_outputs': 'outputs', '_targets_$all': 'inputs'}, 'agg_func': tf.reduce_mean, } params['skip_check'] = True coll.remove({'exp_id' : 'exp1'}, {'justOne': True}) base.train_from_params(**params) ###Output _____no_output_____ ###Markdown Now let's have a look at our training and validation curves that were stored in our database: ###Code def get_losses(exp_id): """ Gets all loss entries from the database and concatenates them into a vector """ q_train = {'exp_id' : exp_id, 'train_results' : {'$exists' : True}} return np.array([_r['loss'] for r in coll.find(q_train, projection = ['train_results']) for _r in r['train_results']]) def plot_train_loss(exp_id, start_step=None, end_step=None, N_smooth = 100, plt_title = None): """ Plots the training loss You will need to EDIT this part. """ # get the losses from the database loss = get_losses(exp_id) if start_step is None: start_step = 0 if end_step is None: end_step = len(loss) if plt_title is None: plt_title = exp_id # Only plot selected loss window loss = loss[start_step:end_step] # plot loss fig = plt.figure(figsize = (15, 6)) plt.plot(loss) plt.title(plt_title + ' training: loss') plt.grid() axes = plt.gca() # plot smoothed loss smoothed_loss = signal.convolve(loss, np.ones((N_smooth,)))[N_smooth : -N_smooth] / float(N_smooth) plt.figure(figsize = (15, 6)) plt.plot(smoothed_loss) plt.title(plt_title + ' training: loss smoothed') plt.grid() axes = plt.gca() plot_train_loss('exp1') def get_validation_data(exp_id): """ Gets the validation data from the database (except for gridfs data) """ q_val = {'exp_id' : exp_id, 'validation_results' : {'$exists' : True}, 'validates' : {'$exists' : False}} val_steps = coll.find(q_val, projection = ['validation_results']) return [val_steps[i]['validation_results']['valid0']['l2_loss'] for i in range(val_steps.count())] def plot_validation_results(exp_id, plt_title = None): """ Plots the validation results i.e. the top1 and top5 accuracy You will need to EDIT this part. """ # get the data from the database l2_loss = get_validation_data(exp_id) if plt_title is None: plt_title = exp_id # plot top1 accuracy fig = plt.figure(figsize = (15, 6)) plt.plot(l2_loss) plt.title(plt_title + ' validation: l2 loss') plt.grid() axes = plt.gca() plot_validation_results('exp1') ###Output _____no_output_____ ###Markdown Finally let's plot some example outputs of our autoencoder. The images to the left are the outputs. The images to the right are the inputs. ###Code def get_validation_images(exp_id): """ Gets the validation images from the database """ q_val = {'exp_id' : exp_id, 'validation_results' : {'$exists' : True}, 'validates' : {'$exists' : False}} val_steps = coll.find(q_val, projection = ['validation_results']) pred = np.array([val_steps[i]['validation_results']['valid0']['pred'] for i in range(val_steps.count())]) gt = np.array([val_steps[i]['validation_results']['valid0']['gt'] for i in range(val_steps.count())]) return {'gt': gt, 'pred': pred} def plot_validation_images(exp_id, n_images = 24): ''' Plots n_images images in a grid. The ground truth image is on the left and the prediction is on the right. ''' imgs = get_validation_images(exp_id) fig = plt.figure(figsize=(16, 16)) for i in range(n_images): pred = np.reshape(imgs['pred'][0,i], [28, 28]) plt.subplot(n_images/4,n_images/3,1 + i*2) plt.imshow(pred, cmap='gray') ax = plt.gca() ax.xaxis.set_visible(False) ax.yaxis.set_visible(False) gt = np.reshape(imgs['gt'][0,i], [28, 28]) plt.subplot(n_images/4,n_images/3,2 + i*2) plt.imshow(gt, cmap='gray') ax = plt.gca() ax.xaxis.set_visible(False) ax.yaxis.set_visible(False) plot_validation_images('exp1') ###Output _____no_output_____ ###Markdown 2.) Training and evaluating a variational autoencoder on MNIST 2.1.) Define a simple variational autoencoder as discussed in class consisting of - one fully connected layer (dimension n_hidden) in the encoder that results in the latent variable, - two fully connected layers (dimension n_latent) that take the latent variable and reparametrize the latent distribution with it's mean mu and log of the standard deviation, and - two fully connected layers (dimension n_hidden & output_dim) in the decoder that take the reparametrized latent variable and decode it into the output prediction. The input dimension is 784. Use tanh activation functions in the intermediate layers and a sigmoid at the top layer. Use xavier initialization for the weights and constant initialization to 0 for the biases: ###Code def variational_autoencoder(inputs, train=True, n_hidden = 100, n_latent = 20, **kwargs): ''' Implements a simple autoencoder consisting of two fully connected layers ''' outputs = {} # flatten the input images inp = tf.reshape(inputs['images'], [inputs['images'].get_shape().as_list()[0], -1]) ### YOUR CODE HERE return outputs, {} ###Output _____no_output_____ ###Markdown 2.2.) Define the loss for our variational autoencoder as discussed in class: ###Code def variational_autoencoder_loss(inputs, outputs, **kwargs): ''' Defines the loss = l2(inputs - outputs) + l2(weights) ''' # flatten the input images inputs = tf.reshape(inputs, [inputs.get_shape().as_list()[0], -1]) ### YOUR CODE HERE return loss def variational_autoencoder_validation(inputs, outputs, **kwargs): ''' Wrapper for using the loss function as a validation target ''' return {'l2_loss': variational_autoencoder_loss(inputs['images'], outputs), 'pred': outputs['pred'], 'gt': inputs['images']} def online_agg_mean(agg_res, res, step): """ Appends the mean value for each key """ if agg_res is None: agg_res = {k: [] for k in res} for k, v in res.items(): if k in ['pred', 'gt']: value = v else: value = np.mean(v) agg_res[k].append(value) return agg_res def agg_mean(results): for k in results: if k in ['pred', 'gt']: results[k] = results[k][0] elif k is 'l2_loss': results[k] = np.mean(results[k]) else: raise KeyError('Unknown target') return results # number of hidden neurons n_hidden = 100 # dimension of latent space n_latent = 20 params = {} params['load_params'] = { 'do_restore': False, } params['save_params'] = { 'host': 'localhost', 'port': 24444, 'dbname': 'mnist', 'collname': 'autoencoder', 'exp_id': 'exp2', 'save_metrics_freq': 200, 'save_valid_freq': 200, 'save_filters_freq': 1000, 'cache_filters_freq': 1000, } params['train_params'] = { 'validate_first': False, 'data_params': {'func': data.MNIST, 'batch_size': 256, 'group': 'train', 'n_threads': 1}, 'queue_params': {'queue_type': 'random', 'batch_size': 256}, 'num_steps': 10000, 'thres_loss': float("inf"), } params['validation_params'] = {'valid0': { 'data_params': {'func': data.MNIST, 'batch_size': 100, 'group': 'test', 'n_threads': 1}, 'queue_params': {'queue_type': 'fifo', 'batch_size': 100}, 'num_steps': 100, 'targets': {'func': variational_autoencoder_validation}, 'online_agg_func': online_agg_mean, 'agg_func': agg_mean, }} params['model_params'] = { 'func': variational_autoencoder, 'n_latent': n_latent, 'n_hidden': n_hidden, } params['learning_rate_params'] = { 'learning_rate': 5e-3, 'decay_steps': 10000, 'decay_rate': 0.95, 'staircase': True, } params['optimizer_params'] = { 'func': optimizer.ClipOptimizer, 'optimizer_class': tf.train.AdamOptimizer, 'clip': True, } params['loss_params'] = { 'targets': ['images'], 'loss_per_case_func': variational_autoencoder_loss, 'loss_per_case_func_params' : {'_outputs': 'outputs', '_targets_$all': 'inputs'}, 'agg_func': tf.reduce_mean, } params['skip_check'] = True coll.remove({'exp_id' : 'exp2'}, {'justOne': True}) base.train_from_params(**params) ###Output _____no_output_____ ###Markdown Now let's have a look at our training and validation curves that were stored in our database: ###Code def get_losses(exp_id): """ Gets all loss entries from the database and concatenates them into a vector """ q_train = {'exp_id' : exp_id, 'train_results' : {'$exists' : True}} return np.array([_r['loss'] for r in coll.find(q_train, projection = ['train_results']) for _r in r['train_results']]) def plot_train_loss(exp_id, start_step=None, end_step=None, N_smooth = 100, plt_title = None): """ Plots the training loss You will need to EDIT this part. """ # get the losses from the database loss = get_losses(exp_id) if start_step is None: start_step = 0 if end_step is None: end_step = len(loss) if plt_title is None: plt_title = exp_id # Only plot selected loss window loss = loss[start_step:end_step] # plot loss fig = plt.figure(figsize = (15, 6)) plt.plot(loss) plt.title(plt_title + ' training: loss') plt.grid() axes = plt.gca() # plot smoothed loss smoothed_loss = signal.convolve(loss, np.ones((N_smooth,)))[N_smooth : -N_smooth] / float(N_smooth) plt.figure(figsize = (15, 6)) plt.plot(smoothed_loss) plt.title(plt_title + ' training: loss smoothed') plt.grid() axes = plt.gca() plot_train_loss('exp2') def get_validation_data(exp_id): """ Gets the validation data from the database (except for gridfs data) """ q_val = {'exp_id' : exp_id, 'validation_results' : {'$exists' : True}, 'validates' : {'$exists' : False}} val_steps = coll.find(q_val, projection = ['validation_results']) return [val_steps[i]['validation_results']['valid0']['l2_loss'] for i in range(val_steps.count())] def plot_validation_results(exp_id, plt_title = None): """ Plots the validation results i.e. the top1 and top5 accuracy You will need to EDIT this part. """ # get the data from the database l2_loss = get_validation_data(exp_id) if plt_title is None: plt_title = exp_id # plot top1 accuracy fig = plt.figure(figsize = (15, 6)) plt.plot(l2_loss) plt.title(plt_title + ' validation: l2 loss') plt.grid() axes = plt.gca() plot_validation_results('exp2') ###Output _____no_output_____ ###Markdown Finally let's plot some example outputs of our autoencoder. The images to the left are the outputs. The images to the right are the inputs. ###Code def get_validation_images(exp_id): """ Gets the validation images from the database """ q_val = {'exp_id' : exp_id, 'validation_results' : {'$exists' : True}, 'validates' : {'$exists' : False}} val_steps = coll.find(q_val, projection = ['validation_results']) pred = np.array([val_steps[i]['validation_results']['valid0']['pred'] for i in range(val_steps.count())]) gt = np.array([val_steps[i]['validation_results']['valid0']['gt'] for i in range(val_steps.count())]) return {'gt': gt, 'pred': pred} def plot_validation_images(exp_id, n_images = 24): ''' Plots n_images images in a grid. The ground truth image is on the left and the prediction is on the right. ''' imgs = get_validation_images(exp_id) fig = plt.figure(figsize=(16, 16)) for i in range(n_images): pred = np.reshape(imgs['pred'][0,i], [28, 28]) plt.subplot(n_images/4,n_images/3,1 + i*2) plt.imshow(pred, cmap='gray') ax = plt.gca() ax.xaxis.set_visible(False) ax.yaxis.set_visible(False) gt = np.reshape(imgs['gt'][0,i], [28, 28]) plt.subplot(n_images/4,n_images/3,2 + i*2) plt.imshow(gt, cmap='gray') ax = plt.gca() ax.xaxis.set_visible(False) ax.yaxis.set_visible(False) plot_validation_images('exp2') ###Output _____no_output_____
Cardio Catch Disease - Machine Learning.ipynb
###Markdown Cardiovascular Disease Escrevendo algoritmos Classificadores Encontrando a acurácia e a Precisão da ferramenta ###Code from matplotlib import pyplot as plt import pandas as pd import seaborn as sns #Abrindo os dados limpos da seção anterior df = pd.read_csv('cardio_data.csv') df.set_index('id', inplace=True) df.head() ###Output _____no_output_____ ###Markdown Preparing the Training and Test set. Separa os dados em treino e teste Separa aleatoriamente em 70% dos dados para treino e 30% dos dados para teste ###Code X = df.drop(columns = ['cardio']) y = df['cardio'] from sklearn.model_selection import train_test_split X.head() #X_train = recebe os dados de treino #X_test = recebe os dados de teste (30%) #y_train = classes associadas aos dados de treino #Y-test = classes associadas aos dados de teste X_train, X_test, y_train, y_test = train_test_split(X, y, test_size = 0.30, random_state = 9) X_train.shape, y_train.shape X_test.shape, y_test.shape ###Output _____no_output_____ ###Markdown Criando modelos ###Code from sklearn.tree import DecisionTreeClassifier from sklearn.ensemble import RandomForestClassifier from sklearn.neighbors import KNeighborsClassifier from sklearn.naive_bayes import GaussianNB dtc = DecisionTreeClassifier() ran = RandomForestClassifier(n_estimators=90) knn = KNeighborsClassifier(n_neighbors=79) naive = GaussianNB() models = {"Decision tree" : dtc, "Random forest" : ran, "KNN" : knn, "Naive bayes" : naive } scores= { } ###Output _____no_output_____ ###Markdown Treina o algoritmo e gera o modelo ###Code for key, value in models.items(): model = value model.fit(X_train, y_train) scores[key] = model.score(X_test, y_test) ###Output _____no_output_____ ###Markdown Performance Metrics ###Code scores_frame = pd.DataFrame(scores, index=["Accuracy Score"]).T scores_frame.sort_values(by=["Accuracy Score"], axis=0 ,ascending=False, inplace=True) scores_frame plt.figure(figsize=(5,5)) sns.barplot(x=scores_frame.index,y=scores_frame["Accuracy Score"]) plt.xticks(rotation=45); ###Output _____no_output_____ ###Markdown Prevendo resultados com o melhor modelo ###Code y_test[400:403] X_test[400:403] previsoes = knn.predict(X_test[400:403]) previsoes ###Output _____no_output_____ ###Markdown Fazendo Cross-Validation Cross-Validation com 5 FoldsFAZER CROSS VALIDATION APÓS O TRAIN TEST SPLIT EM CIMA DO TREINO* https://scikit-learn.org/stable/modules/cross_validation.html ###Code from sklearn.model_selection import cross_val_score scores_dtc = cross_val_score(dtc, X, y, cv=5, scoring='accuracy') scores_ran = cross_val_score(ran, X, y, cv=5, scoring='accuracy') scores_knn = cross_val_score(knn, X, y, cv=5, scoring='accuracy') scores_naive = cross_val_score(naive, X, y, cv=5, scoring='accuracy') print('Decision Tree:', scores_dtc.mean()) print('Random Forest:', scores_ran.mean()) print('KNeighbors:', scores_knn.mean()) print('Naive Bayes:', scores_naive.mean()) ###Output Decision Tree: 0.6079662447257383 Random Forest: 0.686464135021097 KNeighbors: 0.7156793248945148 Naive Bayes: 0.7072911392405064
qick_demos/00_Send_receive_pulse_sim.ipynb
###Markdown Sending and receiving a pulse demonstrationNote: this notework is a copy of 00_Send_receive_pulse. It shows how to use the simulator for the QickPrograms in that notebook. In this demo you will send and receive a pulse in loopback to demonstrate control over the QICK. By modifying the config Python dictionary in the below notebook cell, you can change several variables:* The pulse length length in FPGA clock ticks (1 clock tick = 2.6 ns).* The readout buffer length readout_length in FPGA clock ticks.* The pulse envelope shape pulse_style (either const or flat_top or arb )* The pulse amplitude pulse_gain in DAC units.* The pulse frequency pulse_freq in MHz.* The readout "time of flight" adc_trig_offset in FPGA clock ticks.* The number of times you average the read soft_avgs ###Code # Import the QICK drivers and auxiliary libraries from qick import * from qick.helpers import gauss import time %matplotlib inline from qick.interpreter import simulate from qick.interpreter import save_results from qick.interpreter import read_results # Load bitstream with custom overlay soc = QickSoc() # Set the loopback DAC channel to be in 1st Nyquist zone mode soc.set_nyquist(ch=7,nqz=1); ###Output /home/jimk/.local/lib/python3.8/site-packages/pynq/pl_server/device.py:79: UserWarning: No devices found, is the XRT environment sourced? warnings.warn( ###Markdown Hardware ConfigurationtProc channel 7 : DAC 229 CH3 Readout channel 0 : ADC 224 CH0 ###Code class LoopbackProgram(AveragerProgram): def __init__(self,cfg): AveragerProgram.__init__(self,cfg) def initialize(self): cfg=self.cfg r_freq=self.sreg(cfg["res_ch"], "freq") #Get frequency register for res_ch self.cfg["adc_lengths"]=[self.cfg["readout_length"]]*2 #add length of adc acquisition to config self.cfg["adc_freqs"]=[adcfreq(self.cfg["pulse_freq"])]*2 #add frequency of adc ddc to config if self.cfg["pulse_style"] == "const": self.add_pulse(ch=self.cfg["res_ch"], name="measure", style=self.cfg["pulse_style"], length=self.cfg["length"]) #add a constant pulse to the pulse library if self.cfg["pulse_style"] == "flat_top": self.add_pulse(ch=self.cfg["res_ch"], name="measure", style=self.cfg["pulse_style"], length=self.cfg["length"], idata = self.cfg["idata"]) if self.cfg["pulse_style"] == "arb": self.add_pulse(ch=self.cfg["res_ch"], name="measure", style=self.cfg["pulse_style"], idata = self.cfg["idata"]) freq=freq2reg(adcfreq(cfg["pulse_freq"])) # convert frequency to dac frequency (ensuring it is an available adc frequency) self.pulse(ch=cfg["res_ch"], name="measure", freq=freq, phase=0, gain=cfg["pulse_gain"], t= 0, play=False) # pre-configure readout pulse self.synci(200) # give processor some time to configure pulses def body(self): self.trigger_adc(adc1=1, adc2=1,adc_trig_offset=self.cfg["adc_trig_offset"]) # trigger the adc acquisition if self.cfg["pulse_style"] == "const": self.pulse(ch=self.cfg["res_ch"], length=self.cfg["length"], play=True) # play readout pulse if self.cfg["pulse_style"] == "flat_top": self.pulse(ch=self.cfg["res_ch"], name="measure", play=True) # play readout pulse if self.cfg["pulse_style"] == "arb": self.pulse(ch=self.cfg["res_ch"], play=True) # play readout pulse self.sync_all(us2cycles(self.cfg["relax_delay"])) # sync all channels ###Output _____no_output_____ ###Markdown Send/receive a pulse with pulse_style = const ###Code config={"res_ch":7, # --Fixed "reps":1, # --Fixed "relax_delay":0, # --Fixed "res_phase":0, # --Fixed "pulse_style": "const", # --Fixed "length":20, # [Clock ticks] # Try varying length from 10-100 clock ticks "readout_length":200, # [Clock ticks] # Try varying readout_length from 50-1000 clock ticks "pulse_gain":3000, # [DAC units] # Try varying pulse_gain from 500 to 30000 DAC units "pulse_freq": 100, # [MHz] # In this program the signal is up and downconverted digitally so you won't see any frequency # components in the I/Q traces below. But since the signal gain depends on frequency, # if you lower pulse_freq you will see an increased gain. "adc_trig_offset": 100, # [Clock ticks] # Try varying adc_trig_offset from 100 to 220 clock ticks "soft_avgs":100 # Try varying soft_avgs from 1 to 200 averages } ################### # Try it yourself ! ################### prog1 =LoopbackProgram(config) #prog1.acquire_decimated(soc, load_pulses=True, progress=True, debug=False) results = simulate(prog1) print(f'1: pulses={results["pulses"]}') print(f'1: log={results["instruction_log"]}') print(f'1: mem changes={results["mem_changes"]}') ###Output using qick program version 1: pulses=[[7, 'const', 0, 20]] 1: log=[(0, 0, 3, 0, 'regwi', {'va': 69905067, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 16, 'rb': 0, 'rc': 0, 'imm': 69905067}), (2, 1, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 17, 'rb': 0, 'rc': 0, 'imm': 0}), (4, 2, 3, 0, 'regwi', {'va': 3000, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 19, 'rb': 0, 'rc': 0, 'imm': 3000}), (6, 3, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}), (8, 4, 3, 0, 'regwi', {'va': 589844, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 589844}), (10, 5, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 200}), (12, 6, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 0}), (14, 7, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 14, 'rb': 0, 'rc': 0, 'imm': 0}), (16, 8, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}), (18, 9, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 100}), (20, 10, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}), (22, 11, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 110}), (24, 12, 3, 0, 'regwi', {'va': 589844, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 589844}), (26, 13, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}), (28, 14, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 3000, 've': 589844, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (30, 15, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 20}), (32, 16, 0, 0, 'mathi', {'vb': 0, 'va': 1, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}), (34, 17, 0, 0, 'memwi', {'va': 1, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}), (36, 18, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 8}), (38, 19, 0, -1, 'end', {'page': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'addr': 0}), (200, 20, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 3000, 've': 589844, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (300, 20, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 100}), (310, 20, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 110})] 1: mem changes=[(34, 17, 0, 1, 1)] ###Markdown Send/receive a pulse with pulse_style = flat_top ###Code config={"res_ch":7, # --Fixed "reps":1, # --Fixed "relax_delay":0, # --Fixed "res_phase":0, # --Fixed "pulse_style": "flat_top", # --Fixed "length": 50, # [Clock ticks] # Try varying length from 10-100 clock ticks "sigma": 30, # [Clock ticks] # Try varying sigma from 10-50 clock ticks "readout_length":200, # [Clock ticks] # Try varying readout_length from 50-1000 clock ticks "pulse_gain":5000, # [DAC units] # Try varying pulse_gain from 500 to 30000 DAC units "pulse_freq": 100, # [MHz] # In this program the signal is up and downconverted digitally so you won't see any frequency # components in the I/Q traces below. But since the signal gain depends on frequency, # if you lower pulse_freq you will see an increased gain. "adc_trig_offset": 200, # [Clock ticks] # Try varying adc_trig_offset from 100 to 220 clock ticks "soft_avgs":100 # Try varying soft_avgs from 1 to 200 averages } config["idata"] = gauss(mu=config["sigma"]*16*5/2,si=config["sigma"]*16,length=5*config["sigma"]*16,maxv=32000) # Try varying idata to be an arbitrary numpy array of your choosing! # The first half of idata ramps up the flat_top pulse, the second half ramps down the flat_top pulse ################### # Try it yourself ! ################### prog2 =LoopbackProgram(config) #prog.acquire_decimated(soc, load_pulses=True, progress=True, debug=False) results = simulate(prog2) print(f'2: pulses={results["pulses"]}') print(f'2: log={results["instruction_log"]}') print(f'2: mem changes={results["mem_changes"]}') ###Output using qick program version 2: pulses=[[7, 'flat_top', 0, 2400, array([61, 62, 63, ..., 63, 63, 62], dtype=int16), array([0, 0, 0, ..., 0, 0, 0], dtype=int16)]] 2: log=[(0, 0, 3, 0, 'regwi', {'va': 69905067, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 16, 'rb': 0, 'rc': 0, 'imm': 69905067}), (2, 1, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 17, 'rb': 0, 'rc': 0, 'imm': 0}), (4, 2, 3, 0, 'regwi', {'va': 5000, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 19, 'rb': 0, 'rc': 0, 'imm': 5000}), (6, 3, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}), (8, 4, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}), (10, 5, 3, 0, 'regwi', {'va': 524363, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524363}), (12, 6, 3, 0, 'regwi', {'va': 50, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 22, 'rb': 0, 'rc': 0, 'imm': 50}), (14, 7, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 200}), (16, 8, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 0}), (18, 9, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 14, 'rb': 0, 'rc': 0, 'imm': 0}), (20, 10, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}), (22, 11, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}), (24, 12, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}), (26, 13, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}), (28, 14, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}), (30, 15, 3, 0, 'regwi', {'va': 524363, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524363}), (32, 16, 3, 0, 'regwi', {'va': 50, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 22, 'rb': 0, 'rc': 0, 'imm': 50}), (34, 17, 3, 0, 'regwi', {'va': 5000, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 19, 'rb': 0, 'rc': 0, 'imm': 5000}), (36, 18, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}), (38, 19, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}), (40, 20, 3, 0, 'regwi', {'va': 524363, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524363}), (42, 21, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524363, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (44, 22, 3, 0, 'regwi', {'va': 2500, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 19, 'rb': 0, 'rc': 0, 'imm': 2500}), (46, 23, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}), (48, 24, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}), (50, 25, 3, 0, 'regwi', {'va': 589824, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 589824}), (52, 26, 3, 0, 'math', {'va': 589874, 'vb': 589824, 'vc': 50, 'page': 3, 'ch': 0, 'oper': 8, 'ra': 20, 'rb': 20, 'rc': 22, 'rd': 0, 're': 0, 'rf': 0, 'rg': 0, 'rh': 0}), (54, 27, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 2500, 've': 589874, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (56, 28, 3, 0, 'regwi', {'va': 5000, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 19, 'rb': 0, 'rc': 0, 'imm': 5000}), (58, 29, 3, 0, 'regwi', {'va': 125, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 125}), (60, 30, 3, 0, 'regwi', {'va': 75, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 75}), (62, 31, 3, 0, 'regwi', {'va': 524363, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524363}), (64, 32, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 75, 'vd': 5000, 've': 524363, 'vt': 125, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (66, 33, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 200}), (68, 34, 0, 0, 'mathi', {'vb': 0, 'va': 1, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}), (70, 35, 0, 0, 'memwi', {'va': 1, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}), (72, 36, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 10}), (74, 37, 0, -1, 'end', {'page': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'addr': 0}), (200, 38, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524363, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (200, 38, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 2500, 've': 589874, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (325, 38, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 75, 'vd': 5000, 've': 524363, 'vt': 125, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}), (400, 38, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}), (410, 38, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210})] 2: mem changes=[(70, 35, 0, 1, 1)] ###Markdown Send/receive a pulse with pulse_style = arb ###Code config={"res_ch":7, # --Fixed "reps":5, # --Fixed "relax_delay":0, # --Fixed "res_phase":0, # --Fixed "pulse_style": "arb", # --Fixed "sigma": 30, # [Clock ticks] # Try varying sigma from 10-50 clock ticks "readout_length":200, # [Clock ticks] # Try varying readout_length from 50-1000 clock ticks "pulse_gain":5000, # [DAC units] # Try varying pulse_gain from 500 to 30000 DAC units "pulse_freq": 100, # [MHz] # In this program the signal is up and downconverted digitally so you won't see any frequency # components in the I/Q traces below. But since the signal gain depends on frequency, # if you lower pulse_freq you will see an increased gain. "adc_trig_offset": 200, # [Clock ticks] # Try varying adc_trig_offset from 100 to 220 clock ticks "soft_avgs":100 # Try varying soft_avgs from 1 to 200 averages } config["idata"] = gauss(mu=config["sigma"]*16*5/2,si=config["sigma"]*16,length=5*config["sigma"]*16,maxv=32000) # Try varying idata to be an arbitrary numpy array of your choosing! ################### # Try it yourself ! ################### prog3 =LoopbackProgram(config) #prog3.acquire_decimated(soc, load_pulses=True, progress=True, debug=False) # print(prog3.asm()) results = simulate(prog3) print(f'3: pulses={results["pulses"]}') # print(f'3: log={results["instruction_log"]}') print(f'3: mem changes={results["mem_changes"]}') # print(f'3: reg state={results["reg_state"]}') save_results(results, "Loop3") print("Reading back results to check them") res3 = read_results("Loop3") print(f'3 saved: pulses={res3["pulses"]}') # print(f'3 saved: log={res3["instruction_log"]}') print(f'3 saved: mem changes={res3["mem_changes"]}') # print(f'3 saved: reg state={res3["reg_state"]}') # for i in results['state'].instructions: # print(i) for i in results['instruction_log']: print(i) ###Output using qick program version 3: pulses=[[7, 'arb', 0, 2400, array([61, 62, 63, ..., 63, 63, 62], dtype=int16), array([0, 0, 0, ..., 0, 0, 0], dtype=int16)]] 3: mem changes=[(36, 18, 0, 1, 1), (60, 18, 0, 1, 2), (84, 18, 0, 1, 3), (108, 18, 0, 1, 4), (132, 18, 0, 1, 5)] Reading back results to check them 3 saved: pulses=[[7, 'arb', 0, 2400, array([61, 62, 63, ..., 63, 63, 62], dtype=int16), array([0, 0, 0, ..., 0, 0, 0], dtype=int16)]] 3 saved: mem changes=[[36, 18, 0, 1, 1], [60, 18, 0, 1, 2], [84, 18, 0, 1, 3], [108, 18, 0, 1, 4], [132, 18, 0, 1, 5]] (0, 0, 3, 0, 'regwi', {'va': 69905067, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 16, 'rb': 0, 'rc': 0, 'imm': 69905067}) (2, 1, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 17, 'rb': 0, 'rc': 0, 'imm': 0}) (4, 2, 3, 0, 'regwi', {'va': 5000, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 19, 'rb': 0, 'rc': 0, 'imm': 5000}) (6, 3, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}) (8, 4, 3, 0, 'regwi', {'va': 524438, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524438}) (10, 5, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 200}) (12, 6, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 0}) (14, 7, 0, 0, 'regwi', {'va': 4, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 14, 'rb': 0, 'rc': 0, 'imm': 4}) (16, 8, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}) (18, 9, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (20, 10, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}) (22, 11, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (24, 12, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}) (26, 13, 3, 0, 'regwi', {'va': 524438, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524438}) (28, 14, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}) (30, 15, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (32, 16, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 150}) (34, 17, 0, 0, 'mathi', {'vb': 0, 'va': 1, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}) (36, 18, 0, 0, 'memwi', {'va': 1, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}) (38, 7, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 8}) (40, 8, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}) (42, 9, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (44, 10, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}) (46, 11, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (48, 12, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}) (50, 13, 3, 0, 'regwi', {'va': 524438, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524438}) (52, 14, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}) (54, 15, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (56, 16, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 150}) (58, 17, 0, 0, 'mathi', {'vb': 1, 'va': 2, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}) (60, 18, 0, 0, 'memwi', {'va': 2, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}) (62, 7, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 8}) (64, 8, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}) (66, 9, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (68, 10, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}) (70, 11, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (72, 12, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}) (74, 13, 3, 0, 'regwi', {'va': 524438, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524438}) (76, 14, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}) (78, 15, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (80, 16, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 150}) (82, 17, 0, 0, 'mathi', {'vb': 2, 'va': 3, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}) (84, 18, 0, 0, 'memwi', {'va': 3, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}) (86, 7, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 8}) (88, 8, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}) (90, 9, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (92, 10, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}) (94, 11, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (96, 12, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}) (98, 13, 3, 0, 'regwi', {'va': 524438, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524438}) (100, 14, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}) (102, 15, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (104, 16, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 150}) (106, 17, 0, 0, 'mathi', {'vb': 3, 'va': 4, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}) (108, 18, 0, 0, 'memwi', {'va': 4, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}) (110, 7, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 8}) (112, 8, 0, 0, 'regwi', {'va': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 49152}) (114, 9, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (116, 10, 0, 0, 'regwi', {'va': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 31, 'rb': 0, 'rc': 0, 'imm': 0}) (118, 11, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (120, 12, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 18, 'rb': 0, 'rc': 0, 'imm': 0}) (122, 13, 3, 0, 'regwi', {'va': 524438, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 20, 'rb': 0, 'rc': 0, 'imm': 524438}) (124, 14, 3, 0, 'regwi', {'va': 0, 'page': 3, 'ch': 0, 'oper': 0, 'ra': 21, 'rb': 0, 'rc': 0, 'imm': 0}) (126, 15, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (128, 16, 0, 0, 'synci', {'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'imm': 150}) (130, 17, 0, 0, 'mathi', {'vb': 4, 'va': 5, 'page': 0, 'ch': 0, 'oper': 8, 'ra': 15, 'rb': 15, 'rc': 0, 'imm': 1}) (132, 18, 0, 0, 'memwi', {'va': 5, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 15, 'rb': 0, 'rc': 0, 'imm': 1}) (134, 19, 0, -1, 'loopnz', {'page': 0, 'oper': 8, 'ra': 14, 'rb': 14, 'rc': 0, 'addr': 8}) (136, 20, 0, -1, 'end', {'page': 0, 'oper': 0, 'ra': 0, 'rb': 0, 'rc': 0, 'addr': 0}) (200, 21, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (350, 21, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (400, 21, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (410, 21, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (500, 21, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (550, 21, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (560, 21, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (650, 21, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (700, 21, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (710, 21, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (800, 21, 3, 7, 'set', {'va': 69905067, 'vb': 0, 'vc': 0, 'vd': 5000, 've': 524438, 'vt': 0, 'page': 3, 'ch': 7, 'oper': 0, 'ra': 0, 'rb': 16, 'rc': 21, 'rd': 17, 're': 18, 'rf': 19, 'rg': 20, 'rh': 0}) (850, 21, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (860, 21, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210}) (1000, 21, 0, 0, 'seti', {'vb': 49152, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 200}) (1010, 21, 0, 0, 'seti', {'vb': 0, 'page': 0, 'ch': 0, 'oper': 0, 'ra': 0, 'rb': 31, 'rc': 0, 'imm': 210})
notebooks/vae-resnet/vae-cifar10-depth1.nbconvert.ipynb
###Markdown Variational Autoencoder Parameters ###Code img_rows, img_cols, img_chns = 32, 32, 3 original_img_size = (img_rows, img_cols, img_chns) batch_size = int(os.environ.get('BATCH_SIZE', 25)) latent_dim = int(os.environ.get('LATENT_DIM', 256)) intermediate_dim = int(os.environ.get('INTERMEDIATE_DIM', 1024)) epsilon_std = 1.0 epochs = int(os.environ.get('EPOCHS', 1000)) activation = os.environ.get('ACTIVATION', 'sigmoid') dropout = float(os.environ.get('DROPOUT', 0.0)) decay = float(os.environ.get('DECAY', 0.0)) learning_rate = float(os.environ.get('LEARNING_RATE', 0.001)) resnet_depth = int(os.environ.get('RESNET_DEPTH', 3)) ###Output _____no_output_____ ###Markdown Load CIFAR10 dataset ###Code ftrain = H5PYDataset("../../data/cifar10/cifar10.hdf5", which_sets=('train',)) X_train, y_train = ftrain.get_data(ftrain.open(), slice(0, ftrain.num_examples)) X_train = np.moveaxis(X_train[:], 1, 3) X_train = X_train / 255. ftest = H5PYDataset("../../data/cifar10/cifar10.hdf5", which_sets=('test',)) X_test, y_test = ftest.get_data(ftest.open(), slice(0, ftest.num_examples)) X_test = np.moveaxis(X_test[:], 1, 3) X_test = X_test / 255. print(X_train.shape, y_train.shape) print(X_test.shape, y_test.shape) ###Output (50000, 32, 32, 3) (50000, 1) (10000, 32, 32, 3) (10000, 1) ###Markdown Helper Functions ###Code def create_dense_layers(stage, width): dense_name = '_'.join(['enc_conv', str(stage)]) bn_name = '_'.join(['enc_bn', str(stage)]) layers = [ Dense(width, name=dense_name), BatchNormalization(name=bn_name), Activation(activation), Dropout(dropout), ] return layers def inst_layers(layers, in_layer): x = in_layer for layer in layers: if isinstance(layer, list): x = inst_layers(layer, x) else: x = layer(x) return x def sampling(args, batch_size=batch_size, latent_dim=latent_dim, epsilon_std=epsilon_std): z_mean, z_log_var = args epsilon = K.random_normal(shape=(batch_size, latent_dim), mean=0., stddev=epsilon_std) return z_mean + K.exp(z_log_var) * epsilon def resnet_layers(x, depth, stage_base, transpose=False): assert depth in [0, 1, 2, 3] filters = [64, 64, 256] x = conv_block(x, 3, filters, stage=stage_base + 2, block='a', strides=(1, 1), transpose=transpose) if depth >= 2: x = identity_block(x, 3, filters, stage=stage_base + 2, block='b') if depth >= 3: x = identity_block(x, 3, filters, stage=stage_base + 2, block='c') filters = [128, 128, 512] x = conv_block(x, 3, filters, stage=stage_base + 3, block='a', transpose=transpose) if depth >= 1: x = identity_block(x, 3, filters, stage=stage_base + 3, block='b') if depth >= 2: x = identity_block(x, 3, filters, stage=stage_base + 3, block='c') if depth >= 3: x = identity_block(x, 3, filters, stage=stage_base + 3, block='d') filters = [256, 256, 1024] x = conv_block(x, 3, filters, stage=stage_base + 4, block='a', transpose=transpose) if depth >= 1: x = identity_block(x, 3, filters, stage=stage_base + 4, block='b') if depth >= 2: x = identity_block(x, 3, filters, stage=stage_base + 4, block='c') x = identity_block(x, 3, filters, stage=stage_base + 4, block='d') if depth >= 3: x = identity_block(x, 3, filters, stage=stage_base + 4, block='e') x = identity_block(x, 3, filters, stage=stage_base + 4, block='f') filters = [512, 512, 2048] x = conv_block(x, 3, filters, stage=stage_base + 5, block='a', transpose=transpose) if depth >= 2: x = identity_block(x, 3, filters, stage=stage_base + 5, block='b') if depth >= 3: x = identity_block(x, 3, filters, stage=stage_base + 5, block='c') return x ###Output _____no_output_____ ###Markdown Loss Function ###Code def kl_loss(x, x_decoded_mean): kl_loss = - 0.5 * K.sum(1. + z_log_var - K.square(z_mean) - K.exp(z_log_var), axis=-1) return K.mean(kl_loss) def logx_loss(x, x_decoded_mean): x = K.flatten(x) x_decoded_mean = K.flatten(x_decoded_mean) xent_loss = img_rows * img_cols * img_chns * metrics.binary_crossentropy(x, x_decoded_mean) return xent_loss def vae_loss(x, x_decoded_mean): return logx_loss(x, x_decoded_mean) + kl_loss(x, x_decoded_mean) ###Output _____no_output_____ ###Markdown VAE ###Code def make_encoder(): encoder_input = Input(batch_shape=(batch_size,) + original_img_size) resnet = resnet_layers(encoder_input, depth=resnet_depth, stage_base=0) encoder_layers = [ create_dense_layers(stage=9, width=intermediate_dim), Flatten(), ] enc_dense = inst_layers(encoder_layers, resnet) z_mean = Dense(latent_dim, kernel_regularizer=l2(0.01), bias_regularizer=l2(0.01))(enc_dense) z_log_var = Dense(latent_dim, kernel_regularizer=l2(0.1), bias_regularizer=l2(0.1))(enc_dense) return Model(inputs=encoder_input, outputs=[z_mean, z_log_var]) def make_decoder(): decoder_input = Input(batch_shape=(batch_size,) + (latent_dim,)) decoder_layers = [ create_dense_layers(stage=10, width=intermediate_dim), Reshape((4, 4, intermediate_dim // 16)), ] dec_out = inst_layers(decoder_layers, decoder_input) dec_out = resnet_layers(dec_out, depth=resnet_depth, transpose=True, stage_base=10) decoder_out = Conv2DTranspose(name='x_decoded', filters=3, kernel_size=1, strides=1, activation='sigmoid')(dec_out) return Model(inputs=decoder_input, outputs=decoder_out) encoder = make_encoder() decoder = make_decoder() encoder.summary() decoder.summary() # VAE x_input = Input(batch_shape=(batch_size,) + original_img_size) z_mean, z_log_var = encoder(x_input) z = Lambda(sampling, output_shape=(latent_dim,))([z_mean, z_log_var]) _output = decoder(z) vae = Model(inputs=x_input, outputs=_output) optimizer = Adam(lr=learning_rate, decay=decay) vae.compile(optimizer=optimizer, loss=vae_loss) vae.summary() start = time.time() early_stopping = keras.callbacks.EarlyStopping('val_loss', min_delta=0.1, patience=50) reduce_lr = keras.callbacks.ReduceLROnPlateau(monitor='val_loss', factor=0.5, patience=25, min_lr=0.001 * learning_rate) callbacks=[early_stopping, reduce_lr] if 'CMDLINE' not in os.environ: callbacks += [TQDMNotebookCallback()] history = vae.fit( X_train, X_train, batch_size=batch_size, epochs=epochs, callbacks=callbacks, validation_data=(X_test, X_test), verbose=0 ) done = time.time() elapsed = done - start print("Elapsed: ", elapsed) df = pd.DataFrame(history.history) display(df.describe(percentiles=[0.25 * i for i in range(4)] + [0.95, 0.99])) df.plot(figsize=(8, 6)) # Eval kl loss m = Model(inputs=x_input, outputs=_output) optimizer = Adam(lr=learning_rate, decay=decay) m.compile(optimizer=optimizer, loss=kl_loss) val_kl_loss = m.evaluate(x=X_test, y=X_test, batch_size=batch_size) # Eval logx loss m = Model(inputs=x_input, outputs=_output) optimizer = Adam(lr=learning_rate, decay=decay) m.compile(optimizer=optimizer, loss=logx_loss) val_logx_loss = m.evaluate(x=X_test, y=X_test, batch_size=batch_size) print() print("kl_loss = %.2f" % val_kl_loss) print("logx_loss = %.2f" % val_logx_loss) import matplotlib.pyplot as plt n = 10 figure = np.zeros((img_rows * n, img_cols * n, img_chns)) batches = (n * n + batch_size - 1) // batch_size digits = [] for i in range(batches): z_sample = np.random.normal(size=[batch_size, latent_dim]).reshape(batch_size, latent_dim) x_decoded = decoder.predict(z_sample, batch_size=batch_size) digits += [x_decoded[i].reshape(img_rows, img_cols, img_chns) for i in range(batch_size)] for j in range(n): for i in range(n): digit = digits[j * n + i] d_x = i * img_rows d_y = j * img_cols figure[d_x:d_x + img_rows, d_y:d_y + img_cols] = digit plt.figure(figsize=(10, 10)) plt.imshow(figure) plt.show() if os.environ.get('OUTDIR', None): encoder.save(os.path.join(os.environ['OUTDIR'], 'encoder-depth-' + str(resnet_depth) + '.h5')) decoder.save(os.path.join(os.environ['OUTDIR'], 'decoder-depth-' + str(resnet_depth) + '.h5')) vals = {k: v for k, v in locals().items() if type(v) in [int, float, bool]} with open(os.path.join(os.environ['OUTDIR'], 'params-depth-' + str(resnet_depth) + '.json'), 'w') as f: json.dump(vals, f) ###Output _____no_output_____
notebooks/old_notebooks/new_experiments_p_q_test.ipynb
###Markdown Parameters:- $\beta$ = {0.1, 0.5, 0.9}- $\gamma$ = 0.8- $\mu$ = 0.9- $\kappa$ = 0.05- max_infected_time = 10- NSTEPS = 100k- NAGENTS = 10k- NFRACLINKS = 0.1 Dead rate ###Code dr_small_beta_params, dr_small_beta_df = load_results( "../data/new_experiments/p_q_test/dead_ratio_p_q_L1-beta=0.1_gamma=0.8_mu=0.9_kappa=0.05_max_infected_time=10_L2-q=4_p=0.5_xi=0.1_n=100_NRUNS=100_NSTEPS=100000_NAGENTS=10000_NFRACLINKS=0.1.csv") viz.plot_imshow(dr_small_beta_df, "p", "q") dr_medium_beta_params, dr_medium_beta_df = load_results( "../data/new_experiments/p_q_test/dead_ratio_p_q_L1-beta=0.5_gamma=0.8_mu=0.9_kappa=0.05_max_infected_time=10_L2-q=4_p=0.5_xi=0.1_n=100_NRUNS=100_NSTEPS=100000_NAGENTS=10000_NFRACLINKS=0.1.csv") viz.plot_imshow(dr_medium_beta_df, "p", "q") dr_large_beta_params, dr_large_beta_df = load_results( "../data/new_experiments/p_q_test/dead_ratio_p_q_L1-beta=0.9_gamma=0.8_mu=0.9_kappa=0.05_max_infected_time=10_L2-q=4_p=0.5_xi=0.1_n=100_NRUNS=100_NSTEPS=100000_NAGENTS=10000_NFRACLINKS=0.1.csv") viz.plot_imshow(dr_large_beta_df, "p", "q") ###Output _____no_output_____ ###Markdown Infected ratio ###Code ir_small_beta_params, ir_small_beta_df = load_results( "../data/new_experiments/p_q_test/infected_ratio_p_q_L1-beta=0.1_gamma=0.8_mu=0.9_kappa=0.05_max_infected_time=10_L2-q=4_p=0.5_xi=0.1_n=100_NRUNS=100_NSTEPS=100000_NAGENTS=10000_NFRACLINKS=0.1.csv") viz.plot_imshow(ir_small_beta_df, "p", "q") ir_medium_beta_params, ir_medium_beta_df = load_results( "../data/new_experiments/p_q_test/infected_ratio_p_q_L1-beta=0.5_gamma=0.8_mu=0.9_kappa=0.05_max_infected_time=10_L2-q=4_p=0.5_xi=0.1_n=100_NRUNS=100_NSTEPS=100000_NAGENTS=10000_NFRACLINKS=0.1.csv") viz.plot_imshow(ir_medium_beta_df, "p", "q") ir_large_beta_params, ir_large_beta_df = load_results( "../data/new_experiments/p_q_test/infected_ratio_p_q_L1-beta=0.9_gamma=0.8_mu=0.9_kappa=0.05_max_infected_time=10_L2-q=4_p=0.5_xi=0.1_n=100_NRUNS=100_NSTEPS=100000_NAGENTS=10000_NFRACLINKS=0.1.csv") viz.plot_imshow(ir_large_beta_df, "p", "q") ###Output _____no_output_____
elliot_kleiman_20170321_project2.ipynb
###Markdown Project 2In this project, you will implement the exploratory analysis plan developed in Project 1. This will lay the groundwork for our our first modeling exercise in Project 3. Step 1: Load the python libraries you will need for this project ###Code #imports from __future__ import division import pandas as pd import numpy as np from scipy import stats import statsmodels.api as sm import matplotlib.pyplot as plt import pylab as pl %matplotlib inline import seaborn as sns ###Output _____no_output_____ ###Markdown Step 2: Read in your data set ###Code #Read in data from source df_raw = pd.read_csv("../assets/admissions.csv") print df_raw.head() ###Output admit gre gpa prestige 0 0 380.0 3.61 3.0 1 1 660.0 3.67 3.0 2 1 800.0 4.00 1.0 3 1 640.0 3.19 4.0 4 0 520.0 2.93 4.0 ###Markdown Questions Question 1. How many observations are in our dataset? ###Code df_raw.count() df_raw.count().sum() ###Output _____no_output_____ ###Markdown Answer: There are 1595 observations in our data set Question 2. Create a summary table ###Code summary_stats_admissions = df_raw.describe() summary_stats_admissions # Compute quantiles of gre gre_quantiles = pd.qcut(df_raw['gre'], 4) gre_quantiles.value_counts().sort_index() # Compute quantiles of gpa gpa_quantiles = pd.qcut(df_raw['gpa'], 4) gpa_quantiles.value_counts().sort_index() # What is the sample size distribution among quantiles of gre and gpa by prestige level? df_raw.pivot_table(['gre'], ['admit', gre_quantiles], [gpa_quantiles, 'prestige'], aggfunc=[len]) # What is the standard deviation distribution among quantiles of gre and gpa by prestige level? df_raw.pivot_table(['gre'], ['admit', gre_quantiles], [gpa_quantiles, 'prestige'], aggfunc=[np.std]) ###Output _____no_output_____ ###Markdown Question 3. Why would GRE have a larger STD than GPA? ###Code # Inspect gre, gpa std df_raw.std()[['gre', 'gpa']] ###Output _____no_output_____ ###Markdown Answer: Because the GRE consists of three parts: quantitative reasoning, verbal reasoning, and analytical writingand because test takers each have different academic degrees, there is going to be larger variation across testingfor knowledge in these areas. Specifically, I would expect skills in quantitative & verbal reasoning, and analytical writing to vary by academic institution, academic college, academic department, academic degree program, and degree program specialization. This is because most academic institutions, including their colleges, departments, and degree programs tend to put varied emphasis on having different quantitative, verbal, and writing skills. e.g., Theatre arts majors, might not need to have a strong background in quantitative reasoning, so they may not have to take classes focusing on quantitative reasoning (unless they went to strong engineering college). Similary, a computer engineering major may not need to have a strong background in English literature, so the emphasis is not on building analytical writing skills (unless they went to a strong liberal arts college). Therefore, I would expect more variation in GRE scores due to the many different degree programs having a different focus on acquiring different skills, and in varying amounts. Question 4. Drop data points with missing data ###Code # Which columns have missing data? df_raw.isnull().sum() # Which records are null? df_raw[df_raw.isnull().any(axis=1)] # What is shape of dataframe before dropping records? shape_before_dropna = df_raw.shape print(shape_before_dropna) # Inspect shape before dropping missing values shape_after_dropna = df_raw.dropna(how='any').shape print(shape_after_dropna) # Now, drop missing values df_raw.dropna(how='any', inplace=True) ###Output _____no_output_____ ###Markdown Question 5. Confirm that you dropped the correct data. How can you tell? Answer: Before dropping missing values the dataframe shape was (400, 4). After dropping missing values the dataframeshape was (397, 3). The `isnull()` method showed that there were three records having any values missing in a row (axis=1). Question 6. Create box plots for GRE and GPA ###Code #boxplot 1 #df_raw.boxplot('gre') sns.boxplot('gre', data=df_raw) sns.plt.title('GRE: Box and Whiskers Plot') #boxplot 2 #df_raw.boxplot('gpa') sns.boxplot('gpa', data=df_raw) sns.plt.title('GPA: Box and Whiskers Plot') ###Output _____no_output_____ ###Markdown Question 7. What do these plots show? Answer: They show the data's spread, or how far from the center the data tend to range. Specifically, boxplots show the middle fifty percent of the data, and its range.The idea is to divide the data into four equal groups and see how far apart the extreme groups are.The data is first divided into two equal high and low groups at the median, which is called the second quartile, or Q2.The median of the low group is called the first quartile or Q1. The median of the high group is the third quartile, or Q3. The box's ends are the quartiles Q1 and Q3 respectively. The box's midline is the quartile Q2, which is the median of the data. The interquartile range (IQR) is the distance between the box's ends: the distance between the third quartile and the first quartile, or Q3-Q1. These plots are especially good for showing off differences between the high and low groups, as well as outliers. Question 8. Describe each distribution ###Code # plot the distribution of each variable df_raw.plot(kind='density', subplots=True, layout=(2, 2), sharex=False) plt.show() ###Output _____no_output_____ ###Markdown The *Admit distribtion* is bimodal (has two modes, 0, and 1) as expected. Both the *GRE distribution* and *GPA distribution* are approximately symmetrical. The *Prestige distribution* is multimodal (has four modes, 1, 2, 3, 4) as expected. Question 9. If our model had an assumption of a normal distribution would we meet that requirement? ###Code # Test for normality using the Kolmogorov-Smirnov Test # GRE normal? print('GRE: ', stats.kstest(df_raw.gre, 'norm')) print('Kurtosis: ', df_raw.gre.kurt()) print('Skew: ', df_raw.gre.skew()) print('~~~~~~~~~~~') # GPA normal? print('GPA : ', stats.kstest(df_raw.gpa, 'norm')) print('Kurtosis: ', df_raw.gpa.kurt()) print('Skew: ', df_raw.gpa.skew()) print('~~~~~~~~~~~') # Admit normal? print('Admit: ', stats.kstest(df_raw.admit, 'norm')) print('Kurtosis: ', df_raw.admit.kurt()) print('Skew: ', df_raw.admit.skew()) print('~~~~~~~~~~~') # Prestige normal? print('Prestige: ', stats.kstest(df_raw.prestige, 'norm')) print('Kurtosis: ', df_raw.prestige.kurt()) print('Skew: ', df_raw.prestige.skew()) ###Output ('GRE: ', KstestResult(statistic=1.0, pvalue=0.0)) ('Kurtosis: ', -0.33286435465143427) ('Skew: ', -0.146046988215597) ~~~~~~~~~~~ ('GPA : ', KstestResult(statistic=0.98972085476178895, pvalue=0.0)) ('Kurtosis: ', -0.56356989952216807) ('Skew: ', -0.21688893296924305) ~~~~~~~~~~~ ('Admit: ', KstestResult(statistic=0.5, pvalue=0.0)) ('Kurtosis: ', -1.3865881769308692) ('Skew: ', 0.7876691478505351) ~~~~~~~~~~~ ('Prestige: ', KstestResult(statistic=0.84134474606854293, pvalue=0.0)) ('Kurtosis: ', -0.90103795489017591) ('Skew: ', 0.086505552897055041) ###Markdown Answer: No. We would not meet that requirement. Because according to the **Kolmogorov-Smirnov test**, there is zero percent chance that the test statistic values of `D` we observed for GRE, GPA, Admit, and Prestige respectively `(1.0, 0.9897, 0.5, and 0.8413`) could have arisen if the data had been drawn from a normal distribution. We therefore reject the hypothesis at the 95% confidence level that the data were drawn from a normal distribution and conclude that the data is not normally distributed. Question 10. Does this distribution need correction? If so, why? How? Answer: Yes, it needs correction. It needs correction because the distributions are not normal. They are both left-skewed and leptokurtic. I plan to remove outliers and log transform the data. ###Code # GRE IQR q3_gre = summary_stats_admissions.gre['75%'] q1_gre = summary_stats_admissions.gre['25%'] iqr_gre = q3_gre - q1_gre low_fence_gre = q1_gre - 1.5*iqr_gre high_fence_gre = q3_gre + 1.5*iqr_gre print("GRE IQR: ", iqr_gre) print("GRE low fence: ", low_fence_gre) print("GRE high fence: ", high_fence_gre) # Find GRE outliers print('Number of outliers: ', df_raw[(df_raw.gre < low_fence_gre) | (df_raw.gre > high_fence_gre)].shape[0]) print('These are the outliers: ') df_raw[(df_raw.gre < low_fence_gre) | (df_raw.gre > high_fence_gre)] # Remove GRE outliers print('Shape before outlier removal is: ', df_raw.shape) df = df_raw[(df_raw.gre >= low_fence_gre) & (df_raw.gre <= high_fence_gre)] print('Shape after outlier removal is: ', df.shape) # Plot to visually inspect distribution, still looks skewed df.gre.plot.density() plt.title('GRE density') plt.show() # GPA IQR q3_gpa = summary_stats_admissions.gpa['75%'] q1_gpa = summary_stats_admissions.gpa['25%'] iqr_gpa = q3_gpa - q1_gpa low_fence_gpa = q1_gpa - 1.5*iqr_gpa high_fence_gpa = q3_gpa + 1.5*iqr_gpa print("GPA IQR: ", round(iqr_gpa, 1)) print("GPA low fence: ", round(low_fence_gpa, 1)) print("GPA high fence: ", round(high_fence_gpa, 1)) # Now, find GPA Outliers print('Number of outliers: ', df[(df.gpa < low_fence_gpa) | (df.gpa > high_fence_gpa)].shape[0]) print('These are the outliers: ') df[(df.gpa < low_fence_gpa) | (df.gpa > high_fence_gpa)] print('Shape before outlier removal is: ', df.shape) df = df[(df.gpa >= low_fence_gpa) & (df.gpa <= high_fence_gpa)] print('Shape after outlier removal is: ', df.shape) # Plot to visually inspect distribution, still looks skewed! df.gpa.plot.density() plt.title('GPA density') plt.show() # Removed outliers: re-test for normality using the Kolmogorov-Smirnov Test # Observation: skew got better, kurtosis got worse! # GRE print('GRE: ', stats.kstest(df.gre, 'norm')) print('Kurtosis: ', df.gre.kurt()) print('Skew: ', df.gre.skew()) print('~~~~~~~~~~~') # GPA print('GPA : ', stats.kstest(df.gpa, 'norm')) print('Kurtosis: ', df.gpa.kurt()) print('Skew: ', df.gpa.skew()) # Transform GRE distribution to standard normal sns.distplot( (df.gre - df.gre.mean()) / df.gre.std(), bins=5, kde_kws={'bw':1} ) sns.plt.title('GRE to Standard Normal') sns.plt.show() # Transform GPA distribution to standard normal sns.distplot( (df.gpa - df.gpa.mean()) / df.gpa.std(), bins=10, kde_kws={'bw':1} ) sns.plt.title('GPA to Standard Normal') sns.plt.show() # Log transform the data: re-test for normality using the Kolmogorov-Smirnov Test # Observation: Skew got worse, Kurtosis got better # GRE print('GRE: ', stats.kstest(np.log(df.gre), 'norm')) print('Kurtosis: ', np.log(df.gre).kurt()) print('Skew: ', np.log(df.gre).skew()) print('~~~~~~~~~~~') # GPA print('GPA : ', stats.kstest(np.log(df.gpa), 'norm')) print('Kurtosis: ', np.log(df.gpa).kurt()) print('Skew: ', np.log(df.gpa).skew()) ###Output ('GRE: ', KstestResult(statistic=0.99999999721106625, pvalue=0.0)) ('Kurtosis: ', -0.18677728430420748) ('Skew: ', -0.47287836787183568) ~~~~~~~~~~~ ('GPA : ', KstestResult(statistic=0.81696385197730359, pvalue=0.0)) ('Kurtosis: ', -0.31189663172165183) ('Skew: ', -0.43724961849141997) ###Markdown Answer: I don't know how to correct for the skewness and kurotis inherent in this data set.But here's what I found: 1. After removing outliers, the skew got better, but the kurtosis got worse!2. After removing outliers and log transforming the data, the skew got worse, but the kurtosis got better!3. One way to normalize the data is to subtract the mean, and divide by the standard deviation. This puts the data on the standard normal scale. Question 11. Which of our variables are potentially colinear? Answer: GPA and GRE are potentially collinear. i.e., They are moderately positively correlated. ###Code # create a correlation matrix for the data df_raw.corr() sns.heatmap(df_raw.corr(), annot=True, cmap='RdBu') pd.scatter_matrix(df_raw) plt.show() ###Output _____no_output_____
notebooks/exp2-18_analysis_summary.ipynb
###Markdown Exp 2-18 analysis summary.Which parameters did best on the tasks?See `./informercial/Makefile` for experimentaldetails. ###Code import os import numpy as np from IPython.display import Image import matplotlib import matplotlib.pyplot as plt %matplotlib inline %config InlineBackend.figure_format = 'retina' import seaborn as sns sns.set_style('ticks') matplotlib.rcParams.update({'font.size': 16}) matplotlib.rc('axes', titlesize=16) from infomercial.exp import meta_bandit from infomercial.local_gym import bandit from infomercial.exp.meta_bandit import load_checkpoint import gym # ls ../data/exp2* ###Output _____no_output_____ ###Markdown Load and process data ###Code data_path ="/Users/qualia/Code/infomercial/data/" exp_names = ["exp2", "exp3", "exp4", "exp5", "exp6", "exp7", "exp8", "exp9", "exp10", "exp11", "exp12", "exp13", "exp14", "exp15", "exp16", "exp17", "exp18"] exp_index = list(range(2, 19)) num_exps = 50 num_episodes = 10000 env_names = [ "BanditOneHigh2-v0", "BanditOneHigh10-v0", "BanditOneHigh121-v0", "BanditOneHigh1000-v0", "BanditHardAndSparse2-v0", "BanditHardAndSparse10-v0", "BanditHardAndSparse121-v0", "BanditHardAndSparse1000-v0" ] last_trials = -500 exp_index # For each exp, then each task, extract the p_best, and the last_trials. # Init final result p_best = {} for env in env_names: p_best[env] = np.zeros(len(exp_names)) for j, exp_name in enumerate(exp_names): # Gather traces by bandit: scores, # Qs in a big numpy array (n_exp, n_episodes) scores_E = {} scores_R = {} values_E = {} values_R = {} controlling = {} actions = {} best = {} # Preallocate the arrays for this env for env in env_names: scores_E[env] = np.zeros((num_episodes, num_exps)) scores_R[env] = np.zeros((num_episodes, num_exps)) values_E[env] = np.zeros((num_episodes, num_exps)) values_R[env] = np.zeros((num_episodes, num_exps)) controlling[env] = np.zeros((num_episodes, num_exps)) actions[env] = np.zeros((num_episodes, num_exps)) best[env] = None # Load and repackage for n in range(num_exps): result = load_checkpoint(os.path.join(data_path, f"{exp_name}_{env}_{n+1}.pkl")) scores_E[env][:, n] = result["scores_E"] scores_R[env][:, n] = result["scores_R"] values_E[env][:, n] = result["values_E"] values_R[env][:, n] = result["values_R"] controlling[env][:, n] = result["policies"] actions[env][:, n] = result["actions"] best[env] = result["best"] # Est. prob. that the action was correct. p_best_e = {} for env in env_names: b = best[env] p_best_e[env] = np.zeros(num_episodes) for i in range(num_episodes): actions_i = actions[env][i,:] p_best_e[env][i] = np.sum(actions_i == b) / actions_i.size # Get avg. p_best of last_trials for each exp and env for env in env_names: p_best[env][j] = np.mean(p_best_e[env][last_trials:]) p_best ###Output _____no_output_____ ###Markdown Learning performanceFor each bandit env, over all exps. ###Code def plot_sum_performance(plot_names): fig = plt.figure(figsize=(4, 3*len(plot_names))) grid = plt.GridSpec(len(plot_names), 1, wspace=0.4, hspace=1.2) for i, env in enumerate(plot_names): plt.subplot(grid[i, 0]) plt.title(f"{env}") b = best[env] for n in range(num_exps): ps = p_best[env] plt.scatter(exp_index, ps, color="black", alpha=1, s=16) plt.plot(exp_index, np.ones(len(exp_index)), color="grey", alpha=0.2, ls='--', linewidth=1) plt.ylabel("p(best)") plt.xlabel("Experiment number") plt.xticks(exp_index) plt.ylim(-.1, 1.1) _ = sns.despine() ###Output _____no_output_____ ###Markdown Compare exps OneHigh ###Code plot_sum_performance(env_names[0:4]) ###Output _____no_output_____ ###Markdown Sparse ###Code plot_sum_performance(env_names[4:8]) last_trials ###Output _____no_output_____
AdvancedDataAnalysis/Final Assignment/.ipynb_checkpoints/first attempt-checkpoint.ipynb
###Markdown Data Collection ###Code posPath='aclImdb/train/pos' arr = [] for filename in tqdm(os.listdir(posPath)): path= os.path.join(posPath,filename) with open(path) as f: review = f.readlines() arr.append(review) posArr = np.array(arr) posArr = np.insert(posArr, 1, 1, axis=1) negPath='aclImdb/train/neg' arr = [] for filename in tqdm(os.listdir(negPath)): #print(filename) #path='imdb/train/neg/0_3.txt' path= os.path.join(negPath,filename) with open(path) as f: review = f.readlines() arr.append(review) negArr = np.array(arr) negArr = np.insert(negArr, 1, 0, axis=1) print(posArr.shape) print(negArr.shape) df1 = pd.DataFrame(data = posArr, columns = ['review','label']) df2 = pd.DataFrame(data = negArr, columns = ['review','label']) reviews_df = df1.append(df2) print(df1.head(2)) print(df2.head(2)) reviews_df.head() reviews_df.groupby('label').count() reviews_df['vec'] = 0 reviews_df.iloc[0,1] ###Output _____no_output_____ ###Markdown Pre Processing ###Code from sklearn import preprocessing from sklearn.preprocessing import LabelEncoder lbl_enc = preprocessing.LabelEncoder() y = lbl_enc.fit_transform(reviews_df.label) from sklearn.model_selection import train_test_split xtrain, xvalid, ytrain, yvalid = train_test_split(reviews_df.review, y, stratify=y, random_state=42, test_size=0.3, shuffle=True) from sklearn.feature_extraction.text import TfidfVectorizer, CountVectorizer # Always start with these features. They work (almost) everytime! tfv = TfidfVectorizer(min_df=3, max_features=None, strip_accents='unicode', analyzer='word',token_pattern=r'\w{1,}', ngram_range=(1, 3), use_idf=1,smooth_idf=1,sublinear_tf=1, stop_words = 'english') # Fitting TF-IDF to both training and test sets (semi-supervised learning) # Always start with these features. They work (almost) everytime! tfv = TfidfVectorizer(min_df=3, max_features=None, strip_accents='unicode', analyzer='word',token_pattern=r'\w{1,}', ngram_range=(1, 3), use_idf=1,smooth_idf=1,sublinear_tf=1, stop_words = 'english') # Fitting TF-IDF to both training and test sets (semi-supervised learning) tfv.fit(list(xtrain) + list(xvalid)) xtrain_tfv = tfv.transform(xtrain) xvalid_tfv = tfv.transform(xvalid) from sklearn.linear_model import LogisticRegression # Fitting a simple Logistic Regression on TFIDF clf = LogisticRegression(C=1.0) clf.fit(xtrain_tfv, ytrain) y_pred = clf.predict_proba(xvalid_tfv) clf.score(xvalid_tfv,yvalid) from sklearn.metrics import confusion_matrix from sklearn import metrics confusion_matrix(yvalid, clf.predict(xvalid_tfv)) yvalid y_pred[:,0] ###Output _____no_output_____
week06/prep_notebook_week06_03.ipynb
###Markdown Color SpacesIn this notebook, we will explore how colormaps move through two colorspaces, specifically [HSV](https://en.wikipedia.org/wiki/HSL_and_HSV) and RGB. ###Code %matplotlib inline import matplotlib.pyplot as plt import numpy as np from PIL import Image import IPython.display import io from mpl_toolkits.mplot3d import Axes3D import matplotlib.cm import ipywidgets plt.rcParams["figure.figsize"] = (16, 12) ###Output _____no_output_____ ###Markdown The next cell is going to set up two different functions. One will take rgb values and rotate them through RGB space, and the other will plot a nice depiction of the colormap in RGB space. ###Code def rotate(arr, theta, phi, psi): Rx = np.array([[1, 0, 0], [0, np.cos(phi), np.sin(phi)], [0.0, -np.sin(phi), np.cos(phi)]]) Ry = np.array([[np.cos(theta), 0, -np.sin(theta)], [0.0, 1.0, 0.0], [np.sin(theta), 0, np.cos(theta)]]) Rz = np.array([[np.cos(psi), np.sin(psi), 0], [-np.sin(psi), np.cos(psi), 0], [0, 0, 1]]) R = np.dot(Rx, Ry).dot(Rz) return np.dot(arr - 0.5, R) + 0.5 def plot_colortable(colortable, theta = 0.0, phi = 0.0, psi = 0.0): title = "" if isinstance(colortable, str): title = title or colortable colortable = matplotlib.cm.cmap_d[colortable](np.mgrid[0.0:1.0:256j]) colortable = rotate(colortable[:,:3], theta, phi, psi) fig = plt.figure(figsize=(20, 16)) ax = fig.add_axes([0.0, 0.25, 0.75, 0.75], projection="3d") ax.plot(colortable[:,0], colortable[:,1], colortable[:,2], '-', lw=4) ax.set_xlim(-0.05, 1.05) ax.set_ylim(-0.05, 1.05) ax.set_zlim(-0.05, 1.05) ax.set_xlabel("Red") ax.set_ylabel("Green") ax.set_zlabel("Blue") ax = fig.add_axes([0.75, 0.76, 0.25, 0.20]) ax.plot(colortable[:,0], colortable[:,1], '-', lw=4) ax.set_xlim(-0.05, 1.05) ax.set_ylim(-0.05, 1.05) ax.set_xlabel("Red", fontsize=18) ax.set_ylabel("Green", fontsize=18) ax = fig.add_axes([0.75, 0.52, 0.25, 0.20]) ax.plot(colortable[:,0], colortable[:,2], '-', lw=4) ax.set_xlim(-0.05, 1.05) ax.set_ylim(-0.05, 1.05) ax.set_xlabel("Red", fontsize=18) ax.set_ylabel("Blue", fontsize=18) ax = fig.add_axes([0.75, 0.28, 0.25, 0.20]) ax.plot(colortable[:,1], colortable[:,2], '-', lw=4) ax.set_xlim(-0.05, 1.05) ax.set_ylim(-0.05, 1.095) ax.set_xlabel("Green", fontsize=18) ax.set_ylabel("Blue", fontsize=18) # Now we do three colorbars that span the whole thing im = np.ones((16, colortable.shape[0], 4), dtype="uint8") im[...,:3] = (colortable * 255).astype("uint8")[None,:,:] im[...,3] = 255 im_no_red = im.copy() im_no_red[:,:,0] = 0 im_no_green = im.copy() im_no_green[:,:,1] = 0 im_no_blue = im.copy() im_no_blue[:,:,2] = 0 aspect = im.shape[0]/im.shape[1] * 10 ax = fig.add_axes([0.0, 0.0, 1.0, 0.05]) ax.imshow(im, interpolation='nearest', aspect = aspect) ax.set_ylabel("Standard", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) ax = fig.add_axes([0.0, 0.06, 1.0, 0.05]) ax.imshow(im_no_red, interpolation='nearest', aspect = aspect) ax.set_ylabel("No Red", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) ax = fig.add_axes([0.0, 0.12, 1.0, 0.05]) ax.imshow(im_no_green, interpolation='nearest', aspect = aspect) ax.set_ylabel("No Green", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) ax = fig.add_axes([0.0, 0.18, 1.0, 0.05]) ax.imshow(im_no_blue, interpolation='nearest', aspect = aspect) ax.set_ylabel("No Blue", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) if title is not None: fig.suptitle(title, fontsize = 24) ###Output _____no_output_____ ###Markdown Let's create a widget, and see what our colormaps look like in those dimensions. ###Code ipywidgets.interact(plot_colortable, colortable = ["viridis", "jet", "RdBu", "Blues"], theta = (0.0, 2.0*np.pi, 0.01), phi = (0.0, 2.0*np.pi, 0.01), psi = (0.0, 2.0*np.pi, 0.01)) ###Output _____no_output_____ ###Markdown A better representation of this would be in HSV space. Here we convert from RGB to HSV, which can be a tricky process. Note that HSV space is periodic in Hue, so we can rotate around the axis and it will remain continous. ###Code def move_cylinder(arr, theta, phi, psi): Rx = np.array([[1, 0, 0], [0, np.cos(phi), np.sin(phi)], [0.0, -np.sin(phi), np.cos(phi)]]) Ry = np.array([[np.cos(theta), 0, -np.sin(theta)], [0.0, 1.0, 0.0], [np.sin(theta), 0, np.cos(theta)]]) Rz = np.array([[np.cos(psi), np.sin(psi), 0], [-np.sin(psi), np.cos(psi), 0], [0, 0, 1]]) R = np.dot(Rx, Ry).dot(Rz) return np.dot(arr - 0.5, R) + 0.5 def rgb_to_hsv(arr): # https://en.wikipedia.org/wiki/HSL_and_HSV has info; we will use C_2 and H_2 # The colortable is in the range 0..1. alpha = 0.5 * (2*arr[...,0] - arr[...,1] - arr[...,2]) beta = np.sqrt(3)/2.0 * (arr[...,1] - arr[...,2]) H2 = np.arctan2(beta, alpha) H2[H2<0] += 2.0 * np.pi C2 = np.sqrt(alpha**2 + beta**2) ma = np.max(arr, axis=-1) mi = np.min(arr, axis=-1) V = ma S = C2/V np.nan_to_num(S) return np.stack([H2, S, V], axis=-1) def hsv_to_rgb(arr): # H, S, V C = arr[...,2] * arr[...,1] Hp = arr[...,0] / (np.pi/3) X = C * (1.0 - np.abs(np.mod(Hp, 2) - 1)) c1 = (0 <= Hp) & (Hp < 1) c2 = (1 <= Hp) & (Hp < 2) c3 = (2 <= Hp) & (Hp < 3) c4 = (3 <= Hp) & (Hp < 4) c5 = (4 <= Hp) & (Hp < 5) c6 = (5 <= Hp) & (Hp <= 6) rgb = np.zeros_like(arr) rgb[c1,0] = C[c1] rgb[c1,1] = X[c1] rgb[c1,2] = 0 rgb[c2,0] = X[c2] rgb[c2,1] = C[c2] rgb[c2,2] = 0 rgb[c3,0] = 0 rgb[c3,1] = C[c3] rgb[c3,2] = X[c3] rgb[c4,0] = 0 rgb[c4,1] = X[c4] rgb[c4,2] = C[c4] rgb[c5,0] = X[c5] rgb[c5,1] = 0 rgb[c5,2] = C[c5] rgb[c6,0] = C[c6] rgb[c6,1] = 0 rgb[c6,2] = X[c6] mi = arr[...,2] - C rgb += mi[...,None] return rgb def plot_colortable_hsv(colortable, hue_theta = 0.0, sat_scale = 1.0, val_scale = 1.0): title = "" if isinstance(colortable, str): title = title or colortable colortable = matplotlib.cm.cmap_d[colortable](np.mgrid[0.0:1.0:256j]) HSV = rgb_to_hsv(colortable[...,:3]) # Now we scale and rotate angle = hue_theta + HSV[...,0] HSV[...,0] = angle - 2.0*np.pi * np.floor(angle / (2.0*np.pi)) HSV[...,1] = np.clip(sat_scale * HSV[...,1], 0.0, 1.0) HSV[...,2] = np.clip(val_scale * HSV[...,2], 0.0, 1.0) fig = plt.figure(figsize=(20, 16)) ax = fig.add_axes([0.0, 0.25, 0.75, 0.75], projection="3d") x = HSV[...,1] * np.cos(HSV[...,0]) y = HSV[...,1] * np.sin(HSV[...,0]) z = HSV[...,2] ax.plot(x, y, z, '.-', markevery=10, lw=4, ms=16) ax.set_xlim(-1.05, 1.05) ax.set_ylim(-1.05, 1.05) ax.set_zlim(-0.05, 1.05) ax.set_xlabel("Hue") ax.set_ylabel("Saturation") ax.set_zlabel("Value") ax = fig.add_axes([0.75, 0.76, 0.25, 0.20], projection="polar") ax.plot(HSV[:,0], HSV[:,1], '.-', markevery=10, lw=4, ms=16) ax.set_xlim(-0.05, 2.0*np.pi + 0.05) ax.set_ylim(-0.05, 1.05) ax.set_title("Saturation", fontsize=18) ax = fig.add_axes([0.75, 0.52, 0.25, 0.20]) ax.plot(HSV[:,0], HSV[:,2], '.-', markevery=10, lw=4, ms=16) ax.set_xlim(-0.05, 2.0*np.pi + 0.05) ax.set_ylim(-0.05, 1.05) ax.set_title("Value", fontsize=18) ax = fig.add_axes([0.75, 0.28, 0.25, 0.20]) ax.plot(HSV[:,1], HSV[:,2], '.-', markevery=10, lw=4, ms=16) ax.set_xlim(-0.05, 1.05) ax.set_ylim(-0.05, 1.095) ax.set_xlabel("Saturation", fontsize=18) ax.set_ylabel("Value", fontsize=18) # Now we do three colorbars that span the whole thing colortable = hsv_to_rgb(HSV) im = np.ones((16, colortable.shape[0], 4), dtype="uint8") im[...,:3] = (colortable * 255).astype("uint8")[None,:,:] im[...,3] = 255 im_no_red = im.copy() im_no_red[:,:,0] = 0 im_no_green = im.copy() im_no_green[:,:,1] = 0 im_no_blue = im.copy() im_no_blue[:,:,2] = 0 aspect = im.shape[0]/im.shape[1] * 10 ax = fig.add_axes([0.0, 0.0, 1.0, 0.05]) ax.imshow(im, interpolation='nearest', aspect = aspect) ax.set_ylabel("Standard", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) ax = fig.add_axes([0.0, 0.06, 1.0, 0.05]) ax.imshow(im_no_red, interpolation='nearest', aspect = aspect) ax.set_ylabel("No Red", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) ax = fig.add_axes([0.0, 0.12, 1.0, 0.05]) ax.imshow(im_no_green, interpolation='nearest', aspect = aspect) ax.set_ylabel("No Green", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) ax = fig.add_axes([0.0, 0.18, 1.0, 0.05]) ax.imshow(im_no_blue, interpolation='nearest', aspect = aspect) ax.set_ylabel("No Blue", fontsize=16) ax.yaxis.set_ticks([]) ax.xaxis.set_visible(False) if title is not None: fig.suptitle(title, fontsize = 24) ipywidgets.interact(plot_colortable_hsv, colortable = ["viridis", "jet", "gray", "gist_stern", "flag", "magma"], hue_theta = (-1.0*np.pi, 1.0*np.pi, 0.01), sat_scale = (0.01, 10.0, 0.01), val_scale = (0.01, 10.0, 0.01)) ###Output _____no_output_____
mimic/notebooks/model_exploration.nbconvert.ipynb
###Markdown Test of the latent representations of the MMVAE Average precision scores on the latent representations, averaged over the batches ###Code config_path = get_config_path() with open(config_path, 'rt') as json_file: config = json.load(json_file) checkpoint_path = os.path.expanduser(config['dir_fid']) lr_evals = dict() for modality_method in ['moe']: for factorization in os.listdir(os.path.join(checkpoint_path, modality_method)): for experiment in os.listdir(os.path.join(checkpoint_path, modality_method, factorization)): if experiment.startswith('Mimic'): lr_evals[experiment] = dict() lr_eval = lr_evals[experiment] experiment_dir_ = os.path.join(checkpoint_path, modality_method, factorization, experiment) lr_eval_dir = os.path.join(experiment_dir_, 'logs', 'Latent Representation') if os.path.exists(lr_eval_dir): for label in os.listdir(lr_eval_dir): lr_eval[label] = dict() for lr in os.listdir(os.path.join(lr_eval_dir, label)): lr_eval[label][lr] = dict() for logfile in os.listdir(os.path.join(lr_eval_dir, label, lr)): for summary in summary_iterator(os.path.join(lr_eval_dir, label, lr, logfile)): value = summary.summary.value temp = str(value).split('\n') for elem in temp: elem = elem if elem.startswith('simple_value'): lr_eval[label][lr][summary.step] = elem.split(' ')[1] import pandas as pd experiments_dataframe = pd.read_csv('experiments_dataframe.csv') dfs = [] for experiment in lr_evals.keys(): experiment_evals = lr_evals[experiment] if experiment_evals: for label in experiment_evals.keys(): for lr in experiment_evals[label].keys(): steps = experiment_evals[label][lr].keys() max_step = max(steps) experiment_evals[label][lr] = experiment_evals[label][lr][max_step] df = pd.DataFrame(experiment_evals).astype(float) index = df.index index.name = f'Steps: {max_step}' df['mean'] = df.mean(numeric_only=True, axis=1) dfs.append((df, experiment)) for df, experiment in dfs: flags = experiments_dataframe.loc[experiments_dataframe['experiment_uid'] == experiment] print(f'Experiment {experiment} with text encoding: {flags.text_encoding.item()}, ' f'image size: {flags.img_size.item()}, method: {flags.method.item()} \n and trained ' f'for {flags.total_epochs.item()} epochs with batch size: {flags.batch_size.item()} ' f'and {flags.steps_per_training_epoch.item()} steps per training epoch') display(df) ###Output Experiment Mimic_2020_10_30_17_20_31_531755 with text encoding: word, image size: 128.0, method: joint_elbo and trained for 18.0 epochs with batch size: 180.0 and 200.0 steps per training epoch ###Markdown Evaluation of the classifiers All classifiers were trained for 100 epochs ###Code labels = ['Lung Opacity', 'Pleural Effusion', 'Support Devices'] FLAGS.num_features = len(alphabet) FLAGS.batch_size = 300 ###Output _____no_output_____ ###Markdown Evaluation of the character encoding and image size 128 ###Code list_precision_pa, list_precision_lat, list_precision_text = test_clfs(FLAGS, 128, 'char', alphabet) print('mean precision for pa classifier: ', np.mean(list_precision_pa)) print('mean precision for lat classifier: ',np.mean(list_precision_lat)) print('mean precision for text classifier: ',np.mean(list_precision_text)) ###Output setting dataset setting modalities setting model setting clfs setting rec_weights dict_keys(['real', 'random', '', 'PA', 'Lateral', 'text', 'Lateral_PA', 'PA_text', 'Lateral_text', 'Lateral_PA_text']) char mean precision for pa classifier: 0.39415439014746345 mean precision for lat classifier: 0.23401534481353384 mean precision for text classifier: 0.5803474159536921 ###Markdown Evaluation of the word encoding and image size 128 The text classifier precision is slightly better for the word encoding ###Code list_precision_pa, list_precision_lat,list_precision_text = test_clfs(FLAGS, 128, 'word', alphabet) print('mean precision for pa classifier: ',np.mean(list_precision_pa)) print('mean precision for lat classifier: ',np.mean(list_precision_lat)) print('mean precision for text classifier: ',np.mean(list_precision_text)) ###Output setting dataset setting modalities setting model setting clfs setting rec_weights dict_keys(['real', 'random', '', 'PA', 'Lateral', 'text', 'Lateral_PA', 'PA_text', 'Lateral_text', 'Lateral_PA_text']) word mean precision for pa classifier: 0.36514591827396703 mean precision for lat classifier: 0.2211051543545864 mean precision for text classifier: 0.674078788856655 ###Markdown Evaluation of image size 256Loading the 256 dataset crashes jupyter notebooks for some reason. If that's the case, run it from the commandline with: `jupyter nbconvert --to notebook --execute notebooks/model_exploration.ipynb` ###Code import numpy as np from mimic.dataio.MimicDataset import Mimic from mimic.utils.experiment import MimicExperiment FLAGS.text_encoding = 'char' FLAGS.img_size = 256 mimic_experiment = MimicExperiment(flags=FLAGS, alphabet=alphabet) mimic_test = Mimic(FLAGS, mimic_experiment.labels, alphabet, split='test') model_text = mimic_experiment.clfs['text'] list_precision_pa = test_clf_pa(FLAGS, mimic_experiment, mimic_test, alphabet) list_precision_lat = test_clf_lat(FLAGS, mimic_experiment, mimic_test, alphabet) list_precision_text = test_clf_text(FLAGS, mimic_experiment, mimic_test, alphabet) print('mean precision for pa classifier: ',np.mean(list_precision_pa)) print('mean precision for lat classifier: ',np.mean(list_precision_lat)) print('mean precision for text classifier: ',np.mean(list_precision_text)) ###Output char char char mean precision for pa classifier: 0.41602969689539815 mean precision for lat classifier: 0.40079714910746855 mean precision for text classifier: 0.5953110664731457
porto-seguro-safe-driver-prediction/Phase1/Python_Foundation/01-Python Crash Course.ipynb
###Markdown ___ ___ Python基础教程**代码配套视频讲解,在此不含解释**本文档按照下列顺序展开,可以依次点击链接到对应的位置。* [数据类型](数据类型) * [Numbers](Numbers) * [Strings](Strings) * [Printing](Printing) * [Lists](Lists) * [Dictionaries](Dictionaries) * [Booleans](Booleans) * [Tuples](Tuples) * [Sets](Sets)* [比较符](比较符)* [逻辑符](逻辑符)* [条件语句](条件语句)* [for循环](for循环)* [while循环](while循环)* [range()](range())* [列表推导 => list comprehension](列表推导)* [函数](函数)* [匿名函数 => lambda](匿名函数)* [map & filter](map&filter)* [其他](其他)____ 数据类型 Numbers ###Code 1 + 1 1 * 3 1 / 2 2 ** 4 4 % 2 5 % 2 (2 + 3) * (5 + 5) ###Output _____no_output_____ ###Markdown Variable Assignment ###Code # Can not start with number or special characters name_of_var = 2 x = 2 y = 3 z = x + y z ###Output _____no_output_____ ###Markdown Strings ###Code '鲸析' "Whale Project" "It's a trick!" ###Output _____no_output_____ ###Markdown Printing ###Code x = 'hello' x print(x) num = 12 name = 'Sam' print(f'My number is: {num}, and my name is: {name}') print('My number is: {one}, and my name is: {two}'.format(one=num,two=name)) print('My number is: {}, and my name is: {}'.format(num,name)) ###Output My number is: 12, and my name is: Sam ###Markdown Lists ###Code [1,2,3] ['hi',1,[1,2]] my_list = ['鲸析','数据','分析'] my_list.append('data analysis') my_list my_list[0] my_list[1] my_list[1:] my_list[:1] my_list[0] = 'NEW' my_list nest = [1,2,3,[4,5,['找到我!']]] nest[3] nest[3][2] nest[3][2][0] ###Output _____no_output_____ ###Markdown Dictionaries ###Code d = {'key1':'item1','key2':'item2'} d d['key1'] ###Output _____no_output_____ ###Markdown Booleans ###Code True False ###Output _____no_output_____ ###Markdown Tuples ###Code t = (1,2,3) t[0] t[0] = 'NEW' ###Output _____no_output_____ ###Markdown Sets ###Code {1,2,3} {1,2,3,1,2,1,2,3,3,3,3,2,2,2,1,1,2} set([1,2,3,1,2,1,2,3,3,3,3,2,2,2,1,1,2]) ###Output _____no_output_____ ###Markdown 比较符 ###Code 1 > 2 1 < 2 1 >= 1 1 <= 4 1 == 1 '你' == '我' ###Output _____no_output_____ ###Markdown 逻辑符 ###Code (1 > 2) and (2 < 3) (1 > 2) or (2 < 3) (1 == 2) or (2 == 3) or (4 == 4) ###Output _____no_output_____ ###Markdown 条件语句 ###Code if 1 < 2: print('嘿哈') if 1 > 2: print('嘿哈') if 1 < 2: print('first') else: print('last') if 1 > 2: print('first') else: print('last') if 1 == 2: print('first') elif 3 == 3: print('middle') else: print('Last') ###Output middle ###Markdown for循环 ###Code seq = [1,2,3,4,5] for item in seq: print(item) for item in seq: print('哈哈哈') for num in seq: print(num+num) ###Output 2 4 6 8 10 ###Markdown while循环 ###Code 数字 = 1 while 数字 < 5: print('数字是:{}'.format(数字)) 数字 = 数字 + 1 ###Output 数字是:1 数字是:2 数字是:3 数字是:4 ###Markdown range() ###Code range(5) for i in range(5): print(i) list(range(5)) ###Output _____no_output_____ ###Markdown 列表推导 ###Code x = list(range(10)) x out = [] for item in x: out.append(item**3) print(out) [item**3 for item in x] ###Output _____no_output_____ ###Markdown 函数 ###Code def my_func(param1='default'): """ 注释:一般写参数文档 """ print(param1) my_func my_func() my_func('new param') my_func(param1='new param') def my_square(x): return x**2 out = my_square(2) print(out) ###Output 4 ###Markdown 匿名函数 ###Code def quad(x): return (x+1)**2 quad(2) lambda var: var*2 ###Output _____no_output_____ ###Markdown map and filter ###Code seq = [1,2,3,4,5] map(quad,seq) list(map(quad,seq)) list(map(lambda var: var*2,seq)) filter(lambda item: item%2 == 0,seq) list(filter(lambda item: item%2 == 0,seq)) ###Output _____no_output_____ ###Markdown 其他 ###Code st = 'hello my name is Whale' st.lower() st.upper() st.split() tweet = "let's go #数据科学" tweet.split('#') tweet.split('#')[1] d d.keys() d.items() d.values() lst = [1,2,3] lst.pop() lst 'x' in [1,2,3] 'x' in ['x','y','z'] ###Output _____no_output_____
Inverted_Residual.ipynb
###Markdown **Inverted Residual**The difference between residual blockand inverted residual.They expand features with the first conv instead of reducing them. The following image should make this clearshow lower image : ![Screen Shot 1401-01-10 at 18.04.53.png](data:image/png;base64,iVBORw0KGgoAAAANSUhEUgAABE4AAAF2CAYAAABwLd3rAAAMbWlDQ1BJQ0MgUHJvZmlsZQAASImVVwdYU8kWnluSkJDQAghICb0jUgNICaEFkF4EGyEJJJQYE4KKvSwquHYRxYquiii2lWYBsSuLYu+LBRVlXdTFhsqbkICu+8r3zvfNvX/OnPlPuTO59wCg+YErkeShWgDkiwukCeHBjDFp6QzSU4AAEtABfkCdy5NJWHFx0QDK4P3v8u4GtIZy1VnB9c/5/yo6fIGMBwAyDuJMvoyXD3EzAPgGnkRaAABRobecUiBR4DkQ60phgBCvVuBsJd6lwJlKfHTAJimBDfFlANSoXK40GwCNe1DPKORlQx6NzxC7ivkiMQCaThAH8IRcPsSK2J3y8ycpcDnEdtBeAjGMBzAzv+PM/ht/5hA/l5s9hJV5DYhaiEgmyeNO+z9L878lP08+6MMGDqpQGpGgyB/W8FbupCgFpkLcLc6MiVXUGuIPIr6y7gCgFKE8IllpjxrzZGxYP6APsSufGxIFsTHEYeK8mGiVPjNLFMaBGO4WdKqogJMEsQHEiwSy0ESVzRbppASVL7Q2S8pmqfTnuNIBvwpfD+S5ySwV/xuhgKPixzSKhEmpEFMgtioUpcRArAGxiyw3MUplM6pIyI4ZtJHKExTxW0GcIBCHByv5scIsaViCyr4kXzaYL7ZFKOLEqPDBAmFShLI+2CkedyB+mAt2WSBmJQ/yCGRjogdz4QtCQpW5Y88F4uREFc8HSUFwgnItTpHkxanscQtBXrhCbwGxh6wwUbUWTymAm1PJj2dJCuKSlHHiRTncyDhlPPhyEA3YIAQwgByOTDAJ5ABRW3ddN/ylnAkDXCAF2UAAnFWawRWpAzNieE0EReAPiARANrQueGBWAAqh/suQVnl1BlkDs4UDK3LBU4jzQRTIg7/lA6vEQ95SwBOoEf3DOxcOHow3Dw7F/L/XD2q/aVhQE63SyAc9MjQHLYmhxBBiBDGMaI8b4QG4Hx4Nr0FwuOFM3Gcwj2/2hKeEdsIjwnVCB+H2RNE86Q9RjgYdkD9MVYvM72uB20BOTzwY94fskBnXx42AM+4B/bDwQOjZE2rZqrgVVWH8wP23DL57Gio7sisZJQ8jB5Htflyp4aDhOcSiqPX39VHGmjlUb/bQzI/+2d9Vnw/vUT9aYouwQ9hZ7AR2HjuK1QEG1oTVY63YMQUe2l1PBnbXoLeEgXhyIY/oH/64Kp+KSspcq127XD8r5woEUwsUB489STJNKsoWFjBY8O0gYHDEPBcnhpurmxsAineN8u/rbfzAOwTRb/2mm/87AP5N/f39R77pIpsAOOANj3/DN50dEwBtdQDONfDk0kKlDldcCPBfQhOeNENgCiyBHczHDXjBd1oQCAWRIBYkgTQwAVZZCPe5FEwBM8BcUAxKwXKwBqwHm8E2sAvsBQdBHTgKToAz4CK4DK6Du3D3dIKXoAe8A30IgpAQGkJHDBEzxBpxRNwQJhKAhCLRSAKShmQg2YgYkSMzkPlIKbISWY9sRaqQA0gDcgI5j7Qjt5GHSBfyBvmEYigV1UVNUBt0BMpEWWgUmoSOR7PRyWgRugBdipajlegetBY9gV5Er6Md6Eu0FwOYOqaPmWPOGBNjY7FYOpaFSbFZWAlWhlViNVgjfM5XsQ6sG/uIE3E6zsCd4Q6OwJNxHj4Zn4Uvwdfju/Ba/BR+FX+I9+BfCTSCMcGR4EvgEMYQsglTCMWEMsIOwmHCaXiWOgnviESiPtGW6A3PYhoxhziduIS4kbiP2ExsJz4m9pJIJEOSI8mfFEvikgpIxaR1pD2kJtIVUifpg5q6mpmam1qYWrqaWG2eWpnabrXjalfUnqn1kbXI1mRfciyZT55GXkbeTm4kXyJ3kvso2hRbij8liZJDmUspp9RQTlPuUd6qq6tbqPuox6uL1Oeol6vvVz+n/lD9I1WH6kBlU8dR5dSl1J3UZupt6lsajWZDC6Kl0wpoS2lVtJO0B7QPGnQNFw2OBl9jtkaFRq3GFY1XmmRNa02W5gTNIs0yzUOalzS7tchaNlpsLa7WLK0KrQatm1q92nTtkdqx2vnaS7R3a5/Xfq5D0rHRCdXh6yzQ2aZzUucxHaNb0tl0Hn0+fTv9NL1Tl6hrq8vRzdEt1d2r26bbo6ej56GXojdVr0LvmF6HPqZvo8/Rz9Nfpn9Q/4b+p2Emw1jDBMMWD6sZdmXYe4PhBkEGAoMSg30G1w0+GTIMQw1zDVcY1hneN8KNHIzijaYYbTI6bdQ9XHe433De8JLhB4ffMUaNHYwTjKcbbzNuNe41MTUJN5GYrDM5adJtqm8aZJpjutr0uGmXGd0swExkttqsyewFQ4/BYuQxyhmnGD3mxuYR5nLzreZt5n0WthbJFvMs9lnct6RYMi2zLFdbtlj2WJlZjbaaYVVtdceabM20FlqvtT5r/d7G1ibVZqFNnc1zWwNbjm2RbbXtPTuaXaDdZLtKu2v2RHumfa79RvvLDqiDp4PQocLhkiPq6OUoctzo2O5EcPJxEjtVOt10pjqznAudq50fuui7RLvMc6lzeTXCakT6iBUjzo746urpmue63fXuSJ2RkSPnjWwc+cbNwY3nVuF2zZ3mHuY+273e/bWHo4fAY5PHLU+652jPhZ4tnl+8vL2kXjVeXd5W3hneG7xvMnWZccwlzHM+BJ9gn9k+R30++nr5Fvge9P3Tz9kv12+33/NRtqMEo7aPeuxv4c/13+rfEcAIyAjYEtARaB7IDawMfBRkGcQP2hH0jGXPymHtYb0Kdg2WBh8Ofs/2Zc9kN4dgIeEhJSFtoTqhyaHrQx+EWYRlh1WH9YR7hk8Pb44gRERFrIi4yTHh8DhVnJ5I78iZkaeiqFGJUeujHkU7REujG0ejoyNHrxp9L8Y6RhxTFwtiObGrYu/H2cZNjjsST4yPi6+If5owMmFGwtlEeuLExN2J75KCk5Yl3U22S5Ynt6RopoxLqUp5nxqSujK1Y8yIMTPHXEwzShOl1aeT0lPSd6T3jg0du2Zs5zjPccXjboy3HT91/PkJRhPyJhybqDmRO/FQBiEjNWN3xmduLLeS25vJydyQ2cNj89byXvKD+Kv5XQJ/wUrBsyz/rJVZz7P9s1dldwkDhWXCbhFbtF70OiciZ3PO+9zY3J25/Xmpefvy1fIz8hvEOuJc8alJppOmTmqXOEqKJR2TfSevmdwjjZLukCGy8bL6Al34Ud8qt5P/JH9YGFBYUfhhSsqUQ1O1p4qntk5zmLZ42rOisKJfpuPTedNbZpjPmDvj4UzWzK2zkFmZs1pmW85eMLtzTvicXXMpc3Pn/jbPdd7KeX/NT53fuMBkwZwFj38K/6m6WKNYWnxzod/CzYvwRaJFbYvdF69b/LWEX3Kh1LW0rPTzEt6SCz+P/Ln85/6lWUvblnkt27ScuFy8/MaKwBW7VmqvLFr5eNXoVbWrGatLVv+1ZuKa82UeZZvXUtbK13aUR5fXr7Nat3zd5/XC9dcrgiv2bTDesHjD+438jVc2BW2q2WyyuXTzpy2iLbe2hm+trbSpLNtG3Fa47en2lO1nf2H+UrXDaEfpji87xTs7diXsOlXlXVW123j3smq0Wl7dtWfcnst7Q/bW1zjXbN2nv690P9gv3//iQMaBGwejDrYcYh6q+dX61w2H6YdLapHaabU9dcK6jvq0+vaGyIaWRr/Gw0dcjuw8an604pjesWXHKccXHO9vKmrqbZY0d5/IPvG4ZWLL3ZNjTl47FX+q7XTU6XNnws6cPMs623TO/9zR877nGy4wL9Rd9LpY2+rZevg3z98Ot3m11V7yvlR/2edyY/uo9uNXAq+cuBpy9cw1zrWL12Out99IvnHr5ribHbf4t57fzrv9+k7hnb67c+4R7pXc17pf9sD4QeXv9r/v6/DqOPYw5GHro8RHdx/zHr98InvyuXPBU9rTsmdmz6qeuz0/2hXWdfnF2BedLyUv+7qL/9D+Y8Mru1e//hn0Z2vPmJ7O19LX/W+WvDV8u/Mvj79aeuN6H7zLf9f3vuSD4YddH5kfz35K/fSsb8pn0ufyL/ZfGr9Gfb3Xn9/fL+FKuQOfAhgcaFYWAG92AkBLA4AO+zbKWGUvOCCIsn8dQOA/YWW/OCBeANTA7/f4bvh1cxOA/dth+wX5NWGvGkcDIMkHoO7uQ0Mlsix3NyUXFfYphAf9/W9hz0ZaBcCX5f39fZX9/V+2wWBh79gsVvagCiHCnmFL6JfM/Ezwb0TZn36X4493oIjAA/x4/xfFxZDFmlNreQAAAIplWElmTU0AKgAAAAgABAEaAAUAAAABAAAAPgEbAAUAAAABAAAARgEoAAMAAAABAAIAAIdpAAQAAAABAAAATgAAAAAAAACQAAAAAQAAAJAAAAABAAOShgAHAAAAEgAAAHigAgAEAAAAAQAABE6gAwAEAAAAAQAAAXYAAAAAQVNDSUkAAABTY3JlZW5zaG90bYbIhwAAAAlwSFlzAAAWJQAAFiUBSVIk8AAAAddpVFh0WE1MOmNvbS5hZG9iZS54bXAAAAAAADx4OnhtcG1ldGEgeG1sbnM6eD0iYWRvYmU6bnM6bWV0YS8iIHg6eG1wdGs9IlhNUCBDb3JlIDYuMC4wIj4KICAgPHJkZjpSREYgeG1sbnM6cmRmPSJodHRwOi8vd3d3LnczLm9yZy8xOTk5LzAyLzIyLXJkZi1zeW50YXgtbnMjIj4KICAgICAgPHJkZjpEZXNjcmlwdGlvbiByZGY6YWJvdXQ9IiIKICAgICAgICAgICAgeG1sbnM6ZXhpZj0iaHR0cDovL25zLmFkb2JlLmNvbS9leGlmLzEuMC8iPgogICAgICAgICA8ZXhpZjpQaXhlbFlEaW1lbnNpb24+Mzc0PC9leGlmOlBpeGVsWURpbWVuc2lvbj4KICAgICAgICAgPGV4aWY6UGl4ZWxYRGltZW5zaW9uPjExMDI8L2V4aWY6UGl4ZWxYRGltZW5zaW9uPgogICAgICAgICA8ZXhpZjpVc2VyQ29tbWVudD5TY3JlZW5zaG90PC9leGlmOlVzZXJDb21tZW50PgogICAgICA8L3JkZjpEZXNjcmlwdGlvbj4KICAgPC9yZGY6UkRGPgo8L3g6eG1wbWV0YT4KqOdScgAAABxpRE9UAAAAAgAAAAAAAAC7AAAAKAAAALsAAAC7AAIUoFsjBPYAAEAASURBVHgB7L0H/C5VdS78HlCKghiVJgTRmEhRQQEpii3qjdgSG5aY2HKTqLnXkmtuNOa70agpIsZ2xUixROw3igUMoEZAVIqAinQQNQL2Sj3zrbKftZ+1Z+Z/znnhIOCe8zuznrXWs/eaWe9u77wz8181yLaY2NS4iu1iGMQAm5ZaVRTnUglAk8NikFIo51WCwAEyHjHE0OP3/KMd9fbX+18ff3g0pRET0GQff/v80+dfzBvcY/KKI2voQmEVQ19/9PUH2lFff/T1R19/8GhKIyagyb7+6OuPW976YxUunIyaN88MsXqoYMSvLkGp51TPyAyDlmiaV49fZ6aawUCjfIVHAfIKWZxQIYPX8z/KZ29/vf1hZZT6liuj9pI46GCQvf9ZBpAOyD7+SFr8q9ioPfXxp48/ffxJoyoro/7CzhhXYqBxL1TI4PX1zyifffzp408ff9Kowsqov7AzxpUYaPr4oxlAOiDDcPMbf1cNq2WExGX08uHHeem5qrt0ILYXaoiRrzE0qpUz24SDTT1+z39vf/hyha9Z0e0CcJ+pnSvcNEQ1tlHBOr4ps/e/3v96/+v9z8YC2TVLBTXbNhpGGkOj1jITDjb18aePP3386eOPDhg8LtgAQruRrzE0qpU024SDTX386eNPH3/6+KMDBo8LcceJj0Hsaqmtr6nJKwijs2XP97eCM3s1u43BOuNS0YQJp+cu2ff449Vuz//Mrwltg2KdcW9/loGJlPT+50mJfR9/+vjTXm3o428ffyd/zW0HVNYZ9/mnzz+SgYkmAaO7ZN/nnz7/9PmnDJgYN6Vf9PE358S0dkBhnTHyKLJtW2VQcrbsb6Hjj1044auqPvBqYkpGLAORBssYclWt2WIk3ilREzrZWMVDd7X0+JYsyVfPvzUha2S2sy6pttza3Gfc6ZVEaVLC6+3P09Tse/+rv6r08Uf7k259/LE02PDiY0yTmbREMG4ff2q78YTUtGhu+vjLWQncx98+/uJX7T7/NKOsDb22s9FVO01f//nQUbOSMxIDC4ClVHZ9/EVGkuzjbx9/13X89TtOrAdSN7SOJm0L/TE1s2klFUnKmD9yq2GV7XzpOSKM62gtqUhSWuZoanJDj68fQs+/ZmEN7Wfcopo2tYbyI7caevvr7a/3vz7+YCjQQabPv5qFtdrSmJqUcfGRu4+/ff7p8690lL7+i6FAh40+/o4HzxlLGlOTMi4wckfSe/uLVGjaevsbN54ZS2pTSRkXGLkj6Wvf/lbJK070ZhrarJaiU4hiZq+SXJ9xolZxDzIx1XfropapV+zAh9pVyhHOhHDzjLPH9wxIenr+e/vr/Q8jXRkvZFDhVhGdJWYs5enWx58+/qZmYK1Cd96SSntCswpvAWLv4y/3NCSq9z/OircW5EY1xbr18aePP6kZWKvQnbeW0ma46QTDSX384Z6GRPXxh7PiTQa5UU2xbn386eNPagbWKnTnraW0GW46wXDSLW38kTtOVsvp0pUWfLegE5/LB1FWhKm8KrpFHPfaPhGdpvsZcyWsAaXyqujW43seuOmnRBW3iBlzJawBpfI9/56t3v5Kq/HWYfvUUGqjmjFXwhpQKq+Kbj3/nofe/60xWBtJDaWkR8SMuRLWgFJ5VXTr7c/z0NufNQZrI6mhlPSImDFXwhpQKq+Kbr39eR56+7PGYG0kNZSSHhEz5kpYA0rlVdGttz/PQ29/1hisjaSGUtIjYsZcCWtAqbwquvX253no7c8ag7WR1FBKekRMmevLYae8KIuX2Qln6l0voE3JWu3U1V0qUYlkLLDHl44uPb3nv7c/aQYx5o97yshSu1Xvf+NfVyhdNVFkLLCPP3386eNvn3/6/Nvn3z7/9vXHeIUwa6nLir7+6uuves/3qMHUhjJyyUtA+/rrJrb+mnhUp3xu9mEpnv6attLnbJdo8PFPFi+lRcxeiOnxS+onEzh5FQwp7/mPTMw0397+LDG9//XxZ+6LQB9/+/hrU0+ff2g2CVhmkNATUCe2yfT1+afPP9Iw+vzb519pBtNDhDQOc0x6+/pfxtfpzIijj7+YfWaS1OcfS8z1GH/jUR3LdMlnzbq3wdpAa8Kt1U7wvaw74IaMepOBFILgZlPRYIQEOaQ74IZs3EUlL0Fws6loMEKCHNIdcEM27qKSlyC42VQ0GCFBDukOuCEbd1HJSxDcbCoajJAgh3QH3JCNu6jkJQhuNhUNRkiQQ7oDbsjGXVTyEgQ3m4oGIyTIId0BN2TjLip5CYKbTUWDERLkkO6AG7JxF5W8BMHNpqLBCAlySHfADdm4i0peguBmU9FghAQ5pDvghmzcRSUvQXCzqWgwQoIc0h1wQzbuopKXILjZVDQYIUEO6Q64IRt3UclLENxsKhqMkCCHdAfckI27qOQlCG42FQ1GSJBDugNuyMZdVPISBDebigYjJMgh3QE3ZOMuKnkJgptNRYMREuSQ7oAbsnEXlbwEwc2mosEICXJId8AN2biLSl6C4GZT0WCEBDmkO+CGbNxFJS9BcLOpaDBCghzSHXBDNu6ikpcguNlUNBghQQ7pDrghG3dRyUsQ3GwqGoyQIId0B9yQjbuo5CUIbjYVDUZIkEO6A27Ixl1U8hIEN5uKBiMkyCHdATdk4y4qeQmCm01FgxES5JDugBuycReVvATBzaaiwQgJckh3wA3ZuItKXoLgZlPRYIQEOaQ74IZs3EUlL0Fws6loMEKCHNIdcEM27qKSlyC42VQ0GCFBDukOuCEbd1HJSxDcbCoajJAgh3QH3JCNu6jkJQhuNhUNRkiQQ7oDbsjGXVTyEgQ3m4oGIyTIId0BN2TjLip5CYKbTUWDERLkkO6AG7JxF5W8BMHNpqLBCAlySHfADdm4i0peguBmU9FghAQ5pDvghmzcRSUvQXCzqWgi6h0nYECi5IR0yjxx7GktVQ8EADkRFyanzBPHntZS9UAAkAg2IZ0yTxx7WkvVAwFATsSFySnzxLGntVQ9EAAkgk1Ip8wTx57WUvVAAJATcWFyyjxx7GktVQ8EAIlgE9Ip88Sxp7VUPRAA5ERcmJwyTxx7WkvVAwFAItiEdMo8cexpLVUPBAA5ERcmp8wTx57WUvVAAJAINiGdMk8ce1pL1QMBQE7Ehckp88Sxp7VUPRAAJIJNSKfME8ee1lL1QACQE3Fhcso8cexpLVUPBACJYBPSKfPEsae1VD0QAOREXJicMk8ce1pL1QMBQCLYhHTKPHHsaS1VDwQAOREXJqfME8ee1lL1QACQCDYhnTJPHHtaS9UDAUBOxIXJKfPEsae1VD0QACSCTUinzBPHntZS9UAAkBNxYXLKPHHsaS1VDwQAiWAT0inzxLGntVQ9EADkRFyYnDJPHHtaS9UDAUAi2IR0yjxx7GktVQ8EADkRFyanzBPHntZS9UAAkAg2IZ0yTxx7WkvVAwFATsSFySnzxLGntVQ9EAAkgk1Ip8wTx57WUvVAAJATcWFyyjxx7GktVQ8EAIlgE9Ip88Sxp7VUPRAA5ERcmJwyTxx7WkvVAwFAItiEdMo8cexpLVUPBAA5ERcmp8wTx57WUvVAAJAINiGNIn9URyVtKAmz3m8iGM/UwO1Wul0KDkiqkmB4DYQ2wVCfbj1+z7+0hd7+oiug0+XeAw3Se0+7D6+B0IgGm0rdev/r/U/aQu9/0RV6//ORASNF1rLVfXUfXgOhVULchK4+3fr408cfaQt9/Imu0McfHxny6AEN0jntPrwGQiMabCp16+NPH3+kLfTxJ7pCH38kFfJHdeSdPRgsfKiY2k8zYIXMJfWazCp9qYsshtKLgYSOdujrpOnyXNs0A1ZILiExevye/97+ev/r408ff2W215nINpku+vzj66A+/0qL6OsfTYL3jZn9zAqrlJvx9vVXX3/19Zf0kf79p3//6+uPmGFuAesv+6s6adoLJcDMVFrNYMo86X/8pchggBCGDJI7lACZPKGB2eP3/NvLl6VB2HyNtoIGAr2RyR1KgIY9VsHs7a+3v97+9IJ17399/KFxEgMkmRgmdygBmDqJwUS7gwwyCGHIILlDCZDJExqYiAsZVBDCkEFyhxIgkyc0MBEXMqgghCGD5A4lQCZPaGAiLmRQQQhDBskdSoBMntDARFzIoIIQhgySO5QAmTyhgYm4kEEFIQwZJHcoATJ5QgMTcSGDCkIYMkjuUAJk8oQGJuJCBhWEMGSQ3KEEyOQJDUzEhQwqCGHIILlDCZDJExqYiAsZVBDCkEFyhxIgkyc0MBEXMqgghCGD5A4lQCZPaGAiLmRQQQhDBskdSoBMntDARFzIoIIQhgySO5QAmTyhgYm4kEEFIQwZJHcoATJ5QgMTcSGDCkIYMkjuUAJk8oQGJuJCBhWEMGSQ3KEEyOQJDUzEhQwqCGHIILlDCZDJRat/jrgYnJ4LJdvoqCbrJWOty5Ds4tc+YgGmWMWYbD1+860EmZuTnj31GpJdz7+kcCZdqa0VTrL19tfbX/pWPNOQwuytR1VDsuv9r/e/Pv5EB0kgjbXFk2x9/O3jbx9/U59ZWfHeoxxDsuvzT59/+vwz3WvSXFMoydbnnz7/yPzTXDjxJmLtJWCAbBYNbcgYmVaaHMSKTpC0RvlfunTAAMZjrcf3Nmw54cRQRh2u6CQ28QIG6PmXDHA2evvr7U/X8NYmuGFQj3K4opPYxAsYwHis9fbX219vf73/2ZjAAwONKH380QysmBzKFvECBjAea3387eNvH3/7+GtjAg8MNKI4XNFJbOIFDNDHH8kAZ+NXOf42F06aA5MDnbsyCSafCLWAEXReZRuqavDZxDgIAdy7MifIJeGVbaiqQWQT4yAEcO/KnCD3+JKKVdTsLW8TyWMT45pJIPeuzAF33Fqt3ERhNjGuNQG5d2UOuD1+my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkD7g2f/7hwMn8AeLERMQIGqEdoaM6uTvggvWjW3AaPv1iIGAEDcAHBc3alwQfpRbPmNnh6fH2xEWUoYABOmOA5u9Lgg/SiWXMbPD3/Pf+9/VEPCRiAO4zgObvS4IP0ollzGzy9//X+1/sf9ZCAAbjDCJ6zKw0+SC+aNbfB0/tf73+9/1EPCRiAO4zgObvS4IP0ollzGzy9//X+1/sf9ZCAAbjDCJ6zKw0+SC+aNbfBc1Psf3LhZLUcM91XEmcQgM8i4czABZZCSU5RyoOVuL2mVpSINa+R4MpsUS7Z43sD6/m3DKTG0dtf738yxuljNdIU9PbauqWG0sefSEeAmqoGZUYff/v4qwvssqXGIUqf//v408ffPv/IUNDnXwySKtNASWpj5yIFZ0aff/v82+ffG2v9EXecUI+1blm7YUWjvpt7brhbM3SXvI8iAsByW41aEbMN5yLhbs3QXfI+iggAy201akXMNpyLhLs1Q3fJ+ygiACy31agVMdtwLhLu1gzdJe+jiACw3FajVsRsw7lIuFszdJe8jyICwHJbjVoRsw3nIuFuzdBd8j6KCADLbTVqRcw2nIuEuzVDd8n7KCIALLfVqBUx23AuEu7WDN0l76OIALDcVqNWxGzDuUi4WzN0l7yPIgLAcluNWhGzDeci4W7N0F3yPooIAMttNWpFzDaci4S7NUN3yfsoIgAst9WoFTHbcC4S7tYM3SXvo4gAsNxWo1bEbMO5SLhbM3SXvI8iAsByW41aEbMN5yLhbs3QXfI+iggAy201akXMNpyLhLs1Q3fJ+ygiACy31agVMdtwLhLu1gzdJe+jiACw3FajVsRsw7lIuFszdJe8jyICwHJbjVoRsw3nIuFuzdBd8j6KCADLbTVqRcw2nIuEuzVDd8n7KCIALLfVqBUx23AuEu7WDN0l76OIALDcVqNWxGzDuUi4WzN0l7yPIgLAcluNWhGzDeci4W7N0F3yPooIAMttNWpFzDaci4S7NUN3yfsoIgAst9WoFTHbcC4S7tYM3SXvo4gAsNxWo1bEbMO5SLhbM3SXvI8iAsByW41aEbMN5yLhbs3QXfI+iggAy201akXMNpyLhLs1Q3fJ+ygiACy31agVMdtwLhLu1gzdJe+jiACw3FajVsRsw7lIuFszdJe8jyICwHJbjVoRsw3nIuFuzdBd8j6KCADLbTVqRcw2nIuEuzVDd8n7KCIALLfVqBUx23AuEu7WDN0l76OIALDcVqNWxGzDuUi4WzN0l7yPIgLAcluNWhGzDeci4W7N0F3yPooIAMttNWpFzLYLJ/gVNlGoHg5lmHwW0HT9OUGqjks+GkYMK73CuxxJj+9X4Xv+6ZoxtSWHtCdfb3+SDMtH7399/Onjb59/yqRqQscGGRfSnMx+x33+7fOvv+Syz7/xmy2tMRzSnnx9/SHJsHz09Udff8h8kuYabRt9/sk56fNvm4Gb4/qD7jhpTidNDupjQ/mKz6YoTl//wx+gVEN6lGvAiMKGHt+WOJySSF/Pfyz/Ij8BevuzVFA+ot00YERhQ+9/vf/JVwxuEtF8+vjTx5/y9TPaR4A+/loqKB/RbxoworChj799/O3jb59/ZMxIFyp0DOnzb59/+/xr3SKmzAA32PqDLpx45RSiXipho83vZDDouu51q3258oAgncl797A/cADwyWDQdd3r1uN7HkorMQUZgwSjyppD5C+4AcAmg0HXda8byvf4NU9AkJ4p3ruH/YEDgE8Gg67rXreef89Db3+1nQBBIkNVuof9gQOATQaDrutet97+PA+9/dV2AgSJDFXpHvYHDgA2GQy6rnvdevvzPPT2V9sJECQyVKV72B84ANhkMOi67nXr7c/z0NtfbSdAkMhQle5hf+AAYJPBoOu61623P89Db3+1nQBBIkNVuof9gQOATQaDrutet1+H9ueP6uBkKR+WAdxDI0q4Ahij7lr7pA5jkVlQkFJtjy+t0JshMlc/iJp6Q0FA7kTWFlxyC1JOPKyjunv+e/57+8vdKzpL6WcQrX1Sh7HILPr4h/RETsXQ219vf5KBaBoB0EiKbO2TOoxFZkFBUKcQevuzZCBz9YPo+U8ZiASh7Yjs66+aIssPklRkFr3/IT3IWl9/9/G3zz/WG6JrBEAnKbK1T+owFpnFOo0/9Y4T6qR8SE0odgUGZxw5KD7fVmJxkKHHj0GiZq1mlDLFbsPVV1FLMs/ITYae/57/Mkhz20ELgWQfcPVVBB+keUZuMvT219tfb3/oLiHRQyDDQaD6KiK3QfOM3GTo/a/3v97/2m4T14mop8xy6mptROnrX0nJ+NEaymoff/r408ef0cCBHgI5Ioih+ipqeeYZucnQ+9869b964YQzTfk0c6sHd+wYW4JcgZAGGUXLk1jVDtRW0urgUZOBaZYKgsoev+e/t7/e/9JPgjRAtINIqwd17BhbglyBkPr418f/Pv/xT/K1e4ym9dlONXaMLVQvYO9/ffzp83+f//v8jxExy3YQbfVgjx1jS5ArEFJf//T1z815/ZMvnFiDlgsvaOIzV6Hg5gWOd5im20A1qTu9ZiHpKna4ub70EuYef/IqGOcLH5bnsskoVJO66/nv7a/3vz7+yEAggzyGBxsYfHCQBU0f//v8V1pEn3/7/DvxKzCPF3394dnwsbQZUaGa1F1ff/X1V19/9fWXDAR9/XWzXn/6hRMM8Da0+y6bSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0grMN9xEgWJCNuEyVxmJ2fAAHFlqVomfu1EnFppskQl2VoqopoDBoii1dLjcy7alE5mZ66A2ckZMEDPvyRYf8WuGcm457/NAGeq+CZM5jE7OQMGiLxXS88/56LN/mR25gqYnZwBA/T8S4J7/8+tqraOcevLzOKfK2B2cgYM0Ntfb3+9/0kbqD0i43EPZGbvf5aBiZRUOzkDBoi8V0vPP+eit782AxPZmTD19icZsLxQcgIGWC/9Ty6crB4GuT8bd2TWcPXDhA1SPcCQld0iZjAGT/7ueI/f81/uj59uIfNfPKb4aFkumcEYrN7+ev/r418f/8ejRR0h+vgzd+FnakRF3sYZnWL38bePv3387ePveLTAOIJRA1LtwJDgjiUzGIPZx58+/vTxp48/Ph5MjxDj9R/dcVKKmODibNfK2YfBZ+1llBaQnqdHvUYIVo0XpgBrH5SYUVpAj+8NwtNTMmMistTzn9qlZopz45lbl32UFtDbX29/5XphbVfWQKKVNHZtaexTfd22KC2gt7/e/nr7Q/8pPcNE9BJxsl257EPZtZdRWkDvf73/9f6HvsP9LHpJ7W9hCoCC6ySjtIDe/3r/6/0P3af0DBPRS3r/w3wfKXFAF06QQJfBg1kMPNDwe+OcSyUATcoVXblRsjZQrRAEVD6WI4YYevw60PX8Sy5Ko/K2Qi0G0GRvf73/9fGnj788x2CAYFvGI4YY+vzT5x/0oz7/9vm3rz98zPSxkkZMQJN9/dXXX339hXmDe0xecWQNXSisYujrj5vO+iMunIyGN14ZxKdXwYhfXYLwsUMWJ1TI4GmJZnjp8evKJOXWlVG+EgcJhuz5twwgHZC9/UlafEgftafe/3r/wzeDNLa4MmoviYMOBlmcUCF7/5PE9P6nrWPUnvr408efPv6UgXMsRv0lUTDAQhYnVMg+/kpi+virrWPUnvr428ffPv6WgXMsVukrTprbQWg4lQ6l7pLAGG/H9aQy5m7IjVopEw429fg9/739YXLHND/ugNxnaueqvJFfXGabcLCp97/e/3r/6/1PRxIeF+rI4mjkawyNaoXMNuFgUx9/+vjTx58+/uiAweOCjzp1P/I1hka1gmabcLCpjz99/OnjTx9/dMDgcSHuOPEhiF0ttfU1NXkFYXS27Pn+InBmr2a2MVhnXCqaMOH03CX7Hn/8bbvnf+ZqctugWGfc259lYCIlvf95UmLfx58+/vi6A7OfdBFpHZO/5rQdinXGffzp449kYKJJwOgu2ffxp48/ffypY6+iPv72+afPv7lPmNZOKKwzLkUnTL8u849dOOGrqn7impgy2lpyPEO61w3jcLVmi7NobwVlN9lYdRyrVzV7/CbLlmTb2Tqp57+3v9zbvG14b2Pc+19kQNOivaePP5ESBn387fMPflXr868NFtI9+vrHxgibUmzX1x9l0OzzryeitoqckZKmKqxLya7PvzUnhPr82+ffPv/yGKKd46Y9//odJzYC0jBoA109durjszAVScq4yMithlW28x9RRoRxHa0lFUlKyxwtDd3Q4+uH0POvWVhD+xm3qKZNraH8yK2G3v56++v9r48/GAp0kMFaQvEatjSmJGVccOTu408ff/v8Ix2lr39iKNBho48/48FzxpLG1KSMC4zckfTe/iIVmrbe/saNZ8aS2lRSxgVG7kh6b3+RCk3bCu1vlbziRG/mpM2KFp1SXMzsVZLrM07UKu5BJub6bmXUMnolUdToRZWnmxzhTAg3zzi9sJXt8Xv+e/tDTy/9RToVtwrvLvCppli33v/6+JOagbUK3XlrKW2Gm04wnNTHX+5pSFTvf5wVbzLIjWqKdevjTx9/UjOwVqE7by2lzXDTCYaT+vjDPQ2J6uMPZ8WbDHKjmmLd+vjTx5/UDKxV6M5bS2kz3HSC4aQ+/nBPQ6JuvuOP3HGyWs6CrjThuxV98DhNMq0TTOVV0S3iuNf2ieg03c+YK2ENKJVXRbce3/PAXT8lqrhFzJgrYQ0ole/592z19ldajbcO26eGUhvVjLkS1oBSeVV06/n3PPT+b43B2khqKCU9ImbMlbAGlMqroltvf56H3v6sMVgbSQ2lpEfEjLkS1oBSeVV06+3P89DbnzUGayOpoZT0iJgxV8IaUCqvim69/XkeevuzxmBtJDWUkh4RM+ZKWANK5VXRrbc/z0Nvf9YYrI2khlLSI2LKXF8OO+VFWbxMSThT7xoDbUrWaqeuLlGJSiRjgT2+dHTp6T3/vf1JM4gxf9xTRpbarXr/42veKyRq5IqXyfX+1/tf7399/BmPELOWPv5ivurzT59/6j2/ow5TO8rI1edfSU5f//fvP339eZNZf048qlPGLbtYoXj6a9pK45xdosHwN1m8lF6pIfT4JfWTCZy8CoaU9/xHJmaab29/lpje/+YH4j7+9PHHht4+/tJoGrCMoKEnoE5sk+nr428ff6Vh9Pmnzz/SDKaHCGkc5pj09vWvjK/TmRFHH38x+8wkqc8/lpg+/i49/sajOtbSSnuqrc77YO2gtcFZr53ge1l3wA0Z9SYDKQTBzaaiwQgJckh3wA3ZuItKXoLgZlPRYIQEOaQ74IZs3EUlL0Fws6loMEKCHNIdcEM27qKSlyC42VQ0GCFBDukOuCEbd1HJSxDcbCoajJAgh3QH3JCNu6jkJQhuNhUNRkiQQ7oDbsjGXVTyEgQ3m4oGIyTIId0BN2TjLip5CYKbTUWDERLkkO6AG7JxF5W8BMHNpqLBCAlySHfADdm4i0peguBmU9FghAQ5pDvghmzcRSUvQXCzqWgwQoIc0h1wQzbuopKXILjZVDQYIUEO6Q64IRt3UclLENxsKhqMkCCHdAfckI27qOQlCG42FQ1GSJBDugNuyMZdVPISBDebigYjJMgh3QE3ZOMuKnkJgptNRYMREuSQ7oAbsnEXlbwEwc2mosEICXJId8AN2biLSl6C4GZT0WCEBDmkO+CGbNxFJS9BcLOpaDBCghzSHXBDNu6ikpcguNlUNBghQQ7pDrghG3dRyUsQ3GwqGoyQIId0B9yQjbuo5CUIbjYVDUZIkEO6A27Ixl1U8hIEN5uKBiMkyCHdATdk4y4qeQmCm01FgxES5JDugBuycReVvATBzaaiwQgJckh3wA3ZuItKXoLgZlPRYIQEOaQ74IZs3EUlL0Fws6loMEKCHNIdcEM27qKSlyC42VQ0GCFBDukOuCEbd1HJSxDcbCoajJAgh3QH3JCNu6jkJQhuNhUNRkiQQ7oDbsjGXVTyEgQ3m4oGIyTIId0BN2TjLip5CYKbTUWDERLkkO6AG7JxF5W8BMHNpqLBCAlySHfADdm4i0peguBmU9FE1DtOwIBEyQnplHni2NNaqh4IAHIiLkxOmSeOPa2l6oEAIBFsQjplnjj2tJaqBwKAnIgLk1PmiWNPa6l6IABIBJuQTpknjj2tpeqBACAn4sLklHni2NNaqh4IABLBJqRT5oljT2upeiAAyIm4MDllnjj2tJaqBwKARLAJ6ZR54tjTWqoeCAByIi5MTpknjj2tpeqBACARbEI6ZZ449rSWqgcCgJyIC5NT5oljT2upeiAASASbkE6ZJ449raXqgQAgJ+LC5JR54tjTWqoeCAASwSakU+aJY09rqXogAMiJuDA5ZZ449rSWqgcCgESwCemUeeLY01qqHggAciIuTE6ZJ449raXqgQAgEWxCOmWeOPa0lqoHAoCciAuTU+aJY09rqXogAEgEm5BOmSeOPa2l6oEAICfiwuSUeeLY01qqHggAEsEmpFPmiWNPa6l6IADIibgwOWWeOPa0lqoHAoBEsAnplHni2NNaqh4IAHIiLkxOmSeOPa2l6oEAIBFsQjplnjj2tJaqBwKAnIgLk1PmiWNPa6l6IABIBJuQTpknjj2tpeqBACAn4sLklHni2NNaqh4IABLBJqRT5oljT2upeiAAyIm4MDllnjj2tJaqBwKARLAJ6ZR54tjTWqoeCAByIi5MTpknjj2tpeqBACARbEIaRf6ojkraUBJmvd9EMF5uArdb6XYxOCCpSoLhNRDaBEN9uvX4Pf/SFnr7i66ATpd7DzRI7z3tPrwGQiMabCp16/2v9z9pC73/RVfo/c9HBowUWctW99V9eA2EVglxE776dOvjTx9/pC308Se6Qh9/fGTIowc0SOe0+/AaCI1osKnUrY8/ffyRttDHn+gKffyRVMgf1ZF3VmGw8KFiaj/NgBUyl9RrMqv0pUayGEovxhI62qGvk6bLc23TDFghuYTE6PF7/nv76/2vjz99/JXZXmci22S66POPr4P6/Cstoq9/NAneN2b2MyusUm7G29dfff3V11/SR/r3n/79r68/Yoa5Bay/7K/qpGkvlAAzU2k1gynzpL/8uchggBCGDJI7lACZPKGB2eP3/NvLx6VB2HyNtoIGAr2RyR1KgIY9VsHs7a+3v97+9IJ17399/KFxEgMkmRgmdygBmDqJwUS7gwwyCGHIILlDCZDJExqYiAsZVBDCkEFyhxIgkyc0MBEXMqgghCGD5A4lQCZPaGAiLmRQQQhDBskdSoBMntDARFzIoIIQhgySO5QAmTyhgYm4kEEFIQwZJHcoATJ5QgMTcSGDCkIYMkjuUAJk8oQGJuJCBhWEMGSQ3KEEyOQJDUzEhQwqCGHIILlDCZDJExqYiAsZVBDCkEFyhxIgkyc0MBEXMqgghCGD5A4lQCZPaGAiLmRQQQhDBskdSoBMntDARFzIoIIQhgySO5QAmTyhgYm4kEEFIQwZJHcoATJ5QgMTcSGDCkIYMkjuUAJk8oQGJuJCBhWEMGSQ3KEEyOQJDUzEhQwqCGHIILlDCZDJRat/jrgYnJ4LJdvoqCbrJWOty5Ds4tc+YgGmWMWYbD1+860EmZuTnj31GpJdz7+kcCZdqa0VTrL19tfbX/pWPNOQwuytR1VDsuv9r/e/Pv5EB0kgjbXFk2x9/O3jbx9/U59ZWfHeoxxDsuvzT59/+vwz3WvSXFMoydbnnz7/yPzTXDjxJmLtJWCAbBYNbcgYmVaaHMSKTpC0RvlfunTAAMZjrcf3Nmw54cRQRh2u6CQ28QIG6PmXDHA2evvr7U/X8NYmuGFQj3K4opPYxAsYwHis9fbX219vf73/2ZjAAwONKH380QysmBzKFvECBjAea3387eNvH3/7+GtjAg8MNKI4XNFJbOIFDNDHH8kAZ+NXOf42F06aA5MDnbsyCSafCLWAEXReZRuqavDZxDgIAdy7MifIJeGVbaiqQWQT4yAEcO/KnCD3+JKKVdTsLW8TyWMT45pJIPeuzAF33Fqt3ERhNjGuNQG5d2UOuD1+my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkDbs9/my3TJ5LHJsY1k0DuXZkD7g2f/7hwMn8AeLERMQIGqEdoaM6uTvggvWjW3AaPv1iIGAEDcAHBc3alwQfpRbPmNnh6fH2xEWUoYABOmOA5u9Lgg/SiWXMbPD3/Pf+9/VEPCRiAO4zgObvS4IP0ollzGzy9//X+1/sf9ZCAAbjDCJ6zKw0+SC+aNbfB0/tf73+9/1EPCRiAO4zgObvS4IP0ollzGzy9//X+1/sf9ZCAAbjDCJ6zKw0+SC+aNbfBc1Psf3LhZLUcM91XEmcQgM8i4czABZZCSU5RyoOVuL2mVpSINa+R4MpsUS7Z43sD6/m3DKTG0dtf738yxuljNdIU9PbauqWG0sefSEeAmqoGZUYff/v4qwvssqXGIUqf//v408ffPv/IUNDnXwySKtNASWpj5yIFZ0aff/v82+ffG2v9EXecUI+1blm7YUWjvpt7brhbM3SXvI8iAsByW41aEbMN5yLhbs3QXfI+iggAy201akXMNpyLhLs1Q3fJ+ygiACy31agVMdtwLhLu1gzdJe+jiACw3FajVsRsw7lIuFszdJe8jyICwHJbjVoRsw3nIuFuzdBd8j6KCADLbTVqRcw2nIuEuzVDd8n7KCIALLfVqBUx23AuEu7WDN0l76OIALDcVqNWxGzDuUi4WzN0l7yPIgLAcluNWhGzDeci4W7N0F3yPooIAMttNWpFzDaci4S7NUN3yfsoIgAst9WoFTHbcC4S7tYM3SXvo4gAsNxWo1bEbMO5SLhbM3SXvI8iAsByW41aEbMN5yLhbs3QXfI+iggAy201akXMNpyLhLs1Q3fJ+ygiACy31agVMdtwLhLu1gzdJe+jiACw3FajVsRsw7lIuFszdJe8jyICwHJbjVoRsw3nIuFuzdBd8j6KCADLbTVqRcw2nIuEuzVDd8n7KCIALLfVqBUx23AuEu7WDN0l76OIALDcVqNWxGzDuUi4WzN0l7yPIgLAcluNWhGzDeci4W7N0F3yPooIAMttNWpFzDaci4S7NUN3yfsoIgAst9WoFTHbcC4S7tYM3SXvo4gAsNxWo1bEbMO5SLhbM3SXvI8iAsByW41aEbMN5yLhbs3QXfI+iggAy201akXMNpyLhLs1Q3fJ+ygiACy31agVMdtwLhLu1gzdJe+jiACw3FajVsRsw7lIuFszdJe8jyICwHJbjVoRsw3nIuFuzdBd8j6KCADLbTVqRcw2nIuEuzVDd8n7KCIALLfVqBUx23AuEu7WDN0l76OIALDcVqNWxGzDuUi4WzN0l7yPIgLAcluNWhGzDeci4W7N0F3yPooIAMttNWpFzLYLJ/gVNlGoHg5lmHwW0HT9OUGqjks+GkYMK73CuxxJj+9X4Xv+6ZoxtSWHtCdfb3+SDMtH7399/Onjb59/yqRqQscGGRfSnMx+x33+7fOvv+Syz7/xmy2tMRzSnnx9/SHJsHz09Udff8h8kuYabRt9/sk56fNvm4Gb4/qD7jhpTidNDupjQ/mKz6YoTl//wx+gVEN6lGvAiMKGHt+WOJySSF/Pfyz/Ij8BevuzVFA+ot00YERhQ+9/vf/JVwxuEtF8+vjTx5/y9TPaR4A+/loqKB/RbxoworChj799/O3jb59/ZMxIFyp0DOnzb59/+/xr3SKmzAA32PqDLpx45RSiXipho83vZDDouu51q3258oAgncl797A/cADwyWDQdd3r1uN7HkorMQUZgwSjyppD5C+4AcAmg0HXda8byvf4NU9AkJ4p3ruH/YEDgE8Gg67rXreef89Db3+1nQBBIkNVuof9gQOATQaDrutet97+PA+9/dV2AgSJDFXpHvYHDgA2GQy6rnvdevvzPPT2V9sJECQyVKV72B84ANhkMOi67nXr7c/z0NtfbSdAkMhQle5hf+AAYJPBoOu61623P89Db3+1nQBBIkNVuof9gQOATQaDrutet1+H9ueP6uBkKR+WAdxDI0q4Ahij7lr7pA5jkVlQkFJtjy+t0JshMlc/iJp6Q0FA7kTWFlxyC1JOPKyjunv+e/57+8vdKzpL6WcQrX1Sh7HILPr4h/RETsXQ219vf5KBaBoB0EiKbO2TOoxFZkFBUKcQevuzZCBz9YPo+U8ZiASh7Yjs66+aIssPklRkFr3/IT3IWl9/9/G3zz/WG6JrBEAnKbK1T+owFpnFOo0/9Y4T6qR8SE0odgUGZxw5KD7fVmJxkKHHj0GiZq1mlDLFbsPVV1FLMs/ITYae/57/Mkhz20ELgWQfcPVVBB+keUZuMvT219tfb3/oLiHRQyDDQaD6KiK3QfOM3GTo/a/3v97/2m4T14mop8xy6mptROnrX0nJ+NEaymoff/r408ef0cCBHgI5Ioih+ipqeeYZucnQ+9869b964YQzTfk0c6sHd+wYW4JcgZAGGUXLk1jVDtRW0urgUZOBaZYKgsoev+e/t7/e/9JPgjRAtINIqwd17BhbglyBkPr418f/Pv/xT/K1e4ym9dlONXaMLVQvYO9/ffzp83+f//v8jxExy3YQbfVgjx1jS5ArEFJf//T1z815/ZMvnFiDlgsvaOIzV6Hg5gWOd5im20A1qTu9ZiHpKna4ub70EuYef/IqGOcLH5bnsskoVJO66/nv7a/3vz7+yEAggzyGBxsYfHCQBU0f//v8V1pEn3/7/DvxKzCPF3394dnwsbQZUaGa1F1ff/X1V19/9fWXDAR9/XWzXn/6hRMM8Da0+y6bSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0gjSPTlIOqCpFqyiTSCRF8Ooi5IqiWbSCNI9OUg6oKkWrKJNIJEXw6iLkiqJZtII0j05SDqgqRasok0gkRfDqIuSKolm0grMN9xEgWJCNuEyVxmJ2fAAHFlqVomfu1EnFppskQl2VoqopoDBoii1dLjcy7alE5mZ66A2ckZMEDPvyRYf8WuGcm457/NAGeq+CZM5jE7OQMGiLxXS88/56LN/mR25gqYnZwBA/T8S4J7/8+tqraOcevLzOKfK2B2cgYM0Ntfb3+9/0kbqD0i43EPZGbvf5aBiZRUOzkDBoi8V0vPP+eit782AxPZmTD19icZsLxQcgIGWC/9Ty6crB4GuT8bd2TWcPXDhA1SPcCQld0iZjAGT/7ueI//a5f/M888c/Gtb31rccABj7zBP/8jjzxyscceeyx+53d+h1qqtrfe/vwrHPqe52Rd+993v/vdxcFveMPivHPPXey7336Ll7zkJYsNMIBw1YY554xBrP1/9erVi8+feOLiIx/60OJnP/vZ4tBDD001cGlgSNQ2lsxgDGaNr5ZpxvwX3yn+VVddtfjsZz+7+PhRH19cc+3Vi7e97ZASbIq9dvGvvvqqxdFHH7P40Ac/aG37f77whaMvAzijLDkmY2ddddWVi8985nOLT3z8KDnWaxf/921vs3q5DpSCVB8wJPMzZgZjsNbu/OcuPKBGHU+OOuqoxfHHH7943/vet9hyyy1LADBUZXzDxkdtY8kxGYN5w5w/ahtLjskYzB5fx7+vnnXm4lKZjx55wAGp/V991dWLTx0j/e5DH1jc9757LF4k/U43ZBIS2RzLyjjyfe9d7HHfPcu8BGbP/7rOP5o5ZBUS2RxLZjAG85aTf50/TzjhhMWHP/zhxU9/9tPFYYceRpnS871lnz8+UZbf/e5/Ld7whn9ZnCtrlf3uv+/ixS+StcoGGxBluc//Osn1iZLrD334I4uf21rlHVQnQ845Y3CWi4+aIFHbWDKDsTPXx/yv659jdMz84IcW993jPosXvvBF5bDG8bVNrm3/v1Lq/Zysq3Se17XKIbJWmaox54AZjMFa+/hcGhgStY0lMxyfJWuVjx31MVmrfEbWKkcu7nSnrX7tvv/VPN34+a+xFS0XXxotNr2AIpuJgquh2INgnmV2UbOAwKM47Ck4TAGWCV9jSjW5pqKZYA/bNST71v0QorSAwFZN0Uywh+03//gy4A2vfvWrh41uvdHwgP33pwTyeS5//ldffc2wxRa3GzbZZJPh4DccTPU7jJoFBDZX0Uywh+3D8M53HjFsu802w7bbbjNsI1L/q77N1tC3FdvW7hPO9ttvP+x+n92HRzz8vw3PeMYzhre89a3DBeefb8E5ShzNGuIHb3Rma2eImAICr8P5a6kDn/zkQa6T6Cxg/w866KC1Cy6siFniX3HF5cP7jjxy+OM//qNhq622ijp33333UmcpwQXXOtqYyNUENhrHYQ/blcg+r3/1ddcNp592+nDwwQcPj370Y4bNNrutn4fkaO+993ZS2UdpAYHNVzQTji/55iXDIW9/2/C433/csPltN5M67eng4QXPf36qc12Ua/VYT5djff3Bw6Me/ehh880309ea2PHasVJ8r5ePSy1FX5egxI3SAgKbv2gm2MN2Jbr+/e9/b3j/+98/PPvZzx623267aDeao0suuZgiZhg1CwhslKKZYA/blci+XPfaaFFaQGArWDQT7GG7Etm3NhEzJ0oLCGyUoplgD9uVyL5c99poUVpAYCtYNBPsYbsS2bc2ETMnSgtQfJ30h9fofLTRrYf9dT4So7afQw45xPrdZptLX5Z+rH3k+c9/wdLxfV7aYthY5yUZJzSOxq9b0Uywh+3KZl8t/cEPfLDMR3n+sTnK5iq1bzNsTfPVfxx7bK0A9S4ZnypaEcbRCwhsJYpmgj1sVyL7Vgw16YzSAgIbs2gm2MN2JbJvMsSKxigtILCVKJoJ9rBdiewbhssvu3yQH4pk/vzjMn/6HFHnz3w4UVpA4OsRP9e+Zi1i3kjxnyxrFe2/mDtfh7XKEvEvv+wKyfV7h2dKrrfeuq5V7hNrlZve+bdHtL7m/0sulrWKjpmP+/20/nl+s1ZZl8//utWyVvnKabaOf8yjHi31bm7zvI7F++yT11Xtec7p6xLf6ygluOBc5TP27+la5X0fsLXKnWWtgvWWrp+/KWu8cZwIJq7rH58PK2oWENgIHIc9bFci+7jmtcNRWkBgK1o0E+xhuxLZt3YxmRWlBQQ2QtFMsIftSnSdLpxw9XCTTfipOlIcjgylwOpUzmskLoVgOGKIgW2rSXE4MvT4lpKbTv4vv/zy4aEPfahNZPvtu++g+txGn6ZTxMC2lT7/s848a9jxrne1QfbAA588yJXqJgzX1LiKOmKU+F84+QvDS1/60mGvvfamL2uysKYLCXZBYZUvYHBxIUnh7nSPewxyR8VwzTXXTh7AXHyQVzr/SJRVcsN+/lrbRhttFAsRPa8/ePzj5bDoiAHXIv4mm2ya8yj16evTdmsXI1IXqtUc/KrOfyr/uhib+/z33vt+6bi9PJ8Jaszym9+8dJQX+VnC+o4uRpY9f73IhS+C3iZrO+WLPKMjvInlf/PNt8j5of6nizjb7CRu2PaPT2nZ/NOBoapJeVPP/y3h/GM+kjFn3zIfXXrppTb+pPHaxqTF8LwX6IUT35Y5/7POknlpxx2t/z3lwAMH+WUW1Y3kun7+p59+qv0goRcSb7f57SRG7dd8Lnoh/3++8H8MB73udYOe69y2rvHT6IzCJnv/QzpqrseW6nM0YoiBbWh/+gORfr7t/LP7fe5TCvz65n/1dbpWuXWaJx7frlXaxBedc20mMWy8qaxVYp6p/cvXKlQC0ORNK//rY/7/po6ZkRf/EQbjz/Of/7zUbmda9+hT0NTZsUrbtvatsqx/VL/f/faJMp5uJF3MgCZ/9fnf/HYyHpfzMCm5wsWTSy7BhZM4HQM4hbCKgW3o/+p3O3kBTf7qzx+HE+eSzqRaGY3KiIFtv+rzjwsno/TykfEZFTziJw5OETIKOQhzAElKxUbq8VNGW2WUr0RALiGLEyok5XxU3w2c///6zneG3/6tu9vEc8AjDxiu/OUv5aDiQG7wz/8y+SVmt93ubQPWIx/5yOGaq6+u4SJsgKXi/9VfvTQPiDI4ym2gw7//+/8bzvnGOcOJJ54wfPCDH7QJYI/73tcHy2aCedCDHjR874rvLRU/feRJwXlB3rCf/6677uqTWTmXt7zlLSn6RKLdj8OBlLM+44wzhi+e/MXhMY99TM2l1Lv7brs1dWZ11F6TGwEgixMqpMTHNqpvHdr/L6Uty6Miwyc/+Un5JWSfch6+sOKLEREuwgYYff5Xy5eqr8idIcced9yw884719xIG7MLJ1EZzoAl6oUsPlF/eeWVw5lnnDV8yo41X/zLx9qU5eoFj/KV/CgLWeMbCnOAcX1ryP9FF180nHLKl4e/+Iu/SLnRhUm946TWn+KGOcA6x/9Vn3+PXz+71PSmP2inoIjI78h8dPe7/5a1nQMOOGD4xS9/YZyrr7nafuU89j/+Y9hF+x2N1/zr6bL5v+zyy4bd7r2bxX3kI39vuFrnJdlG9a2h/Y/4Vovv9C6+3/iNO6R+oQv1/fbZd7j6mmvq8CxRsY3qux7xa4Bav8WBCtnjI/1Lf/5n6vz5xS8Oj33sY9PnvfvucuGE8vvrmv9Yq0j71z6gd/zWtERDXKv821rliydHrnGxYHx3D+qFLB8zVMh6IGsVPxqLfbJRCZvnArldiqyP+V/HML2L9Vi5g62uVXz9w2NmnG4ceoDJ89fvCGfKxea8rvILKfeTH6TmN9QLOZeW6l+f49/FF188nPLlLw//4wWyVqH5xNcqfuFkfcaPPOF0IeMD0S4RRqf38T/SNgUWbb6UxCmU5yajXEVhCjDyNYZGtXJmm3Cwqcev2ago0h5g5GsMjXqj5v8KuTCw8y672MT+kIc8ZNAvmrrZMU0cGJuuz+d/2WXflcWxXKyROxie9vSnW0zeXd/4F154QVqs6MRst3vHydVoGutzn/vscLe73S2V0cFTbyfl7YY6/7k6YV/2/E/58inDAx7wAHkMabvhRS960fDTn/40jRlWP5+EGBq1Uorj5C98IeUFi5Hr8/njPFWuKf4cd13j6ySPyVEXVrgYcX3iH/TPryt1+mJk6lGdUf2NoVHtdD+hxyrtT//rXT5Tx7qu5x95bAI2qtHMNuFg05ri66MWW2zhd5/Erzh6+ytXItEa9QaLj/Md1d8YGrXHx2cykRg2renzXyb/V1xxxbAL5qMH1/moret1B0m/s76hv3bKHSf2qA5YWfIxm6cxsCrviJJ56bet7qc/bWZeKtUve/5/8tw/oV+Bfdw44ojD00HbMfGBISaxlo3fdriJME6ZcLDpphZf3pUxHC8Xs9uNj9l8jaFRK2XCwaa1Pf+TZP7E+Kdtdffd8KirHynXiWM324SDTWsbn8vUk0OkX934e8opulbZf9BHJF70ohfaWgVHtez5f+EkX6vwhZOb6vmnc4VS5PqY/18XaxW/CwoXTkb5kWNYl/x/8lO6VvFxTNs51io4pVH9jaFRrdi6xNcCU3WsS3xeq2DdFXfHoqImzs29/+G0pnJ3Y+cfxwJ5feLHHSdeWXt6rDMuoSdMaF7ukv0Uhy7G4CR6fM1AmyzWGZesTZhQh7tkP8W5EfP/y1/8cthjj/vawnPbbe8sd1dcIcc0dVDr5/z1sR27XVMWE/KSME/cDRT/Jz/+iS1+MQjql8/Hyp0T9UzG+ddfiDZNt3z6F9fjP3O8FGvzwjpjP40RPUf2+iaK3Zj5nzzGifx/45y3Y4h8AABAAElEQVRvRC41j37hhA+e8U33/E897dQ4D20Xe99v4lncifP3M2rP0fVDDzs06tTcxIWTlr6On/9pp50W9epiWx8ryhsHYFxYEya0YXfJfoqzjufv0SYqKia/OOoXfzTnevuru2Q/Ueym2P6nz5EPnvFNK//laNJB+dHKfuKwf5X5118x7yOPMWh733bbbQe9iFK3fLCHHvoO7x/C1XaFLwHBz/R1Ov+zvnrWcOtb+2MEb/iXMi+NksUBGKdQcTgOnPeyl788+rZ9wZNzkJdLF8pEXeZp7awzLiEnTPjA3SX7Kc4N3P/L0aSDWh/xr5S79e4jd3I865nP9FhT51ZOeH3Ez+cpGsX/xje+YRe/sRaJx0eIE+VvpvmP4weYOrcbIf/nSK6RZ5X4kccPKz55UQVPHeNNJP/rY/6PMVPyYmPm8ybex7bE+Z9W1lW6/okLJ1O5vRE+fzS/yc92DfF/237I9Qvxmp/8qE57QqwzLkcwYUKDc5fspzhL5H+yoqm613D+duS3gPh24YSvanmCKCMGXdc9eVKK/KNkb/lwVVjBGZ+6UyKNXAtbMS/beHp8y5LnxhPGuKbwV5n/F7/4JTHJfPSjH6WDqnB9f/4vf5kvIvXdHF+W2/vbbdn411x3bf2VRwZBHdAfJ7fMjjZruPWzedlf/7XkxK+e6wJeB9A//e//vRQzcq3CinnZxnOLav/nnnNOyom+UNe2m9n5yxvTy90h3h723luexbUPzj/D+sFWtKb2d9jheuGktBdpK89/gS9GvEaul3Gtfy6+Pl6E9ueLEX1uWOugegy63ngKi7hcjsJ7lcxj5/Uf/3faaSfrQ3oO48UITmn9xa9nMxNDzWmOqyUUrenzx+dh1VBRj8YxGTfEHt8S8uKXvLiM2asGzEdz+T/ssMOda2P0KrnjRN4t1KQVn42b2dsQJ/L/cr24IXXrBRS9lbvWQfUYdF335Cl4bNF6XvOa18R5ap/Q//IXVzxEs587f6NZ9R5D9+NoY0tTvReaOH/wbi7xn/Pc51oen/ksv3BSsxJnApClJY7z1LhTboxcCVbMyzae9Pmfc8655fP2L2Xpy7wV9DpqxRXdXPK/0vn72cyc4w18/uecq2sVz7POOTXXN058fHIejWMyBkukmlMbc9/6mP8PP+ywtKawF2rPxMcRrk37O1N+AMU4pmOm3nHiZ8vnzBi167nr/xmfudhn5FrYXO5vPEvF32nnndK4rI8Vr8356wHdEPHriVXU4/vn6xlpsmwu97PH7zhpnKNPqOZ4FnGlayqfuFpjj48k1FRYTmbTPXKknCZlRB1/POsp/5/53GfkdmH5wicD3X777RcHMjq89RQfAX/4ox8Ot72t/6UTfVTmRz/+secZhCXjXysXTvQLrV4Fx6CuzxqPE4xALj/04Q8b37/oedn9H/gAc8ahGMjlVtKUHkWSMi41cltBLx3Q1XHhGYvSo0hSxgVGbjGccy5+xfF8tLcaj2vJllRnUjJPtZFbDeXoAxpQ+9ptqNMWI2gPZYJvawA37BG0HEVDOOzww1IbiztOooLmnJryRDMIt13kaY41DsVAW3JeR50pwAw9cZUTQQmuQ/ydd8rvgLnk4m+WiqYP4IaOj1OIQx4FyMcxcltBLx3Q1VxwBU3pUSQp40IjtxX00gFdHReesSg9iiRlXGDktoJeOqCr48IzFqVHkaTUAp/97Gd9PpIvO/vut291WEEvHVDAYfolIL4g0Z1etWQgLec1FBBKUAKA+8Mf6rykfy1rMdztt+46/BjzEghRYs0gFRHlNa9+TXyBwcVRXDhJXK1aDeXoAxpQ+9ptqc6kjMuP3BbLdvVQXB0XnrGkOpMyLjByWyzbzcY/9B31rr+444SqTnUmhUgFjtwW2naz8ce1ZMs37IcHv0Cma5KV5s/1ET/VmZR8nKqN3GrwM1/6/FOdSdG68zZyr2P8+JFH5nfNtd6FlOpMSo6t2si9jvHHNTZ1jgLkEnCvj/nfx8yyHpb8jO7SW/L89c5xrLHrjzz1vHBOZklK5QCN3Gq4kdoffuTRc9HzuOQSrFXsIG4W7R95ZJlymhRmOR657dRtd7M5/4UfLp8cWxQXPYso4OYZJ1jizi+fKfzG6nT4VFNc9CycGt4ZJ1ji7vFLjiwnwDkrni74VFNc9CycGt6x8zr5s8N3/219v4hP5v/23n+LMlFn86k4odRliuJcN3uV4nrmWFHeiftP//xPy5XexfCyl/918V6/879WzhHnB2kXTji2YonPkfSllrqY9Qsnnp/t5K8d+FbOxRTFRc+icOGdcYIlbo6PUq3V6aUuUxTnutmrFNczx4ryTtxrin+OPDtuObTFiP6Koy+H1Xpz3UWL2l3PnHACiHtN8UGNeGbQenPdRQu665XjF07KAlbOxZ7FFff1iR9f4LQvyYVIvnDSxo8DYzAT/6wzz6jtV+r154bLuVh5xUXPImp384wTLHFfn/OPQ6lHg5rt6Hba6R71PGwx0rypfj3Hb3MUBwfQ46/Xz39t8n/NddcMfJv0e+XPirat0j+u0pZFOYwekdOxSb8EVC+zi7V1OsUOb679/+mf/VkZ91YNL3vZy6QEV6I4181erd71zFG73nGCCyYYV3HhBKVUtvVVH2rPdbd81zNHS6ZN3DkSaslWLwOfaoqLnoVTwzvjBEvcOVLhN1anw6ea4tX28tWNN9o4Pie+cOLsUoaLemW+F/v1iW+VzIRw8+rhnHThhO+CsFNY7/H5GP2kaX8jnP+NGd8vnJQ5XsYFX6vIEZTPiM7c4U30/NfH/J/XKuXCyQ1w/meeIXfHlvWPXzjxR6A95SXxRdyU828XTnSNW9b/F+MvAMZB80kozufGXi3ieuZEVQDiXt/jj4VqD+4WHH+VplSumtooYNcJpVW2m7bYCXNLm9VTeVV0iwrda/tEdJruZ8yVsAaUyquiW4/veSjZtRylRBW3iBlzJUwg+WsyC3npqad5gw0WV1x++eKOd7yjV6b8tci/vA9kcc555y0uvPDCxY9/+MPFFrffYrHffvdf3P/+91/IX6+ZiDpt0uM/5uijF/LXdSzsZpvdbnHxJRct7nCHO4jHz872Myc6Y16svu66xYa3vpVVoaejPLlwspBbwAXVLZUX5T3/9t7FM57x9EoQ9Ft3v/vi/HPPo7xUdyovZrlgs5BnmhdnnPGVxdkiN9l4k8UD999/sc+++y7kcaRasKBUXhXd5IDPO/88qeOMxfnnnS+5vf3iXrvec3HPe91zcXvB3/72txfbbL31YsNb3Wr28//Fz3+++Hc51w033HBx4IEHer0Te47/pZO/uDj1tNMX3/r2pYs73elOi3vf+16LffbZV/RvL+RPNNvno3z5qwCL008/zWpT/fvfu2IhLxWWO+Q4wLC4zW1ua/XAKr/aLn78ox9HHvVzWS3/VW655ZaLTTfZ1KlqsM2PzvYOLSfyEsDFRRddtPjhj360uN3tbrd4wP33W8gLceW6xcrtTv7kqJzTvVH5Yp+9916cdPLJFj/CKVghfhQu4LDDj1g899nPsjJ6/vIFbvHmN7/ZvN/73vcWxx133OKC88+XdnHdYtd77rrYc489F3fZ8S5RTTkt11XRTeLLexYW975XPVZ5H8vi5C+e7H7ap/LF/nP57M8884zF6V/5yuJbl35rsfVWWy223/43Fw956ENKv6oVpPIU3xnu1f2PfvBD+8xPPe20xc9//ouFXBCR87nn4h732GmxkfazmU3e5L+Q586tnSpFXri22OEuOwT7+z/4weLnP/uZ6TrHrZaT1xvhdBvk34YbbLiQC5d+7018LuYuHPq4qnmtkZ9hoXtYqtC9tk/EWv2MuRLWgFJ5VXSL83Sv7RPRabqfMVfCGlAqr4puN3L8D9F8tGrVhovLr7jM56NydnaM6UAXiyMOP3zxrOc8RxKg61zpdy94/uJNb+J+d+zi/PMvWMhL/xb33HXXxR5Nv7PzlF2qtjn/o4/+lMxLBxh18802l3np4pX7Dypdg3zta1+7kAsxdtwIKRdOFveX+dO2mfxfffVVi0/JXHn22Wcv/vf//t9G1fIXyDx89DFHL77znf+Sfr79Yq+99lrsteee8jlGRYuf/OQnix/JeMknrF6Mv1rZdtve2eYU/vzlT697EUqUPL60kHfQaBGuzvTrO//pGHrc8cfaOP+Sl7xY6ly1kLtQFx98/wcWcuv8Ql4kunjOc5672HjjjRZfkLFb/uqSrT/0mLWszvFvfOMb48g222yz0pbs8Ea7a2S+1vFJ59pvfP3sxUYyBz3wgQ9YyF1PMq7d2uLbqdP5cyVslr+gs5B3Uyy+9a1vybx3x8W9d9tN5ph9Ft+R+fN3ZP7EJo+PyFh6uqlc3pKp1vjY3Gv7RPSa1HTsscctzjzj9MVLXvKXZtQ59gMf+MBC/lqIjJuSq2c/R3K1sRdo9vID2uLrcu66ltM2tamsVe6//wNlHbdPWausHF+rO1fWJ2ed+RXpa7JO2eL2i3vKnKD/sU7ZWtYpt5J1ytymEXStouuyW8lY/2Rdq6zF+VuuJYfyZ7sXW0qud5Ncy6O3tja6B3It9cjdPZFrPQb5K4kL+StddjgaBu3/Nre5zeJOd7yT2TX+j3/yY+svMhuN2v+WW2612HTTTYw78bF4HbI/Q+bgc885d3GxrFV+IPOnrpG1j9//AXWNnMqropvEvz7zv1cy3h8uY6b8SfRwyI88izfKWkXzoGuV46UtnStj5urV1y7kLx4t9thzj8WOd9lxdP5RgQA95K/pumq3e8f6T37kWZwsfXNttp/JZy9319hn9G1Zq9xpq60XO+yw/UL+UIWMtb8hVYzzz/Wm/BWH3C24OF36oa9Vfr7YeaedFrvI+ciFkYWOXbxx+V122nlx9jnniNs/CB3v77LDXeL8f/DDHyx+VtYqVoe+oMrmH/3NTNctG9j4y/WvCXP8Enat2j/qTeVhXAeZyvtp3/zj46IQLmyFzgDPh9mVK3asGdeLUPma16hkJY5c8Xxaj7/ixzROHK5IqufGzf/97nc/7SL2f6+99KWTaxdfn7c74p3vlD8l7H+yUQbcqAf1bbXVVsM73/muqdMd2dCsfvKTHw8b3GrDqMt/3SM6iGQKONP+8x0nfntie8dJrbaev/5yiXNRqef4h0//wwg3AiX+6utWD//6jncMcgFKynu8Ws8qexzpr/7qrwZ9hAjbVHxZ8AyPf/zj4xg2lD+hXOtZDNv/5m+arn9BwDY6f/1zlp/4xCeGp8vx4vEnuWiCcCOJ+PpXc+4hz3dyHOAdJN7rD3p98tXnhqVKiX+ve90r+XH+T35S/otEr3jFK/IvrdZ+PFev/PtXjY4PBs3tO7XdlT8VimPj9rfVlltZ20SZKcnPDWsd95NncX2rn/9UuZU69mGH1tvEtU5tP7+Qly4/73nPGzbZZJOSl9oeVm24gXH0UQDkf6r/4e4YnKPfcTJxdPT5nyx/+nKPPfaQP7td+5LnyuPr+xoe+7jHDmg7K8VHJP0F7+EPe3jz+Xq/12PT9vG+971PT4HOB6WHwW9/red/sf5VnbJp/Ic85KFet6+PRnE2kKso3/n2d1BkLOn86/mMaVOWyl/+8+/zX8miiJrPqWyPbeDfb6+9YlyQL/3zRPLor6foG9bv5EWHP//FL0q/27RpR6vsT9Fr3/yBPBqKDfGn+p9y5GLDcKsNa18azUtKWqL92R0nNvZpP/K+8fkTp99x8ks5p4/++0eHZz3rWcPtt7i9nZfOsbr94PvfH54i47s9chv1eZ2Pe9zj7PEiI8ruda/zv0JUvodQfkrflP6nd0bwpn/1hHPsY4nMidIn49n7cv7Xd/674KILh1e96lXDXe96Vzs2lbp99jOfGeQLLR3vYnjYwx42vEPmWv5scGw83inW8XBq0+P3+brU3Yw/On/qfK13567YsKUe+ZI47HQPnj/rePebOn++/uB6/BIH8+ea2l8cdyWa6cILLxxeKbm6G+dKOMdrru6k51Pja67aTc9d83cHXqvQ+du5v1TWKnruujXx1XSRrFOeQOuUDVbxOmXVoOet+cdco2V4u0bWKh/HWuU2t7H+P7tWofhysazMKfmuYI2la6PXv17XKvX8kWvEvue9p9cqTzpwYq0idU61f22nsUkubROhSHOr61+skdv2qP3P18jvpLSO559l5n9KUxweA7vjhD5nfS+UrVX+/M+HjTfFWqWu6eUHULuD9oc/qGMm12dYzhfH6ueqd/Lq+9hW3vTPdN9X/jBFrFXouLQeW6vIo/Xn6Tp3pROj/OsjcQ9/uK5V6ufP+d9hhx2GI98va5Wy1Wo9/75WqecfL4ctxIfIX3nDOaoctQ1dq3xnhbUKAs/Eb9xVrQdabUB0/ivRQGdZ+eP2x7y1zX+tL5WeVSr/ho0/8ahOOQZLVg3bHtm8R5jqxP+2oOmltIiCxqwevyRxnBq1zOYNTiXMkopDxDxFPbPeFTzDcNqp+GsdPri84hV/q0dFW6lXBEfQpv1ncusyBotdd91leM973jNcfvnlw7cuvXR4bnk5m/llAPnXf/1XqpOgVor/ZN5zzz19UJIBVC8+XHftdSk+UaX8RAVEUG++cOKDoV04QVGVzaZ/Vec37uCL0zpArho+9rGPZWYTX89f7rSx47+tLAKe8tSnDn8tL5l9przhX37JSIPtQ+WL4ve/9/2c3FK7fpne4S6+4Njuzne2hcU111xrC+BD3n6IXQxB/s8668w4pgsvuMC+jMtdG/H54PhHi5Hm/N/1rncNm2zstzrLXT72qNQxxxwzfPyoj9t5xCRUJjaNv7tcOOMTuObqq4f//M//tMUSjk/jt3/KWf80sj4P+8Z/eaPExJcbfxN7/VJSPhgRdqiy03aHBb/8ejy8593vlnZ32SC/Nlm7Q0xd0L9D2l2pIfIDkCd4PP4CL2QpLWKuHm5/7Re4Z/zhH8qCYI/0meOzYLnjXe5q54CoVXrUuP21LAL23gcXeYTZtL9rZRH6t3/7t4P8smef/5byxeof/uEfhuOOPXY4/PDDhyc84QkDL2xvs+lthre++a01ZEIe/1ppd6+QOjfe2BdUd73bXYe/+Zu/GY6SviC/9tst0Dgfzf8Tn/hEPyw9PKpvtBjR21+VQKTLL7vMvthafdLOdNH9lje/ZdCF0JVXXkW1Fdicf0ugqltXjT1LKg4R8xT1zHpX8JRiKxYv9YqYjXALOn/96xHov/r5v+Jv/8Y/szWcPz+qo+X/UPqdfknmutA+WcovqMPll12+UnJT/D33lIs6Ur+OPzovyZ2MUlY/mdlPZwWPF7N3nMTi3se/0aM6Uv3rDz5Y/spbmT9o/N1Kxnm5o1H+5Pz2dmx8zowf/ehH+7nIXu68GU468aThN7f3+SXGdTmOpz7lKcOFF15YuOW8RFwn5yl3RgwPfOADY/x98IMeJBcyv+0nUkpcn/nvDfLX9OzCmea45FnlXXfccTjxpJNk/vT3n2H81/PT///n//x/w+///u8PeqGNz/ku8uXocWJXn/7/u7/7u3KU9fQu/abM1/s9wOJp/U/j+fq2mJf8eB76UJmv5QJV2ujzb+fPl8sjXUfr/Pnxjw9PlXq57QHbO05KmlO9ptT8t5R/ecMbcq40X/JfL6CcZLm6TcqF+nRO/MEPfhBh/Nx1rbLK1iZPferTBn0hfqxVSjvTsg+ZOnep6YdS3w47lHWK/CnhT8oFEL0Q8mP5S4ZvP+SQ4TblnXVah9zlGbEVXHC+rFXkIueWd9qq5Mbbv3J9rTJ//jnXv2GPz+la5ROUa24LWmd74eTqq2St8jlZq2y/Q4pf1yoe/6c/+elwphz7G9/4Jv8BhPqf5muq/2uz+HM83iex5a6NWCNr3rFG1uNatdjA1iopOaZ4/HWZ/9s6SgZbc30vlMTXz9/HzD1zm6HPX49TeTtKX7z8ChkzsVH7V5O/j8Xboua//iCFAlVec7WsVeQ7x6029L9attVWWw//KGuVY2Ot8vi6FpC6tH++5S1vqRUoauLrxU1d/8idVXYuO0p/iLXKEUdYG/Bz8WN84hOeKINbrlI1W6vQ+V8sfwGw3S7TtYr0KdSna5W3yvHpRWesVebyb3WpE//byoPgnNl6mvNvq5ktp0TEniUVh4h5inpmvSt4bpz4chsQHRxBPX/dsqloMEI6lfbugBsyCMlACkFws6loMEKCHNIdcEM27qKSlyC42VQ0GCFBDukOuCEbd1HJSxDcbCoajJAgh3QH3JCNu6jkJQhuNhUNRkiQi3zd6/4pOr0Ocp+Rl/KlLZWrypvf9OYop4Pp17/+tVH722/f/YKjvxJdLV+ox5vXiZoh61/4kUFJBqYv218ygFdqIYg6s6loRegXPwxukH7HiRNQFvLrXz/bfgnQc9P4VkYG0d/7vd/zX15ARPBySLog3V8XlpJLfRHmeeee54zC119c5HbVOBa9u+85z3lOnA5X+0z5VRHxjz/++BKhBBSiLsb8XFYNp5zy5ajj3977Xvuz0ve6Z/trChYjpQ4THlH3r/77v4/j2nrrbeyXg6i0cA9t7qjQ+PKoDtVkUHarh4P5lzXh1cUIOCo9vr4/gBc5L5O/rmSbux0LVx57iWPU2F/72teKDzUNgzwGZRy5WdJ+nbR2l+rxInrhRPPvOVwM8qhOPl2lpXKkEMQBqImfG9b4qFseMbEvHHr3kP5pVdg5vjzCNVx5VXNhoMSxizx0rFN/OhmH9Hi5MKL1a3xtgxdddFFzHsOgn6Muojn+5EWmUqkuRPBlxb486MU+2r72ta/GOeHcDjv0sBq31NNeOPnmNy+2Worb0v2iF71I6vLcPfc5fzL8RC6w+QaWaASLszEVAniQIId0B9yQjbuo5CUIbjYVDUZIkEO6A27Ixl1U8hIEN5uKBiMkyCHdATdk4y4qeQmCm01FgxES5JDugPuf/vmfSzvyz/+zn/lcMNOHjgLFe7i8lBntDlLbP/rdH/xB7nfBkT61P/W7ptoUUpUXv/jFKc6X4i/sNFS0RlQISWfjcLW/48T6ax2LTjjhRLijxH9+/nPGve9975uOQfvKFlvczmw6P/2L/Mnkd737XcPDfvdhDW8h88QpUZ+CT3/66OBg/NW7UWJLx+3KQQcdZGVuJXes2dhSyOq9vvPf78odEXbBK+5YkHYg49Tmm28udxBsN+hf29Mvv/j8VOrFpF/84ueW8fe9//3h08//mX/8TD+6dB5xdvKd61q/ECT1yC38w7nnlfm6UM4957xh62228TrL+KvzNT7eWtMwvCrNn1vZjwLBK/Hb+VPHXz0fHB5k1JsMpAjkXPE4vpnmavvtBn3Xi16UybnadJBHN616/awe9KD9zX8POffz5dwpgt0dsvVWcu60/rFz19JEfOaznh3ziK9TiCC8uk5ZlHWchbfdkUf+m33e8phnmv+1TctjOikOK3//qrxW0R+O6JDsAHX+4XNX3F44QZ0Hvf6gFP/JTy535qZKXbEfbqQu1I21SqLK0ejFfnBU8loFGZDHoIKzpdwddFW7Ri6Vrsv873WXgjgoSAQWaT/y0JoC/V8e55Y+8aDhD3Stss2d0+eP89n/AfsPVzVrFYTAD1JY/9jdsXBSfIVPfGK9m3qXnXceLrzoosyQcofq3YTaBulY//Xt4x9iEUJ/4NHj1Pj6w+T39EInnFL7V79a1yroN++Q9ZBuRKt3x5b2f0lZqzARaxWtR/+Sl96V6FupCRVCFm8V7oAbMvzJQApBcLOpaDBCghzSHXBDNu6ikpcguNlUNBghQQ7pDrghG3dRyUsQ3Gwqmoh6xwkYkCg5IZ0yTxx7WkvVAwFATsSFySnzxLGntVQ9EAAkgk1Ip8wTx57WUvVAAJATcWFyyjxx7GktVQ8EAIlgE9Ip88RHHnCAD0hlUDr//Lxw4KEkahHw0N8tt9SXCUSe2xxFf81rXxuTgg56nz7m0yNO1Bket7xXXlCLQVrla17z6jqooRBklB0Dp6y2R2IwkKNevXDCVVx55ZXDBz7w/uFRj3rUoJMHeJB77bX3IM+F15Rw4RL67+W2TeXrQH/CiSeW+jPxTW96U6p7A3lcQ55hLTVULn/Blmcp67GCIvIxj3mM1XXizJ+v1Nsv7fgl/3r+thihNKGqc887126HxLmeGLeKg1FP+ykHPiVNYuPFiAf48Ec+Us9T4h/4pLIYmYj/j//0j5Ur+RvdcSJl9Ej0SzvORyUv3FGtvDMg1XXMMUfDlSQmeJuQ5fj88Zd6vk6ueiAASKr18MMOT7H1GPfaa8/h7G+cTaxh+NhRH6uP7khsLAre/OY3JR6y7sfqXya1TrvjZCK+3hlk+RGOSn1cB3U0FQ/4AoTz1y8lX/qS8nlbPRx33HHlVx+5ECOLG/2VyLYm/s677JoWn0980pNG7b+9cMK/4uivlM+SPx+qx73pppsO737XuyVMDRIIAJIPt8FOmSeOPa2l6oEAIJuYrDplnjj2tJaqBwKA5IANdso8cexpLVUPBADZxGTVKfPE1nPAAY9M7Zfno+ACQErAw2VxrV+2sPjXNqR3H3z9bO13lah3C9ZH5sqXH+l/b37Lm8thV64bqq7ovXJR2vqX9lmJoY/ZrLR56VpHy1XPa17bvBxW6j3hhM8Xai0LpHfIYLywYxH+baS/vO/9H0jV6xfjvfcpX8zkeHX8f6k8bsKb1vmI//aIVJ+8T4AoiFqz+OhHPdr4f/SMPyKewxtq/jviiPE4etvbbj587nNyIU0O6VNHf2q4+2/dfZB3lgyv/LtXxnF8gC6caG7+WO7yrEcetAD6iIvmctUGq4YTZb4ec1cPdb728Vfv1rP5uqbGHiG41a03ivanc79uTiGi2J4id/Tgc1O5+312N67vMhc1qC88AEUecfgRJW5t/5qXz0qulPKpT33SXv6/ueTv715Zc6WPmGh8O/eTyoW6GsUO501veiMdqz/ipneJ2lbib7vNthH/5z/zizJOwDGvHh7z2MdY+/t8s07BqSi/fSz6QLt4URlA8p6Q4dYb+R0Kevwn4iIjgpI8ELku7b9dq6DOj/w/WqtInfVHHjBwLoPcESFrldL/NX5dq1BggXqHjvrtv/Avvsh/JGCWrZG1rlLfeK3i8dd2/ue6FXvpeg7sjx95KL6tVWzMBHO13WHtY2Zdf+jx6o+nU1usq8q57733/aZow1FHHRW50fr0cR0ccS3gx37QQfpoYY2/0cayVvnilyqtoOOPOz7uUNlJ1ir6uPq4zmHYZZdd0mfzpCc/qdRQczVaq+hfANRNKNdcK2uVZz7L6pB38dldz+7Me6+t1pm9U0fWcqseCACyrZR0p8wTx57WUvVAAJAUr4VOmSeOPa2l6oEAINugpCtlQXqBKKmSMEGUgcl1aJBgZRleA6ERCTaVhAmCDJPr0CDByjK8BkIjEmwqCRMEGSbXoUGClWV4DYRGJNhUEiYIMkyuQ4MEK8vwGgiNSLCpJEwQZJhch7Z6kJdp1gFEBjq+jRMsrzo0q+LO8ugIL1LroyLCK9T3vPs9vigpA6g+JtDGL4aROEYusuiEgyvCj3jEw0ecCGQBcXw1Po5DC8pL36Q+GXjLBKV1bySTr34R1y/M+rzj1MUSjX/HO9xxeNvb/q/9mpYPgmJKMHkJrDwe4Rdc9FEdjg+mlv/+97/nV9BLXvRY/vIv/zJV/d3vfjed/8c+9tHkd2X1IC9RM96xx/5H8efzP/LII+lzWjUc+BS9eIGjgRzSe1T0nTe6hddAaMNHygURfP71r+qUUoV61MfkS7zkD7zxYqTW2b43RRcj4TXg2p23k19ASt60Xr/1V31gr5ZJTNodcivx9VESePUIwZUXplae8P3CiTMqC/Rcg7NgU+n4sMMPTXVuudWW8ijW9ybjf/rTx6SLVXrMOqFjQ+1aOI61tN98kcfj6y+JO+54l4ivvzhji7rM4Nrq1dcN+ie/I1cSn2/nV5a+N2EbufsInCPfdySqFIlaVa62X7rBU/m8P39+cMHcqXl3Dp4b1guD8kJoi7ON/MJ76qmnRu0OUENUOYrvHuGBCknM4BgggjvSPrwGQiMObCoJEwQZJtehQYKVZXgNhEYk2FQSJggyTK5DgwQry/AaCI1IsKkkTBBkmFyHBgnWMJqP9DEAr3rM5ZiHHdb0O3l8RV78GBVz6TX1OxSKMgZcO0b6LNq4jj+PeMQjhK4+sEUSHNVlhkzI7zjxL1r6qA5YDkKzGvS9ExhX9XjShRZQRR4ht6ZX3vT4r182cE4q9aL/eeedPxn/O//1nZgnTz0Vd69owBt2/rvgggvtmDD/63G99a3t44R0ouVo3y/vLOBzeaZciNUNTFOK9g25mK2PMyrf5mt3xh5l9FHaOI4y/v7lS3S+BmO1zJ9/EHHvZ++JI3eGMn9+OLgau/0yjwNA7R4mNLij1gsuOD/VF7miIgSt3GitQrUCahkd/9MPTnL+da2yesA6BTn/2FGyTkEwSKlHXkwvx7hqOPa4Y0WDA9Ij6loF9aiMH3mMVrl455u2a1+rqA9+kQR1rcLt3y9SEcFD1y/xUqfFfvKToxoHKCOPzNl7U+q65mUv40d1avw7b7tdWv+cddZXSzTUJWsVebydz7muketpaPy1mf+98hq/ngDVZSSPPzlmyloFG45S64kxs7R/PeZdd9m5UJUJtrzj5Iwz6zkJf3wn72q76ymtVabevUMHEmsViv+YRz8qxddH6LbZpt7NfeR761oFR+cFJtYq8g66dtO7sKLfy/m2axVtV75W0XEQEUQSRJ0wuQ4NEqwsw2sgNCLBppIwQZBhch0aJFhZhtdAaESCTSVhgiDD5Do0SLCyDK+B0IgEm0rCBJW8cB+sVL6B0wxYIXOheLmXBEkMUUI3EFqugLRpBqyQVEBgj4+83Hj5/+lPf+KTShmM9OVPq+VlpTgS+4RECd2Aa/brgJVbNewrj0bwBn69a8SvFOsv3FQbFxl9/l/60pdk8MUV5lWD3srpB4LaU/GkTDGulSvEPEHp7ac68NmkSoMx6//rf71Untf9pDw+UR4xsoqnavfwB8sz6Ijx4Ac/eDj66KPL/2NEfqpgl7eWX/c5fv3S7p+/XsDi89f3VJx6yqkSKMfX5yi32GKL8muZHwcz/MIK8qiP6jw55UqVE+XXJhy3HtNh8mtfbFJZ1Gdg9XD22d9wfvn824Uf+PGLQvkc/cIJvBHBAO5+QP7jVxyhRwkBL3jB8yIv8pcOciVF00eVcD4qDzro9eJBLZDyLK7c3ot4ysPFiMrwYqEbCG0ydvyKI/VpnU9/+tMLD+Ug3azP0/PFPD2ek04+qdYtdC2hz1drfT6RT7+P5RB5nlw5+K+/0tQNcSHd89rmT6FuKL+86osZsb1L7vrwC46rhu3uvJ3d2g5fK/WRKL2VW18yqO9RufTS8guNET1u/Dni0v++Kc8NXybvp9Hb8/W4d5Y/V3zhxRfVqqVYHLGB0CqnQdMMWCFzoT7/IC833vzDn4DmX29z5v6oLwpM77ayQ8Rxcmm97TzfoVD7HXgo5/KpT3taGn+17X1BXoqtrQ1MKylK6AK+/OUven/Qfibj3y67yrzUbMFPdlghq9PuOCn9Fucf7zgRepQw4Jq9dLyMv3rs+lJHbMEXg95FUecR+RLD70ZCAZHof4j/0pe+1L1SWdQn4B//Ue/mW2V8Km7whpz//As55q1Vw6bybrArf3mlxImjacOb/v7mjhN9V4dvKAc5DG+Q+Rrn+6AHPcjm52Nknv6U3KHoc7fM2586Wu5ukQvcMl9rnjH+7iN38mBrc3xY/EAEhkrE1fnzbGk7df2hX+avz/iTc7UYNpG7j/Tu2bzV+Go/+OA3RPt/8EMeLHel+Lpk6vz1Tho8rqPnv88+dd79vrwoVPOCNqbvVTtFLnq325VX/rKsU6Zfeqznr2sVfB5a54FyZ6umDUeu4KSTTohY2v8Ol7tt2i344rBcxzpSL1Lp+9jGm96JZp+vnovUGz/yNPH1aOLlvqX/xVqlVIv49Q4aXSOjvcDr5CN1rWL1+Lz9epmzE6PEX5v5H2eVysMYtVavPt7I+fYxc378e+pTnhrt3z5zOe6TvnByRAA4Qx6B9jbh5+Qvh0Vcl4e8/e3R/vT86+OB8/H1XVDof1q/3ql94YUXIezwbnnXHeJuJ+/a0Tvu6pbj650oz5W1yp10rfLEJwzf/CbWKjX+TvKYs9cn45D0V/kLgPIeuu8OeM+V/gh00cRdRIhUYyuCFbLxSvv3rcY3XczwOAgtV0DaNANWSCog8PqMP7kmOt7kQFzI5Fyv8e2OkxQ2lAD5aCY0MPE5QQYVhDBkkNyhBMjkCQ1MxIUMKghhyCC5QwmQyRMamIgLGVQQwpBBcocSIJMnNDARFzKoIIQhg+QOJUAmT2hgatyLLrooBhodIPRloLWXThQWUy2/ejjh8ycM8qeD5Z0fejtc3i6Rv5ahL5vCQKYD9Ctf+Soq7/y589e3Z3NZfS+IboiflWQ1V7vTF0b5F0AfzBXvKY9QHCWPNugLzPSikcbDryt6vPoluN1SpFAc6MvnbCKSsjwZer0aVwfgGh/np1L+XJ+HijqH/5+99wD0rSjux899gBQFVFQUCzyqCIoiGjWKLRbs5hdLVCyo+VuxJhYsUWKJUWOvoMESTbEDRhODAqLGmIgFRZSHgiiCUUSk+e73P5+Z/czM7jnf++67POQ92PPePTM7MzuzOztbvqfsmWHwj7zYGGub2d+86lWxKHJZR9ri6mOWoUNe1dEnTsSPJQsg7iCxLKg/NmicdyDbqbLxVcjPX4zEhRO7QIX3hlnSbB+24i6OLZYPe9GLJ4uADQrxowILW+xbYwe1zmb4Ia5x534eZEPAeDyZkrA/fqT0j5oAGxeB+ZXjCUfKKwPR/lg85YOSrP+XZGM6+tLaaUH3H8l5gC9dVtOKL/fk+MPdtvZo7f/sZz/3u660j31keOCHB3Xi7jrzK98TjjDbXBhfa7J2/vd//4/Zbrvtpj7AqxW/PE+eMFjiqCx5wpElchqLkvQ/oWekgBNqpGJ7wpFaeCJFSdoldFEKOKFGKrYnHKmFJ1KUpF1CF6WAE2qkYnvCkVp4IkVJ2iWE6Jo1a6rxV+ejRgfzK9kTizpWW/za+CqfI25yWpJZvvSlE6p+h7zYf6IO8LGKU9O8hDzyGXgRotaxfEuhJOsNmJ844fzzZZlbpw7mt6+12biKcvxOvrYzdeDJgvDLwmy/8lpIto982GjT5Kxf7ihfJZvak0w+o6tyuEgb1bZSbcj57xwZl1Aeln316tVaPda/TgTVX9Up439cONEcXmTU/8EPebDbCFtWf455tE8+oc7Xxezzn/8814N8y5s/zQ70VTceoipW4OZcsUvi5z//WWWfvmqyapL5H/zgeEKGdQJk/C1V/3atcuPmKVCsU14lr1aPL95EyLTxh8Lh4kW2P9rIXmSel9YqKK98ZlaorBW01AfXKlaf+fGPdWCu/yOaLwBSKyxxrWLyeFXnRWRXED9GuVbRrzEJlyVF/eXztvrFQ+jh3ytfcXilg4nlzP+UXQpm++1NHl2rUGBCCTb8ZzkJ8RXBqJVl0rKm9VfcFAzJp8rTqNQB6GuVJezjAuFm5Qkx5n27rlUsU7VWkdcPpw6qn4o/laeAJPiqDvsCNqzdtVqryN4pbaZCmQeofjn2p3Qwf222ok5lcxolr0729cKJe0AQcwJdYZyKRu/kTEvioUsxOQVlnNF4tURF6/bHTluSEr5UTE5BGWc0Xi1R0dbhf2y4ykEBP+h3221XN6J65FRrd7YilS2hrJVH/z9z9DGyR8j9ZFM+uxDhFwtE/yvT+7W1JqbC2jnydR4OjoBbyIAZV0VNvrUPakVL9V8rT9K4PikL6p0/R/yKV74i+GUiw+ZqJ30l3f1nMQvMtmDqujtcV3TYgujO8qrOi174QvmazotnL5KJFZMrvqyjf0gXnNDesTeNrMcbXm+b8KHc3k6C77777rNjjj22rutEmXB15NOflvdIMYnhT/JOLUbux31uSr3P46slUpwoUVReFyNyBZ7+zAs/k7czvriS29/v4oQqx9749290fdD7InlVB8an7DMTbeFiytGIO3l0c0HuUtf+Wqje62ZeQEzw7lfxT34XV3Uv037WeWS58029T3/6MzJ7EtdPNZb2QdmxA3xrP+9UDxnsXcD6U+mBdz3Q/C182D8e+wEs40A8MT6g+3nPe57bv8lNb+L+fKp8pjAfrX3wKlrqf8ynixGpK/1DCPvPe95zKeb2TZ+TK6SyVTgVbcJ+pWCUCGuKySkoI+G6rt2+esD8Vby2nv7/+tfxlGH8iNht192W7X994kT7kI1Lz8AFy3XYR7/L8feSlxzmjaw1kFOpidPx1TgtYxn/8GpmljG8UNZhn0r1wknq/9Cve0GImqyb8oA76Gdm6asFf+KktX+q7AeRx/88Vmd9F19ysXy6dofK//8se6Zk+9iwFmXD/hkX+J5bxaKADTn//ay8qsp4wJdi2qOqa2Fyc1i2a3vhxHVI2+DCHOX++M4yX78I87XN0S8EfOGLZgoF5zxN/tvkyxm0fz/ZE83nOZn/zz3vXDMzp/1P/cGpbhf2o01MIzIrJqegeMkdMR5fl5G4L/G/y+pdTGaOfdzT3uG619G2hH28poQ1ynLrzwvrtG9Pi4Z9tBn07i4/Mo+RJ3aXe+B1ZLY3IPY4aeuvaxXh8QmY8+jrYoRl0qTUf7nx/5mjZZ0EveXvYfKqDoy39qFXL5yk9c+LD2vGDeaa8D+egjj6mKNnD5CYwc06+Eltlv7PNbLaTfaXM/+jbFbeUuoJ+5DhgQsn2f7Tnx6vq7T2mUfXKsn/ecyEDPLhtSLXK/2h+gJgUXTXA21TYsodf8LxJbcJzLOva5VkP68ZbqpfdrLx/2lYq6yj/qUoCRS/lZLkT4qznIwPXMAz6cgDRRXtctqHslp7Kmprq7C6fTiieK34v7lwklzqqCPqxpxiGyotM4rDAyzJDDEWDhTP4ojK5VS3b65Tn2THGDmdl2ROy3kWR1Qup6b8j83DOBAA3kI2d0yNmWxNoaEdG4JhMt1jd7tzvO21tp1hU9LD5Acg9NrjdXbhRHNF1gnFxsRCzgarMiHLxIJ9EOJIShx1RMVyCp/xtbrawIpFRr5wgosyBx10v/CH8GH/RjfaaXbWT88Ks44l7YKeJjvS58H1wx/+cHQLyTPlf1dVIaEXn7r804f+v2rhy/aCLexHsWbNmtRkkRcqkdK7OGVSRt6Hl8dfweexyy67uG82k8cf13Xgwkmuq35OUTPV9j9dFiNsf1w4UYlaTHPaHielbaScLz5s+omTKJu9J4u4wyaBqNt2215rhqcuXnqYxR19dbhcsJvy//guDh6nnShcGE1YknM07nzTNp84mbJPZXity+St/k+Qrym1x7fktSLqBLxD3nCt2L+eXriLxd+/fuxjrmZ97OM1Gxw/PuMMsRn979BDD3V9lZ9S/ZNA5Una34uPv6aYZL0WZNPFoyVmlne40VSURBMlOUX7SsuMkbElmUk6yTnqiMrlVLdvrlOfZMckj1ZPX0l87JP2+0liBU1KBK1+BEj/ePoznrHO9o9+Z33mCU84RHQnvWOjehcd8ZrHv99ewHmpzptTS7W/fY4YZYjx70TfnHuiEEJq+7o9cZItWk10rE59bT98Aa0WcwMvlAsF7IuYH+/V7DuA1/DAf/KTnyx5khJBN/T89/Nzfl58bOMPPn9eH7V94y3O5r2q0/r/tB+eFnWVOn34Qx8u6pPe2mCTCjl80poxgX3SeISEeKskAE79vjyxmdokLpwgZ85FTVMw5H5ens5hTK7exXwVErV9fOkv24+6r8w+XqfDviO07zEkfgX+wAc+wF+p0DLlgqWq6U2ekgfleyS+qtMcXKvAlvva9TmiuZBabvzb07HR/x4+8Uozi4ILJ7muL/anY8f2mee3ZY282+57qE+23U7WyLJWwU2S8NeC3Fx8BbNUcDnzfxs7uTQ5/hhi4ydOcJMn56qKoAmOmay/rlU8iyHtRR68Htja36G5SPuxf+VaxZWNjQulta9rFclyhqxVWCb489BnPauqSWt/UrkSwz6+spXXP9FO8pqQxCe+FlVnibyg59RK7Jvyeeek3VFHNFNOXZ3tNxdOmoaZ59/kwuzIpcRNLqQVi6RnzaSMu4Ajxl1axoVLwIW0YpF0wUzKuAs4YtylZVz4amO//eGIV0NwtN7S9ITzcKf/yPcdWT7bK0+FyOcJD5Or77+Rb97jwBcI8mCWX5lQgTknmNINVDmJCtxcniSYfuLElEwUL2mXr+roqzo2kXPR8KAHPijJmM1ddo7NNVn2A+QVgovTO+TMlG3ivV4bWG3yrb+2YJJZnjqmYJbGJ9+e8cxnhh/Tggv2MAFjA8B5uu2pj1JvkW+fOMGGovqpt+Jr1PnX558vxQqNikWyelUHvsTCL7G9SkfLrun0Icr6sDmPvyIDH3+lfH78tbWPOMBj9dj3BfYRd3hChZ+C0y9fFD9BH54mwtGWsY1/PlJqciHd2ldljT5KY3M3jYNiX+98F8uUYX5CvFtssWM/yF4gTyrl0iLft7nhWtGLzxFnfRfJhUbTURZ/IvfmN72pmDDJLE/bgAcffHBlH58zhn08Bp3LlTeORb6sL+Pg1UfY5xMn1Hub29zGbYB2HXllzT4JGhoVi6SrzqSMu4Ajxl1axoVHraX5JjJnUsZDEzHjLi1DWfo1pBWLpAtmUsZdwBHjLi3jwlda/asnwCQWbpLno4nCZ9KR/jlii//6VZ3p+j/6UdHvEHsv1H63tP+xASFjF3AzPHEi41EuS3iS2LR9cutXdWysjs1eS3kaA9iwXMtRxoMLL5p+VecH6bVKyNc/0lkCg2ecsSY9KYofB7Lf0Y9+pHW7UC4O4UkTjLf8pHEu0oae//D6CecClHvXcjEglzjbJ/5P8pRM9ssT0ld1KAMdUV7zd56vTS6kFYukFwGkC+UzyHn+xA+t88/nJ0kpapmpAk+caBmlXlNtUktPtz80U1+9QevCbPVqu3BC65Sk/Df+h2uVcd1DL6XXbR95Lr70ktkzZZ3iF/9KXLIN8Tnpqf1Ioox8VcfKBL/wwomWRE66VpE9uLLvzv8N1iqsYdZm+HLjn68Vs7wP96+sjOvPtYqWQ+pZbWTfFkHXKkdUa2TcFKrWKiUOoE9v8qiO2v/rmv9Dui0A0sZtZcYXTuz1xlZa0yXzox/TjJnylFZYUFSeOPm2X5iDP/kkL+3jpmhuQ+Bv8rXKuLTZ/sGPjbUK8mF/Ehzf8LWKxccD5GJdHGaZ9oM+jVHaLpxYPKIeulYpcQ3b28taBRdgcSytmxqn7bXUVlrTEwYyKeOtPpZuaZnIZXIhrVgkXTCTMu4Cjhh3aRkXHkWr5pvInEkZD02G+YWT+UKcwJOEo440eufRIUYeoWWtU1ldt2++SR5y1JHssOTjhqxJ5iE0mTqV862f/88+u7wbq4PBwgw79dcHLRGGfVxBz3cZMFmffPI3q6C3TzdiEzQbzA4//JW1ek3VuimAO1gYnPi0wvXlnWse0znAnV9/7MNiumLifdCDHqx5qBcQry9tueVWPrDT/sHp04tT9vkuORcO+EywHi7siNH9PI8OAfIWdS+WnfWiDstvkBP9W97yFpW3HMwni5FP4eJFkZV21g3X3PZsdtZZZ5ZJrugVmand31OW2fd1MR72YzFe+98ef432n9qYlnWc3nCN9SCcyVNHF84eKne36Gc8wn3yySeX4pl935S4xPXhsrcO7eR66AQvMWYX0qY2XKVdQstdp7JGXEhMm1SK/ac/ja/qzMu1OHvOc5/j9UFZ/tF3g2ce2akeixGUtbSlbbiWbc9m17nOtZ0P2ef/5fNTtamrzgOB1v7LXvZyFTrllO96/4PdvWR/AxzzNIFjvCThqCGx4ZrV5Zvf/N/ZLqt3iXJL/W95y1uVp8uYmVDNXy77piGfa92ZEzWtZepUzrHu+mdpw+dru7rZP/vsn3pfRPxe85rbFHfVPqpTJhI/Amxc4pNeESzjXM95znOtT5VxAuNGfTAPIZ6s+KHlKf3wBrIZZhwra3/bHDbGX/Q12xyWdgnNElI77HA9LQfHfz5xYpIhf+r3f1CN/7fGEyfVEbIgP+jBD4q+KG2A1zcQh/jRizbZf//bhkuTng09//FiAOff1at3dWt1iZ2s5cQnmW2clPFF2vXxj3t8EahzWXnD5w9sbqJYJuYhrKlI/fSss8Ke+Adth03H/fCsjqQbD2a//hyx5xSEeQiNV6dmzZdtcOFkl1Diwo7o1//MR2YfnwqePpiHcNp+5JVXg2WvkJ13vpn5pPSrvP54y1veGuKKhW7doLXkQflivWAyZ6mvS5uV/vcd2TQ9jnH/W278c48TrgXqjeyjjLD1Bv2qjpUD/S82h63t//bC38oTw3gSx2R3laemTpYbIPnQtUqqM1/VMRnaXd78j3ixHMwnWhx1xM3r642I2WLfx0yXYB7C2czHzFInbG4bh9mvvwC0IE/H3kFEQgfkr3MdvCpW2lLsP//5fxlqHGMewrH9l7/sZSp9yimnlH5oOvfa09YqYTZ0uHpF5tGxx8le1frnm9/8pny1cHVl51a3vGV6En79/G/lmG8/fFbL1Klcm27ffJM8JKhcOEkE+MuTjmQvVngtQQcXkYopiZKWi6XN0RA86UgjH8laotuv/NEmSvqK9v+ll15WBgFM+Pb3+8suXVb757vk21xz6/TjNdq8/bpJTArrbv+v+Vd1bCDcWz99VjnKy5mQMN5g9jniqCfqaxdORLBSuzh797veLZMJ7PLPJpfXv/7vaq2eb1Ef4c53nq4hu/D/+te/dnkXVcq4/vjU2Zo1a4QrkkW4bf/fyd0t+BAXuNheCmXiwWOrP/zhD90eEV2MlLaFrD1xEvbxJM4Wm2/h+rAQ+Izsi4Kjtc+Cje6Y3WZ6p3rexWFZY4+TsG+GJp44kSeXpuxXcbf1NiXuau/aBbtoa/isljD7cTHCZPlVHQpP2dfy8uRKHYnNYYvPuRgJCWSu6/9YuShHH8H/3ymfLcz220d19S6OKzVk//33dz3Qd395lYuHiyqhtv+4xz7O88H+v/zzv6jUhRdeJHTrf1hc4TUuXND0w5U64qwWoQQ3XLP6yk71Evdf++rX/LPMsA/eI+VrQ7n+po9ainZPOtKa9XQtUdef7e02inC37+4rSO3F8FtDb7NJupaY9j82I2X7c/zFK5Zx1FpC6fxX5Ji3zmn2H5viHjH3ne98x0pahKfa/2tfw1d1YmzZe298krPWTpsZ1hJ1/eNVnXgF6ET5Gs6UfercIb+WJ31z3uawuh9VKq9d5K7t5+J/Xr4oo/Ur89+O8ilytMtdyp4E73rXu1iEVO0NP//9/Jxzythj44+9qlN7Mcod9Lmv6kSpBVuc4Qk9vBrItoz5WnQVdUv5/8eyseea09fok6xbXGNz1wN9eNKyPaKEeH3ENuzl2BqvuiLX8uyr/qL05+fE5rCwv3p1+8SJV0mz4Qs37F+Qv4Z8NSfWKuu2r3VfsyYpLQVR7bPZRbJRMebcbdp1itiydcppdLHlKNn1tWKR0TaR+Hv4wx9ZNBrAzS88Xar8Ioc1RipIJY/EcuOfaxWOP/gU8rz2f+Mb31CV4cWHTW8O+6j0JCm+CtXe4EH52rXK4Xg6tvgj21/O/A998466haz/+4WT4ktbq4jkhH3Tuzh77GNjrYJ28DEzGW5f1cG6qrV/G12rlLWF6ImnWZe2/7jHxVoF9v/lX7BWWdTxL8cF4iyvVVr7VbpNlPRee+GrOrH+Qdxj/G9j8JF/Pn6lLLlD0cqElLdKt4mSzu1v+ipBBBz7TwAAQABJREFUb6eEtGY9Xee8etn3J05aR4UbAnOPEak9R2rdgEKlmMF89ixJymhhNbAsrTgVN4yWzHS2TFpkrSlhNbCQLVidxdktmWmD+exZBKGU0cJqYFla8TqLs1sy0wbz2bMIQimjhdXAsrTidRZn7ySPQ8eAszD7xbnnJvFp+8cd958pD/bNkI20Sq5s5h8/Ep+FxYQUF06KuIAsDyrTx8rn8TiJYfA66KCDIpNLGSlqHVgSVnRts8cJdGOPE9qjPNMHywSR7cNHGIzxNZcopeWi1ZvcRHwpi1gOtu945zuots3idNi74IILZnvsscfs7ne/u5fnl/I54jvJJ+zww5JlYqafyp4r97nPfao2gM2Xv/xlI1leOGFd2idOoHvXXXctumySwOPNZjOfaT0v/GyRs59/4o+eMNm84Rrsx4WT0MW62eOvZh+yujlsVYrZ7LjjjqvqzNeOwqpheGJDY1rbosQdDYVp3Rw2Yh+PlHLDVTbXdP3Jpaps314ZiMmWF05GDVMywwLf20UbXmPLLeVrQZdVNYfMvNeKoIb2H/3oR1Xxd40trzHDZ631mKi/5Z3N7nnPe1btf+r3v+/28TUrxjN8xU8WmlKqpnKWZDb7gXx9ZOdddp7hs+J6FBE+ccJ4xJcFcLTtD1t/LwtUHNSuiQlKWA0sZAs2VjKhKWyZeD5njbWysBpYlp40VARqTd0+5iPGBuC5Mh+1PgovmRPhdf8RUMZf73dsiLGS1O8GecoQ/e73bsvE89kU4bOteczAvBStHhjNOpywDx7I8apOjH+6OWwpzTirbO4pn9HM5bDPEY/t84cjfdo+cdLqxuu3e8pcBN3M8+rXvFbTeFXHXjGoc9Hqhpz/sDks7WP8Wb06XwyYtg9v2uawMf765rB1Fm2Wm9zkJpUP3/mOdxaPM8IsU5v1ggt+4/M1FOGijo+REn+Pf/wTVH91Skp03w3xL9svntikXcJp++RS/8/0qzpR5/x0DmW8YoVw06bu73iHrVVYzGyZNGS1tcruulZBGvPLHXSd8hNJMRLAkadx5AkyrlNyW7785S9XPk5ZN27Y5Lh7ZFlXZplYq5j/0L5hNTAaWG7824WT6H+6OWwpXbYPve2FE2ysa0fY17VKWX+gTo+UjW6njg/LU26MA8TQK+SCEw6zGeflzP+t/yt7bSWEGU/pmS+fIRvZUywssyymLdYqNmbiwnbU2jC/IVXqrzekqLgU6jHpohLqjwuXiCWKzbMfaxUr8/dlvyDa33FHe3WbfRGvFflBxU4wBGRdq+xsaxWKAe4tT5xY21hcYC2Oo21/yOCJ6fX1P3TRHnAcTBvMZ+PbmVJMMU1PZNmCU6RhtWSmDeZzzkgpo4XVwLK04nUWZ7dkpg3ms2cRhFJGC6uBZWm9cMKrUJVI0pNNKZ54ajAUZN1WmEq2YZdkZE/CI9QIek68bl+cEQ5sHAxeQ5pIRvYkPEKNoOfEm/L/Y5u7bt+VR96WOmD/BS94oQ/2eJT25eXRfuRLlmfvl703dBArA+gr9Wp6KVBVrpIz0bC5akwow+y1r32tFmul9cdds6wPeN4c1nyjJvSEuya3utWtRnm23172YPgB7pikwhb0rvpVk1hwXkd27f/5z84JpcgTFfD68CmKzx77WZfl1xswueWD2S+Tuy/YpI+PMmNhcvDBj6n8j2LhVR27mGMTzSOwWzyOUmagWPy7b+ROD35EnCN3/KYO2Ned6sUe8sD+rfeLJ05MrZ3x2G5u/wfIV2/a+tPGYfLer5ZB73QOs8PwVZ3meMELXlDKaROZL8BSXYAe9Q/vL3Ioo2y45q+ICZcOFNQWI2XBKXXBviHrOiJ7Y7RkxKs6sUjMr+pAoLYPChYeN7whLk6YPx8/sTEs5OqF08Ls9hObw37+859XPdn+u94pnw31Y2wf9dEfPMX+bvKFHTyFxAN3glg2a+9VssfBN6Qm0/UH9cLfXjDbd99bzq4tj+P+6v/iqSvU/+blc6bU+eMfY8GNY3H2YHl1zu6Emi82ly9pHffFLxq7nNflfytVOqdiTvk/lItgJRucjHX75o2l2p+OVHdWPoWPC6GiQyd4s9ljmzuK3/3ud7P7U/akQND3yzjJmEL8+4+AJGY2jHCZPG15wxveUPJY/3/84x+n7MrYROJDsoko88CezkvJhqHpnHjZfo61V8unW1l2xr+9qjMuAN1nX7Cx8Rd5MV/hSJYVx6sKefz3sboqV8lZaH//938f5SnjAmw86UlPmut/aNiQ8x9f1aFfVpcLJ6z/vPj7yEc/Wo2/j3zEn6No5ZAKhgIp712rtrzudWS+lgs2Sx1r1878E7K4iIbjoPseVNncSi5YY/5Ud1Z+Nvv8MW91W5B9ZzB/gqfqljxF8UP45z9r9oNZvYvqUIkQKzaMcNcDUXebHwGvK1/ZwSazSx3Yy+dRui/QwuzYzx6rolynoP/hMO1x/v1la2dP/osnV/GEPbXGdZXXfD716Wr8540RVVxO9xVf5/631VayVkltFpbNxHLj/zOfEdvwR1l/PDA9rZntw/+Hpc3nsf6JzWFZrUVZI6e1CtbIerFIMkcDqtp/+IejKt9M3VyE4HLmf8i19a/9XNu3Cyex/nn6054OFXOPyy67bIYn0Oj/qYuSsP9t+aqOxZbFV3xVJ+x//nO2VqEuyL/rne+aaxsMxN+Nb4yLnVbm3eSrTfhaJuuI9SX0cP2zIF8t+u+vfyPpDPvMg1e/9913H33N+Vf/96skG58jZl30JiYqKIe90hjjL776+cUvHrde/jdNPKNsxOfDCJ8kPEKNoOfEUwOhoDEigpVswy7JyJ6ER6gR9Jx4V5Z9vXAyWZ2qcJDIhDLFZJIrSdOP8x0palLa8zXISCQTun31RnaJu+/K9/8HPvhBH2gwQBxxxHu9dBZG44Jj4yUOThjEHv2oR0UexSxPTB42mL30ZS8dybn2Bjn88MOrQfCrX/1KkzclPS9pmWA+xqCoC0hcxMGf1PVef3LP6CmexZEZdp+/9vbbqSwHT0C8NvQb3UC12CtZsGFplgN+u9sdYE/xhFoWcnaZbKj2GFlEwJd3ufNdKmVckDx47vvHM328dgt5zNZsLujmbKYkjH384x9TvraX1PvPHvZnbh8dHJL/Ko87tuXWzRVVTehixi/J15hy++8hP7brw/KccMIJrhfyN77JjWe44BOH2V8rP9Rx0SKX4TnPec5o/PEf8VIP6LOnH0RbU8QXvOCvKl0veanEXSODMnxL9uSBTdbFX9VhATXPREbyCZNIexcHn/ibr2Zx9iG5QEj7uIt8rnzu1I+UEY/5un+k/neQx1/9KPaxsNi93C2m7B//8Z2WtF8/2TXMjj32GFeLjJ/7t8+GXfEV9G4nfaJ6HL3YR0Y8sYZd9CH3+te/XihljCsyN99bdqov/Q/1PqPcxYEt3HGq348fZjve4PqzM888M8o0hSX7xs6E2n6dvfBA9CyOFFpK15kjNRLJhG5fvZFd4p4b+/+DH/iAxVuJkSPzfOT5GkR0+xMnJUbxVTc7pv2PC/Psd9fTfmdPWqZAmGx/nZfK+IMY/6q8Zhaxw3Llyk7bz2V7afVlDetjn/v85ybt08IOclHe+rj9iLAnTsgN+9/97ndUTusq5a730xj7H0Z/9atfyWsW20S+4tP/+i+p69RRzG3I+Y8XTthGu+xiFwOmzGf/8wlL880gT9Pdw9jhElfxviOPGNUR8zWectJD80RG3HzRjbTFjwfehfP1TF5txL4q8QMUtpeKvy8d/6XKLp7wiTqEvbkF9xoYggsn2f4u8qRfHNPxh7aibwlvJ5vgz5t/vO5SN7y2xeMXvzhX6+KvPSsj1UEq8St5Zfkam2/h669nHvrMkr2Ov49//OOqS9tOfPxn/w9rlaxLfI21ivBYZtT76fJ1mvqIPF/60hfD15Jvjz3F136E/ROOj7UK7OtaRS6utvZxUwEXArL95zz72a7RkEV9TZYxCPgYueA0PhZnfyUXWELXMHvpS18SYloNq8ty5v8mY0lOtz+Y73+frVdp31/VoaJkH6QPf/hD1j7iRx0zq37CTDN5JcnWVaz/7bFWiSZRQaxV9tzDvoZI+/gsdiXY2Pen/Ur7H3usXbyjbjwNTpuE28nHEz6N17ka+yjEeef+QtryDur/1/8d1io4QhCfI2bZoA+v6vCwtcrOYo/9fkE3AP6JrlVCB/Q11aAKteWSY6QUxRkpX4OORDKh208XTswxtXuKMzNRSYmgqKVxTpwqRTph0ZyAcTLfcUcongiKWhrnxKlSpBNSU0DjZL7jjlA6ERS1NM6JU6VIJ6SmgMbJfMcdoXQiKGppnBOnSpFOSE0BjZP5jjtC6URQ1NI44w+71+MVFA4Qj5NXVEyiLR/1zWbPkokv//jZ4hpbzP5bdmrXQzJfcvHFszfLZqXXSO+jQv8TD3miiCzKBYNLdTKlHc9oiNq/853v7IPgdeULAu0Pbojm/I47UpQVqdNOkztvUgb8sa77yddg4oiMxAA/8YlP2OAogzUHSeS/v2z+in1TeEAWj3qv3mUXs1MGd9i7mWyW9k55bcfu6Czq3agPfuhDs31lYynw8YTHf33tq1SlJeaFk1W4ai53+OvDSoizPlaNssnfRz/y0dopwn9/+doE63yf+9xbVSGvaZFyy1X73fRx4/ANyqWfdOMlZs1lG9Htf5tbV+2PL0t8T97bdsVF9uyf/tR8Ibpo/yUvOczFYB+b9N373vd2PuUOPPBAL19RNzv00GepPpWR+m4u8fU/xTfQdcnFl8ze8qY32zuoyf/4jCaOSy+5VC422V0FyB/fXNi5pbQHDvqFUInVyTiZT/z9R5Y738X+Ew85JHKqUOS9SPoJn2pCnY444giVpS5CEE84EQs7m6Qhu688DZX5xD//b5+bIWby4hLv/Ydik8QZf3zaCDof+tCHhljhY4Gz9963KO0d9hdkvxM8HfPBoz4wO+V7p8y+KXeZXve3r5vdSO/i2w80vHIU3jR0t9129bZGjJ1++ulqk3JfkwukaFfw+IenV/7vV+WVI5WOOpTM4QtjkSwwERSNvIlTyZFOmJQVNHSQ57KOjDilKJG3Fo0UMUJqCmiczHfcEUongqKWxjlxqhTphNQU0DiZ77gjlE4ERS2Nc+J46hyZj1bJfMS2xzv14yN0kIdX5LgJI/I+8UnW5ysrmm1R56dbpqcJj5Af0FmO5SKkDUDMSxx/8EShxXjKPcqUCIpaGmdynvrUp3qfUN1S/g/IBSTys33DF2dbyFMNqCfL8ht5ykuPJtOXTzwx5EQeGx7iyPaVkKyBh6cZ2QYYd9oLLtRheU3fhpz/8HnRbB8Xa+OwSuaqEv+67o1m8yF8gw2G8eUbPUTove99z+wdb3+71vZSGZ9Wr15d+Qd5birzNV5dwetCOM455xdygftDumk1yoQnSr4qF5FoE/Pnah3X+CPK7D8bn0RdXKs69CQZsN7av/mSGNZfp37/ey5HvYTOcMQ45KuvdL4x+9ttN/YV7TPqUPddSt09jqRuO9/sZrO3Y60idYd+PDmDp6xueSusVexpVK5VwD9XfKNzjcwH38BczEI1ZeXrX7D1kY98RMVqUWxAXD+teW95Hbk9cJNF1ypSXy+34Ic++1mjoMaTKLeRNZ7LiW3s08VXUbP9s+X15ywH/DC5oJkPvBbia5Vk/0B50th1FeTQZx3qfRq6sBZW/xRJzP1venNZqwif/f5JTzxEJXChCheceCx3/o+CMKeXrLSNpXE+snlKj+sk5iCEJnwJ51a32s99dIRceMt8xwXRG2Yaj9ZGiB09VMgkcf63z5W1Sqr/P//TR12vSVpWnA866H5u/yF/+lCXA4+q/Us4yT7WQ4fIWuUDWKvIhvffkptQr/3bv53d8EY3Un14VU73doQiOWh3V3miBW2n6y7Rd/rpa4SLwyS+8tWv2m8cxILaW5jtu88+8Xp0kg3U8uJMO8pLKdIJjZ/Pxsl8xx2hfCIoammcE6dKkU5ITQGNk/mOO0LpRFDU0jgnTpUinZCaAhon8x13hNKL2Bw2GWsF0g8cZzlCJQW29Mk0iQXWIBWEOimfWEGqC9DSJ9MkFliDZKTbVw9cjva3zzLapKt3dtTX8/3/KTxSqQOKDYzArymbXx0iPxIf85jH2GdiGz5kcIHgcfIo9r773HL2XPmigR80VQi//tX5s83lxzht4BPH7GUu6kjJtI76v+IVr3B9GAixyMZmrvoFmZEu6CRR7gr8lT3BgMcyrUzmK/zoxDu/ehT7H5QneFjuKWgbyMaXZjbbbPPZJz/xSdPBs5j+xS+wOZ7Zu+1tbzv7v1/aD0eWCsXDY+x4nQFyWPBwg7cs85SnPMX1QO5GMllcdunvacmgZDjmmGPTpldWP8jjDtxL5YmNf/roP81eJY+UX0++IqETSSkby4hHFXeW90TRVtn+He94x8o+5O9y5wPlLtHT9W4gn5jBj2Pqon7kfeYznzG76KKLtDk+/elPFZkoHzbK1biTd2ZvgM8TN+VC2uNu331lV3g+ycJ3ValrYbb99tvrJw+tAqxFgTWI8KAnS/sfddRRVRmwe/xXvtI8LSW6sBi/053uJLJmH+/F4yKFHgoKXrz5xje80V/LQp7ttytlbewj+erXvKYqA+r/xfzKC1TLX/585K677jb7yU/KazONfewmv92221X2p/xM2o122kn35olAsEL+9rflk6al/0Ee/mrl3vD6N1blh9z+t9l/9pMzfxKidI+pFh1BcMwRChXY0ifTJBZYg1GZu336K7kmSHUDtPQm/ehHPcbb3++eFxkXdcRUH6WvhXJ8ls9ao9+d1PQ7EcUPwTvd8Y91/Edcab9b2yjTJGkFCjj//F/L54cx3lqf1XmJNVth/F100cWznW60k+u0PrQge0jcrWgO+yAghVdFTS7q+0m5wM+j5FDhv5OnvrIs9in57QXlQsIoQyGIAnztKud7x9vfocysm9mNYZwNNf/FWB91xJ3sddn/5S9/Wdoo8mGfDTxhhFeNUKf/7y+e4kWP8nIeQL7AY86XH7cy/+Oi3ic/+akU5KbqmKOPTvNn2MZTHDp/ysXrV73qVbPrl/kzLvKZLcyDnD+tkqxpgTWo7LdP2aCOGLPDWV5dQ4quD34QT3exroRR/7zZPXRiTfaJT5a1StHBGzzg6zol7VNB+9U6RdYIWN9VR9EVaxUri65VLru0KjNEda2CJ1jEZv6jr7HPjfr6elir1DJIb7H5NWa77LyLrlVUebEfa5XwxV3ucufZ0+SJlnve456yD4fZrNcqpv+O8uQC5lOsVaDuU5/iWiXsx1rlYHmScsdSNtoyqGuVxz5eXx/RtUpx4vrM/6U662z/o476BylD2Pe1iiqgFrvRd6c73snn/+kxk/L86lDU29dV1pJVuV7ja5Wo/xeP+5JIUt+iroue+Yxnun1dq/grvkVpGX8R99vKhcOpdg9arL932unGsx+fYXuXsHgw/Vu5EK2fX0/xc9QHjvJSEXnDG94wsoWNb3UtxSq4Ypa1IbRymiaxwBqEe6hqhfOPZqcp1wWExNowqc72PM4JVpAoZbClT6ZJLLAGUTxqXqL+8cRJEmI+wMZUZjlOmZB2liMqE4KFngjdvvsqI/QQYeYRD15g5BEqZ8ROhA3sf1yB17vUGCRkcfBj/IBK5qxcQcBrKtjMlFfJY0Cy/EjjHUDo5c71sVDAAmRB7jrkz5ix5gb/9WMf88EIE4m9dxz284+UnJMShKfIhYW3ve3ts4c97GG+UK7KKuXcVn4Q4s4Q7vavWbNG1Wl+KhEKHs+82934TnBMCKg//IDHLTFgMwveW8XGnOYfmxBy/VmGa13rmrqgy3UADj2/wJ0ctEf5u/GNd5phkYcv5+BxWjyyunqXXZSPDUW/Kle/aR+Pz+KqPx4B56eVs308WfA5+XrCmfJDNB//LI8c4+kR2pwH732ve6lMbn8sCrCL/NfSnTjoPu6443Tjr2y/1rsww6Ouxx9/vNn1i1ODTlwoq12cWpSNCW0TXebP9klD3H0PcbcKkyP8F/5H3L1P7mph/wB8Gnvbba+l/Kznvgfdd3b0McfMfiifHa0PeleoS/S/83553myfffYtdyDMPvrW3e56N11A//Ur/nr2kIc8RN8pR5m3kwsgH5anj/KhluSEeltZDy9ljXhA3oPue19ZREpZf/gjzc4SAr71bW+dbbP11l7/rbfaWvf0ebvcSf1budtyYPlKBvzz5498pPwojMUs7ecyfeEL/zHbWvVFGbLf6P8999pz9k3/PLRpQN844r1HyCs8d7A2LjGNPHvuudfs7W972+x73/ve7JJLLtEMP5Av9+BLD9RJiAUY3qvGj0TckWuPXP+WxzRlJgY4ilg/CsFCT4Ql2h/CSdJ1EgleYOQRKmfEToSruH3031Xpiyd2QW/p+p93nvQ7uTCK+UvjReBm0u/uWvrdK+XCOfrddWQvB/DxCPeHPvRBuryC8/z/sXZeavaEYAkJK6UlAR6+9oAfrX/zN4fP8J5+1Y9YfinjreXJhNe97nX65Q1cSMdFATyufnv5MY465Hx4Wg7zJjZbho0fnf4j+Trcu+Ri8LWLLH8g2Rcs/vM/vzDD51KnDi2/nO4kj87DDn5AnH9+3P1eavyDvssz/530lZN0g8wd5ClTbUexT7h69erZu9/9bnk682txwyJVQMst6Yc97OHV+Gu+svrvf9v9yxeIKD2TO9EyX8vmlGon+Z92CeEHXICZV/9/yvOn6LH2ifmHevDUAnGMv5DD/IlXT33+jOKVGiZC6f8nffkk3XD/uvKqmesr5ceGte8RX+HrhH5zRzSFFsNw4XpL1h2+nlP/aq2Sxp984QRlwCsuH/iArFN+9EN5FeLc2T/LOgU34zD/Yy33FX/l2uzjtagTZa3yEl2rbDmyb2uVz9VrFbGPtcrm6cm0tv6oB/x6r+LrvP4wXz9KvpAST/nCybpWKU9yuT74JP09G2sVeVUZtNz/EBsPwVpFfnCjZr+RDYTtKZtx+1Pfgx78QHni9lQZ60xG6cX/KO/7fa2y/vM/6jN15PY/77xzdS+yXI9V8kTOgbL3z0tf8tLZX5cx87rltUA8yfQhrFVS+9MGYgwbWh8uY9q1rrXtqP9hHyCsVU6T9asdUZK3vfWtsbaQ+m+51VazBz3wQfrUl61VDvQ2eCTWKulJHNoHpMb/+MIXZltvs3Vpn/n+31Ne2zq5+Tz0N+Xi7Hve+179hDLbiXDPPfeU3xOyVpF9ILlWwVi+xUT/2f7a28/wkQVcZM1rFZYxSptrYLjKhGARSIQJ/0OIEoQlYwWCF1glQD0jdiJsYvbjwkmuaaqPktu0y44ZY4oLByJCcs0v0i3Wstq0y48ZY4oLByJC3f4SnmpZbdo9OWZkCnag5gCBq6h+iNCU/zHh3Z77UpTBHvn32H0PfeeX+Q855Amit0wMIretDKq6+BCBbJ/ygHhkkGV56lOeOmnf5VslKf1Xf8m9LsI+J1BC2sFE+1YZwEdHqf85Pz9Hn9ZQea9v6PW7j8X+iV8+cXaPu99DFwu5/rCLz5k97alPm+GTi/MOLkgOuO0B8lrGvvIjIF3QSPb/5J5/MvvBaT8INWIfG6vWC6AoJ+sNiEVFe3z961/XC014fDj7BvJYuL5dHnP+/qnfVx7ukt3/fvef4etJ2GRLj+R/6sYdItw5tsWMlQX6biZ3nz4nj2visEdR8YNmO32CBBudXiJ7wCBIcvxp3MnGqFa2qBcuYGExxaA6RB559fKLLV34yQ729uoVFj0lLxdF7lNbKN3iFrew4jf2jZjObX0lfbHcdcIXlW5/u9uXBUGUk3ZxIQMLrdNPPz0pq1GWNfttKv5usU8pK7Kn8uA1qIfJnjZbiS3mo31A7IeCd92XPEQf/X/WWWfpXdu4uBb1wmIUG9vhdan2sA3dbFGe7Uf7GO/L8glWxENFR/tI2zAf64G7mHqk+k6mvTCtYOUqlxohqf4jHgit2jbtmcaMMcWFAxEh+j+ICWuVtGkXHTPGFBcORISuDPvcMBvt/fo38P3zKJZjqRK424snI/7o9tLvZOEccRRxin6HH2NrTl8zajrXmZFUf5+XJB7xeo0eyf5k2nWFIJ5+rOK5xHiU18YgT4u9E084ceZPTfpYFfWiLO4M47iz3CknrZ1/QId93Jlf6uDTc4c++9D5YlEtkynplcx/aL8oc/LBRP/XsR4WJ+yfdeaZ5dUS6liQG0MLsp/YY+Rpm7hYlLNqee/B+Zr5DG4unxvGfI2nlUZHViLMr//X12cPlxs1MUZCh7UTLmYgPvWGktDwSuL97y/z5z/K/DnnIhbqN9X/zFfj9re2FZseI4PMi/8y4ahwHS7O30P2gsGFjdb/W2y+xewpT3uK3sgZ1V0IeoNHbB1wwAHyKtMt9YKn/hBP9lH/P5H95Kp1SinRi7FWWUb83+fe49d2vv71/5497M9aX9sFDfha1yoyB8L+FltsLvuOmK8vlCcfcTRNpzT8uMeFnsoPUhc8oTJeq2wvr4AcMvv8v8taRV4D1iMpPVduYN1exiK2P+cxrFWwLw5F8cRsbi9eoLu8878ViGdaY9rqr2Om3EzBZvPtjRH6YMutt5z96UP/dHb66adHZmJFrZdVx5Y6/nL99+G6SvLnEmFNiRucW8lFE9p1KP7fc489p9cqWQnKVNJnyatXT3rik3Qbgmwf4x7WKofLWuXicqOGVQG8cfWVUfRdq0s1Xkt5bK3yeed7WSfq/93vnhIFS8baoidWoCI01f9doFXSpucKuqtcYhK5CthfQMWkgewQTGY/6W9MC0GeJZx7IGdhG5oIyMSkQpwsYnRolnxkK8OY3b74xT2OptmE/f+zs3823OY2tx7O+cUvhl13XT2c9sPTpDqbyTAjjT2n/dfOFofvf+97wze+8Y1hlYjttfctBnm1w0NEw0T88p9f+MLwox+dPsjd6EE++TpstfVWorKJqJI8+6dnD7vtvusgg9pw05vebPjOt74zbLf9tq20BuSmEP8yMQ1yR0XrL19CGG5xi32G3ffYfZD3pM2x9BbdoXA2rF1cO5x66mnD3nvtPSyIcy+86HfDd7/97WHNmjXD+b/+jfjmpsN+0l473ehGpmEDx9+ZZ505fEt8f/ZZZw832PH6wx/d8Q7DDW+wo9qSxecgG7kNB93/fsP1r3e9ZdmX998HWZxK+c8Yttxqy+F2B9x2kIts3mfOO/e84aSTvjzc+973kfjYUnViNTEVfzKTDKeccsog++oMcrdmuPnNby76JO5S/8NQ+YX//MJwusTdXnvuOchFvkF+TBW9CGk63IrvyeJ/UOfZp5LlxB/KceZPfiJt+YNhzRlrhmtuc81hp5vceDjgtvsP8rRTMa7GtJ9ZOVCIDWMfen57wQXD//zP/w6ymBjkApe05w0G2ZhNYxF8PeiOZdRfXhtT//9Axgh5qmXYb7/9hr322GNYtfnm1DaG1C8cQxMB0kwqxGnD1b8oM/+q3j9c+6ttnFg/RxMh83v9pXFmg3xiVeLqNoP8+ND5CLG2GQZCHuKnpfof+p08qTLIpyaHM2TM3Eb6ndwNH2572wOGa217rfXu/z/92dnD7rvKvHTxJcPON9tZxsZvDdtuv53oKccGHn+hdUOMPyzeSuLvkksvG175168cnvPcZw87yDjvdbXCLel/2l3p/Gcmpsd/1b2O9r/44osHuQkwyJNsg2xkqe0un0dn1rntf9HvLhpOOPH44fQ1pw/Xuc4Owz577zPssefugzzVuV71P/MnZw7f/u63h5+e9dNhxxvcQNc9O97Q5s/fyHj8CZk/5aLJcD2fP6VocPCV2P8vvvii4fjjT5C6/2jAWmVvqfueu+8p8/U1cvioD0FA/1tcu1b72N577TUMq1YNv/vd74bvfOc7w5rT1wy//s2vh5vJGk72GRlueENZp5QA0iq2Go24ovrL5uGyTvzWcJasYW9wg+sP8mTjsOOO5mt5EmKQJxQH2R9juN71ba2yLvu/X3uZrlVOP/2MYesttxoOuN0Bw24yX64qFfjleb8c5GLTcO/73mfYequtij9Ea1p/GFH8s2hrZNmnTtZxslbZ6+bDAbe/nWhihaXKMnZ84T++IGuj04c99pA18h1uP8hFDJFYefzTvsMwVywnAoSQlNOPfyxj5mmn6lplm22uNdxkJxkzZa12rW23ZfNBUOWXGn9NKJ2TOUMToeiDgQt+c8Hwv9/8H12rXHDBhbLmvKH2v71vcYsV2Zen9PQ3irzeqOu//fa7tejbY9h8M8wlJSCTfXOOOkPQP7D/URx1S7dvTbJh/G8XTpp4y20OvHh+hBrhcpxplzCpqkkpldAkvjKUugiTlpqUUglN4itDqYswaalJKZXQJL4ylLoIk5aalFIJTeJz0X///OeH+9z3vhJCs+G9Rxw5PPGJh4QsdREGx/q6p5NAQp29DkTeQx7ed+SRMqatGuT1gOHud7+75aAuwqSnJqVUQpP4ylDqIkxaalJKJTSJrwylLsKkpSalVEKT+MpQ6iJMWmpSSiU0ia8MpS7CpKUmpVRCk/jKUOoiTFpqUkolNImvDKUuwqSlJqVUQpP4ylDqIkxaalJKJTSJrwylLsKkpSalVEKT+MpQ6iJMWmpSSiU0ia8MpS7CpKUmpVRCk/jKUOoiTFpqUkolNIkvG5Wnzob7cj567xGDbPhaLW+oqDaTUgml7Eqh3L0cjnzfkbKUlnnpOJmX7lbmJVFYm0mphK7UruejLkJndPu1S1IqocldK0OpizBpqUkpldAkvjKUugiTlpqUUglN4itDqYswaalJKZXQJL4ylLoIk5aalFIJTeIrQ6mLMGmpSSmV0CS+MpS6CJOWmpRSCU3iK0OpizBpqUkpldAkvjKUugiTlpqUUglN4itDqYswaalJKZXQJL4ylLoIk5aalFIJTeIrQ6mLMGmpSSmV0CS+MpS6CJOWmpRSBa2fOPGMSZC0CZKylJ6Yjjrii4CgtBMzjRBmyUKbIHX74gH1S3KOo45c6f5/8WGHDa959auH7a997UH2CBl22mknLxNbvIZRdqdPkJSn9MR01JCjjz56eOADH6iiL33ZywZ5Lz2yufIWcSXBmCCFosR01BGva1B6/GdfhJOJTXAnSN3/4gH1S3KOo470+BM3+Y2XEmLhnUKowAR3gqRZlJ6YjjrS/S+O2pj8L69ADq9+zauHa29/7UFez9L5qGp+b7FEjeZMREGVnpiOOuLagjIMnznm6OFBDxjPS6Y8SxZzEyTlKD0xHXVk0n5wi/4KTHAnSN2+eED9kpzjqCPd/+Kmjan/I26jdZBqjwnuBElzKT0xHXXEbQWl28++aL0/6Z15GZSemI460v0vDu79r46qiI5x9NWShd9kkAsneIFJHl+BZ+Vo+BUt84gTquDkKUtknML4sk+3f1X2/9q1i4NsojccLYvF/W613yBf4xiuLRdR7Lji2v/kb5483O3udx1+/evzh0MPPXR485vfJCZLoDP8JOJ7/PX+d1XufxHzffwNX3AA6P3/6jb+4VWABz/kwcPRRx8jr+7cajjuuC/JKxQ2H033kPkLzyl5RpbBLGG4bGwsTz3ercxLzxze9Ka39PVXX39quORoYRyRRgg6cULKjmGWyDgl+/h3dRv/2PIGe/v39u/r//Vd/6cnTsqgqiAPsJmOrpZ5dRdcTspzC1K9z0a9KuBSYc9JjizH3EjGcwvS7eefEcUzCtxLG8z/v5P3fO8v+1fgogneFz36M0cP173eDukyxoa1f/LJJw/3uve99H32J//Fk4d3v+s9ujj1mgnS2/8P1/7siN3/xec9/nr/kx+LcRm39AwF3ks22Pjb+595gJ793YW/G+73gPsPx39J5iPZI+szR39m2GEH2a/gCva/fB1quM+97jX84txzhyc9+cnDe9/zntI0pWRXsH3WH/Xs81/vf338aUbG3v+kU/goIc4puJMcoePWC3puQfr408efTXX8SRdO6vj3ACe5CfS8b5nJphxEFcoVTVkehoOgkAJUPoYjCSHkjtbtS6crTjVfJY8RVbjx+B+buh188MHDxz7+MdkUbzdZrB6tm3COW38iQqQuy23/zx772eERf/7I4SJZHP/t614rG9E9r8df5WQGSEWsEiOJ9fC/d29VsvHEX1RwVLtgFWwk0eu/7P7X21+CCGNzj39dhG+s87/PRx/72CCf8JUnUGI+uiL6v3z2d3jEIx4xXHTx74bXvfZ1Mi89dzTukHBF2Kfuidk1WAXr9vOPGnGKOGS5648+/om/+vjXx38dRPr6b2Od/0aDfiL08X/p8d8vnIzCO1+ZSA4lOpInQyHdTliYTBL6DIMxpgmvbl9arr7clF088ldmul/d0cZlktDl/rD+f7M8mvziw16kj0mfdNJXtGyj+qyw/S+77LJhu+22HVav3m046qj3yxd5bh+eYb0Jr6T6h1kviJB6/FeXV1fY/nMCfQ65+5/TQ4+/3v+urv3vzW9+8/CiF79ouLV8HeGkk04qY4WMDRtw/v095iX54tUuu60ejnr/UcPt5CsY5ZeV2uv9r/e/q2v/Qwfo8d/jv8d/+r3X178bdP4tk6wtd33Z78h6jz8L2OKkuR0vSriclgEN7LKAyHQtSDqNeA2hSUY9JhiZ1O1fNf2PT8hecuklw833lE/OpfECgXF52/+EE08Ybnfb2+knihmiWWdFm2BkUo+/q2b8eZzlxi6BkUm9/Xv79/nPBujcLziGEo54DaFJajalTTAy6Q/V/84o89Fe8gnUK8r+iSccrxfytyqfG92Y6s92BLyi6t9M82YnGyuFyKQ/VPv3+ocHuv9jSdrjr8//ff6/6s//Mfotb/7zJ04sYx4yQcnpjBczEyTmMZac8/ONJRs+TTt9N6dVmNMZ7/bVAxMu6f43p/i5x1+sAnr/Mw/08aePv5NPE7QDak5nvM8/ff4RD0yEBInGknOff/r8M7piJHHRx58yiGbQdqicznjJM0Hq/c+c4uc+/vTxp48/eZCRIUJ6x+Ucf/XCSb6qagMP7BRvaw/0bqgFYDsEtabUpZQUBHGaLCzqEVc1u311lvir+x9RY4vTiDSQ6mgzHuhF2NB8VpfKqcdf9orjvf/18Yd3Vfr4q4OF9I0+/uoAocOrnmwaD8/Y0FzONpiYnOHprC6VUx9/k1MC7eNvH3/7+JtXdegbffzVEUKHVBtXdRgNz/TxV31hvlFfVXORUfSsjpNTn3+SUwLt88/6zz/2xInGngWgo/Arx7Lw8VwM+XBoliph9Hwesd3ogoX+SCDnnsarLFViLD9ig1B2knYU2Xr9x86bQ6l8WiXGGUZsd3pvf3cF3Nbjbxw8cyhVTFWJcYYR253e489dAbf1+BsHzxxKFVNVYpxhxHan9/hzV8BtPf7GwTOHUsVUlRhnGLHd6T3+3BVwW4+/cfDMoVQxVSXGGUZsd3qPP3cF3Nbjbxw8cyhVTFWJcYYR253e489dAbf1+BsHT6EsyBYneJgrHeq6kk4hVsiZCyFLz2FSq7BncmEiNv+hltGWUK7RskIOh5Rwjgkjz2FaZs3b7Xf/9/hjTy/9RTpVjgrrLuQhBRxH7399/KnCQKMCJ4uWEjM5dFzChPr4m3saHdX7X/aKhQx9gxRwHH386eNPFQYaFThZtJSYyaHjEibUx5/c0+ioPv5kr1jI0DdIAcfRx58+/lRhoFGBk0VLiZkcOi5hQn38yT2Njtp0xx954mRRapGutPG3VWp4VjOR1gut8iOBw+0YV8+VoInhPIccAuvAqvxI4Oj2zQ+561eOKmwBc8ghsA6syt/9b97q8VeixqJDz1WgRFDNIYfAOrAqPxI4uv/ND73/azBojFSBUtwjYA45BNaBVfmRwNHjz/zQ40+DQWOkCpTiHgFzyCGwDqzKjwSOHn/mhx5/GgwaI1WgFPcImEMOgXVgVX4kcPT4Mz/0+NNg0BipAqW4R8AccgisA6vyI4Gjx5/5ocefBoPGSBUoxT0CpsixOewUl3m5mYrITO01RLEpGGqnri6lHCGYiAXt9qWjS0/v/u/xJ2HgY/64p4wo0a16/8vXvJdw1Ijlm0n1/tf7X+9/ffwZjxBzKX385XzV558+/8Qzv6MOEx1lxOrzrzinr//775++/txo1p8Tr+qUcUsvVgCf/pm21Dinl2g4/E1mL7mXCoRuv7h+0oGTV8Ho8u5/98Sc8O3xp47p/W/+QNzHnz7+6NDbx980mjpaRlBPVwiYPCbd18ffPv5KYPT5p88/EgbTQ4QEhzImuX39K+PrtGeE0cdfzj5znNTnH3VMH39XPP76qzoaaSWeIuqsD0YHjYDTXjshb3mNQTah660IKZFQytakkiKRkMIOjUE2YcMuycRNKGVrUkmRSEhhh8Ygm7Bhl2TiJpSyNamkSCSksENjkE3YsEsycRNK2ZpUUiQSUtihMcgmbNglmbgJpWxNKikSCSns0BhkEzbskkzchFK2JpUUiYQUdmgMsgkbdkkmbkIpW5NKikRCCjs0BtmEDbskEzehlK1JJUUiIYUdGoNswoZdkombUMrWpJIikZDCDo1BNmHDLsnETShla1JJkUhIYYfGIJuwYZdk4iaUsjWppEgkpLBDY5BN2LBLMnETStmaVFIkElLYoTHIJmzYJZm4CaVsTSopEgkp7NAYZBM27JJM3IRStiaVFImEFHZoDLIJG3ZJJm5CKVuTSopEQgo7NAbZhA27JBM3oZStSSVFIiGFHRqDbMKGXZKJm1DK1qSSIpGQwg6NQTZhwy7JxE0oZWtSSZFISGGHxiCbsGGXZOImlLI1qaRIJKSwQ2OQTdiwSzJxE0rZmlRSJBJS2KExyCZs2CWZuAmlbE0qKRIJKezQGGQTNuySTNyEUrYmlRSJhBR2aAyyCRt2SSZuQilbk0qKREIKOzQG2YQNuyQTN6GUrUklRSIhhR0ag2zChl2SiZtQytakkiKRkMIOjUE2YcMuycRNKGVrUkmRSEhhh8Ygm7Bhl2TiJpSyNamkSCSksENjkE3YsEsycRNK2ZpUUiQSUtihMcgmbNglmbgJpWxNKikSCSns0BhkEzbskkzchFK2JpUUiYQUdmgMsgkbdkkmbkIpW5NKikRCCjs0BtmEDbskEzehlK1JJUUiIYUdGoNswoZdkombUMrWpJIikZDCDo1BNmHDLsnETShla1JJCYgnTihByJwT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMT0ETmC445LSXSjhEhnLBLkonMFxxzWkqkHSNCSGMTUEXkozqA6WBOkvG8ieDc3IRso6bHxcggTCoT6lxFPDUhAR6Obr/7X2Khx593BXa6uvcwRWi9pz07VxFPJTHSAHH0/tf7n8RC73/eFXr/s5GBI0WdqqnGi7NzFfFUCPhD+ODh6ONPH38kFvr4412hjz82MtSjB1OEJtOenauIp5IYaYA4+vjTxx+JhT7+eFfo44+4Qj6qI3tWcbCwoWLqPC1BKmGdE9dkFrCpkSyGqo2xRJxxaOuk6fxZ27QEqYQ5h9jo9rv/e/z1/tfHnz7+ymyPmUgPmS76/GProD7/SkT09Q+cYH1jznnOCqvkm8Pt66++/urrL+kj/fdP//3X1x8+w1wF1l/6VZ1q2vOEI3Om0iBTUuZJ2/y5QJeggBNqpGJ7wpFaeCJFyW6/+183H5eA0PmascIAYbqBFdsTjjTS4yQle/z1+OvxhwvWvf/18SeNkxwgEymjFdsTjmTRSZySjDtCF6aAE2qkYnvCkVp4IkVJ2iV0UQo4oUYqticcqYUnUpSkXUIXpYATaqRie8KRWngiRUnaJXRRCjihRiq2JxyphSdSlKRdQhelgBNqpGJ7wpFaeCJFSdoldFEKOKFGKrYnHKmFJ1KUpF1CF6WAE2qkYnvCkVp4IkVJ2iV0UQo4oUYqticcqYUnUpSkXUIXpYATaqRie8KRWngiRUnaJXRRCjihRiq2JxyphSdSlKRdQhelgBNqpGJ7wpFaeCJFSdoldFEKOKFGKrYnHKmFJ1KUpF1CF6WAE2qkYnvCkVp4IkVJ2iV0UQo4oUYqticcqYUnUpSkXUIXpYATaqRie8KRWngiRUnaJXRRCjihRiq2JxyphSdSlKRdQhelgBNqpGJ7wpFauKTic8SFYOJ1poo2KtWk3kQMXYrJye/2JSmila1CrGjdfvOrhJ6bB8174Comp+5/ceEcd1WxVmQqWo+/Hn/Vr+I5geRkix4kFZNT73+9//XxxztIhVRjbeFUtD7+9vG3j79Vn1k6Yb0HMorJqc8/ff7p8890r6nmmiJS0fr80+cfmX+aCycWIhovjjpSkyXFGFKJWqyEHMGSTApBo/yVLu2oIyqXU92+xbD6JDsmedTQJZlJOsk56kj3v3gge6PHX48/rOE1JnJgpB5l6JLMJJ3kHHVE5XKqx1+Pvx5/vf/pmJAHhjSi9PEHHljSOclbSc5RR1Qup/r428ffPv728VfHhDwwpBHF0CWZSTrJOepIH3/EA9kbV+b421w4aQomBZ13ZZKSuSIpAkaoyYW0YpF0+UzKuAs4YtylZVy4ODykFYukC2ZSxl3AEeMuLePC3b64YiGFvfptwnmZlPHwJDHjLi1D2XG0ar6JzJmU8dBEzLhLy1C222+9pekJ52VSxsOTxIy7tAxlu/9bb2l6wnmZlPHwJDHjLi1D2e7/1luannBeJmU8PEnMuEvLULb7v/WWpiecl0kZD08SM+7SMpTt/m+9pekJ52VSxsOTxIy7tAxlu/9bb2l6wnmZlPHwJDHjLi1D2e7/1luannBeJmU8PEnMuEvLULb7v/WWpiecl0kZD08SM+7SMpTt/m+9pekJ52VSxsOTxIy7tAxlN7z//cLJ/AJwY6Mk4agjUULF5tHBJI/QstYpo5FjGwslCUcdyRkEn0eHGHmElrVOGY2cbh8bGyUPOepIdpjg8+gQI4/QstYpo5HT/d/93+Mv9RBHHckdRvB5dIiRR2hZ65TRyOn9r/e/3v9SD3HUkdxhBJ9Hhxh5hJa1ThmNnN7/ev/r/S/1EEcdyR1G8Hl0iJFHaFnrlNHI6f2v97/e/1IPcdSR3GEEn0eHGHmElrVOGY2cjbH/yYWTRSlzeq7Ea+BIrkWF1xK8wFJEKqYkyouVfLwmFFWC4Vd3cEi2WJ2z27cA6/5XD1TB0eOv9z8Z4/BajYQCHq+NowqUPv64OxwJVzVYLdHH3z7+YoFdjio4JNHn/z7+9PG3zz8yFPT5l4MkYDVQpmRDz1kKXkv0+bfPv33+/UOtP/yJk9RjtVtGNwxs1Hfrnuvslsy0wXz2LIJQymhhNbAsrXidxdktmWmD+exZBKGU0cJqYFla8TqLs1sy0wbz2bMIQimjhdXAsrTidRZnt2SmDeazZxGEUkYLq4FlacXrLM5uyUwbzGfPIgiljBZWA8vSitdZnN2SmTaYz55FEEoZLawGlqUVr7M4uyUzbTCfPYsglDJaWA0sSyteZ3F2S2baYD57FkEoZbSwGliWVrzO4uyWzLTBfPYsglDKaGE1sCyteJ3F2S2ZaYP57FkEoZTRwmpgWVrxOouzWzLTBvPZswhCKaOF1cCytOJ1Fme3ZKYN5rNnEYRSRgurgWVpxesszm7JTBvMZ88iCKWMFlYDy9KK11mc3ZKZNpjPnkUQShktrAaWpRWvszi7JTNtMJ89iyCUMlpYDSxLK15ncXZLZtpgPnsWQShltLAaWJZWvM7i7JbMtMF89iyCUMpoYTWwLK14ncXZLZlpg/nsWQShlNHCamBZWvE6i7NbMtMG89mzCEIpo4XVwLK04nUWZ7dkpg3ms2cRhFJGC6uBZWnF6yzObslMG8xnzyIIpYwWVgPL0orXWZzdkpk2mM+eRRBKGS2sBpalFa+zOLslM20wnz2LIJQyWlgNLEsrXmdxdktm2mA+exZBKGW0sBpYlla8zuLslsy0wXz2LIJQymhhNbAsrXidxdktmWmD+exZBKGU0cJqYFla8TqLs1sy0wbz2bMIQimjhdXAsrTidRZnt2SmDeazZxGEUkYLq4FlacXrLM5uyUwbzGfPIgiljBZWA8vSitdZnN2SmTaYz55FEEoZLawGlqUVr7M4uyUzbTCfPYsglDJaWA0sSyteZ3F2S2baYD57FkEoZbSwGliWVrzO4uyWzLTBfPYsglDKaGE1sCyteJ3F2S2ZaYP57FkEoZTRwmpgWVovnPAubCWS9GRTiieeGtQ0bieIar/kAzNCWGoL71KSbt+uwnf/p2vGKZYMTefE6/EnzlB/9P7Xx58+/vb5p0yqCjA2yLhQzcmZb3iff/v8a5tc9vnX79mmNYah6Zx4ff0hzlB/9PVHX3/IfFLNNYiNPv/UPunzb+uBTXH9kZ44aapTTQ7gZUL5iZ9Jnj39/He+I0VNSnu+BhmJZEK3r0uc7BJ3X/e/L//cP470+FNXJH943DTISCQTev/r/U9+YuSQ8PDp408ff8rPT48PR/r4q65I/vB+0yAjkUzo428ff/v42+cfGTOqCxUYQ/r82+ffPv9qt/Ap05ENtv5IF05MeTIRl0oyUef3RFDU0jjjiL4ccsQITTKfjZP5jjtC+URQ1NI44+j2zQ8lSjRBjxFSImD4kP5zWUconQiKWhpnHMzf7YefiBGap/LZOJnvuCOUTwRFLY0zju5/80OPv4gTYoT0UEDjZL7jjlA6ERS1NM44evyZH3r8RZwQI6SHAhon8x13hNKJoKilccbR48/80OMv4oQYIT0U0DiZ77gjlE4ERS2NM44ef+aHHn8RJ8QI6aGAxsl8xx2hdCIoammccfT4Mz/0+Is4IUZIDwU0TuY77gilE0FRS+OM4+oQf/aqDiub/KEe4DM0knCWIyoRp5Y+mSaxwBokI0Vtty9RaGFIz0VDhOsVcwH6TmBEcPEthWrHkzrS3f3f/d/jr+5e3llKPyNo6ZNpEgusQR//6B73qRB6/PX4Ew94aDjCICmwpU+mSSywBskIdYpAjz91Bj0XDdH9X3nAHcTYEdjXX+Ei9Q+dVGANev+je+i1vv7u42+ff7Q3eNdwhJ2kwJY+mSaxwBqs1/gTT5ykTpqL1JjKLMcpM7bsIjbfhmBhJEK374NEeC08mjyV2YoHL7BWSDkjdiJ0/3f/l0E6xw4jhDDziAcvMPIIlTNiJ0KPvx5/Pf7YXRyyhxA6IyHBCyyxFVXOiJ0Ivf/1/tf7X9tt/DpR6ilzZWK1NhLp619xyfjVmuTVPofEm8AAAEAASURBVP708aePP6OBgz2EcCQghOAF1sopZ8ROhN7/1qv/xYWT7OnkTyW3aZcdM8YUFw5EhGYyipY3sYJOrFXSpimXQoakuaIUAOz2u/97/PX+V90STANEO4i0aRcdM8YUFw5EhPr418f/Pv/lW/LRPUbT+txONWaMKUkv0d7/+vjT5/8+//f5nyNiDdtBtE279JgxprhwICLU1z99/bMpr3/qCyca0HLhhSE+5yoU2XmBYx2m6TZMKsQJ1yzEXYVOdtZXbcLc7U9eBcv+YmOZLxuPMqkQp+7/Hn+9//XxRwYCGeQ5POjAYIODLGj6+N/nvxIRff7t8+/EXeA8XvT1h3nDxtJmRGVSIU59/dXXX3391ddfMhD09dcmvf60Cycc4HVot1NNSqmEJvGVodRFmLTUpJRKaBJfGUpdhElLTUqphCbxlaHURZi01KSUSmgSXxlKXYRJS01KqYQm8ZWh1EWYtNSklEpoEl8ZSl2ESUtNSqmEJvGVodRFmLTUpJRKaBJfGUpdhElLTUqphCbxlaHURZi01KSUSmgSXxlKXYRJS01KqYQm8ZWh1EWYtNSklEpoEl8ZSl2ESUtNSqmEJvGVodRFmLTUpJRKaBJfGUpdhElLTUqphCbxlaHURZi01KSUSmgSXxlKXYRJS01KqYQm8ZWh1EWYtNSklEpoEl8ZSl2ESUtNSqmEJvGVodRFmLTUpJRKaBJfGUpdhElLTUqphCbxlaHURZi01KSUSmgSXxlKXYRJS01KqYQm8ZWh1EWYtNSklEpoEl8ZSl2ESUtNSqmEJvGVodRFmLTUpJRKaBJfGUpdhElLTUqphCbxlaHURZi01KSUSmgSXxlKXYRJS01KqYQm8ZWh1EWYtNSklEpoEl8ZSl2ESUtNSqmEJvGVodRFmLTUpJRKaBJfGUpdhElLTUqphCbxlaHURZi01KSUSmgSXxlKXYRJS01KqYLWT5x4xiRI2gRJWUpPTEcd8StLQZm420k7obSiuJKaWhQlzY464lmD0u1nX7QunfTOvAxKT0xHHen+FwfjLnZ4pMa7/1sPZE8V3gRJOUpPTEcdcb8Hpfs/+6L1/qR35mVQemI66kj3vzi49/86qiI6xtFXSxb+vAxKT0xHHenx1+Ov9z+JgegRNT7ugVmy9z/1wIRLgp6Yjjrifg9K93/2RY+/1gMT3pkg9fgTD6hfknMcdeQK6X9y4WRxNpPns/lEZpiLxiSNEBzihCHdYlki45ST7453+93/5fn46QiZ/8NjSp6RZTBLZJxSPf56/+vjXx//x6NFjBB9/Jl34WdqRKXfxh6dku7jbx9/+/jbx9/xaMFxhKMGIejECSk7hlki45Ts408ff/r408cfGw+mR4jx+i89cVKyKMjZMx3KM4+Dz/Kh5xakep+eelXApcKekxxZvtEk6bkF6fYtIMw9xTMK3Evd/1VcwlPZN+a59Tl7bkF6/PX4K9cLI640QDxKGjoiLfOQXr/DcwvS46/HX48/9p/SMxR4LxFmpkM285h3+dBzC9L7X+9/vf+x7+R+5r0k+puTHGHG9YKeW5De/3r/6/2P3af0DAXeS3r/43zvLjEkXTihAw26HMlCyANN3jfOZFMOogrliq48KBkBCoUUoPIxHEkIoduPga77X3xRgspiJUUMUYU9/nr/6+NPH3/zHMMBItNqfCQhhD7/9PmH/ajPv33+7esPGzNtrEwjJlGFff3V1199/cV5I/eYesVRp9iFnCqEvv7YeNYffuFkNLzllYG3XiAj+WAJxmYnLEwmCV0OOZrhpduPlUnlW0uM/FXJ0MGE3f/qAbqDsMefuMWG9FE89f7X+x9/GVRjiyVG8VLJsIMRFiaThL3/iWN6/0N0jOKpjz99/OnjTxk4x2DUXyoRDrCEhckkYR9/xTF9/EV0jOKpj799/O3jbxk4x2ABW5w0j4Ok4VQ6FNjFgT7ejvVUeZTdCDfJEJlgZFK33/3f44+TO6f5cQfMfSY6V8iN+MJS2gQjk3r/6/2v97/e/zCS5HEhRhbDRryG0CQ1k9ImGJnUx58+/vTxp48/GDDyuGCjTpxHvIbQJDWj0iYYmdTHnz7+9PGnjz8YMPK44E+c2BCUWa1oy2s0mQInmrSc8/NFlJl7NbO1kdMZL4omSKyeseTc7Y9/bXf/z7ma3AZUTme8x596YMIlvf+ZU/zcx58+/ti6g7OfdBGJjsm7OW2HyumM9/Gnjz/igYmQINFYcu7jTx9/+vgTYy+wPv72+afPv3Wf0FQ7oeR0xkvWCdLVZf7RCyf5qqpVHI4po606xzyEMw6Ow0GtKSaVzppRTpPBinEsrmr+QeyLPbEoBUSZ5Lw4G1ZJ2Yy2WCoIhrybB9lVIjdbJXytiFUDJBFZ5brwHp/klTPoRbXpFBkjCAP/kRdIEdpo7OeCS+nG9QdRyl+KLi4ZVi0ulrrk+ouA1LGIKW4EoeC/1t90IR31B61kVPTy+38R7Qr/w2iJPyuXnVEg8LX91TTto2ClrNr+lC9qhLdJtz8Kj3CVQ/0vbTmDfxYWh8VVxpAqgil0YKvEU+aTzQSHbyAP1+CwHoV80k+0gSENKmSLIbhQ8tTtLzLCXqWqYV/SKAf6m+hGltmqtQIRZ7+XFCyIzsXNhLkZiisHpHBIQusgeZAXasBDQjDyzL5q1qwRf6BBB3Qh3+WPvw1lHzvfz6yypbbwLbw7kxKjwPA3nAGv22H+F1zqb1XCeXNxMThKMUGMf9qkoK+N+kNG8i6q6CrVjgwiauOe0DcTvulK9iEPdyuQ81XA/8sb/8V3UnGLf6k/qg4vaIwKS2NT+pH4xnxmTtL4Fzlrt+I49b1kgWwZf4xvaiCMLhzzj9BVg5yFDrQAxY0gFPxHXhHRQ9IR/6CUjK195Uj/MxUqNluQnli1P7Si/0MY8aBiphJGUf+Jo53/EVFqCDqgf1Off6V/FSdItdBbOWfaWBP+Lx5TP20ucYP6S87S/sil+YsboZX+h7fl8WHzmztekmgF73/WNt7/tU2EvamvfxBw8BXHP9Rf/4wCP8H/GpfKQu+DXxZl3hG6rGFs/YeRTcjqf/Dh0xXEH2ygqaBL7aAEIAgF/wWt+39Zf2gOiwkVVPsie7n7P2yv1P4K6m8V30D1v7rZxxgrbYVDmw31t36Nmd4iBw7GaC/rL0AsxIFpNo6/GkFQoXpMWY4/MOQQUvR/EGCUDEEleXnjb3G4TLoS1hXodVg9ouwoGf4uE7C29L/NhIK6AMr6DvGPsXIB6z6ha9EkN/oRyiWA4x8KXfV/lYUYZGEHBMHxH3mBKE3Om/r4p3W5nPXHAlAP+AoH/CMHkhpYaDM7CsdY5WwcFS5SCWhGOcH/E0c7/5vIxm1fL5y0zrEOKMWfrudE1c2/YGiW1sNNjhEbhNQ4V7R9vJ2EA6HiHcsnJwlA5UtNWBBFUTPr7oYVvnZurI1NJ3jao5E2A+YUY4NbJMmWHLC3kdjHsm7BF3ootNVzYSYDWam/DXzGQmXgJhvSBUd1hKXny1V/6IdWQNiQU7FvmBKFhAFjaf9rmaxQqXzwPwYDmWhQXy0rNMGWArVrBKNxuF9f+1pA6hSFRK/U+EOVxb34QY4CLWwusS3QFo2YrKS29D98gn9FVkMVaTgKNPm/Fhc1JK2tIWQ9SkyrQyGnNecADM8XHPJaGMiIV1RWoCTR4xbKjw1YUVFtM/lhgYbVBbPJebzoQkJUWciqHs0oJBxePMVhRCgbSf9TD06MP+ZnG2XgJR23tNdZ9WL8kbrQ1xBkbbXSIMiiRPlYdFnVNd43s3wQ0x8LcL4wNJtc1dKmLtqgRVWjvUUX8LAPIVAkpwW4CasiIRtHobFFdhPw//LGH7QKOpVUDw5TJ2lVjVZqba7Qy38iWuqvDpaxqGRVUeYXCDYO5Reo+oUAsTJSmskiuzz/F2Ue/2HLlAkf9tWy9T/Y03IrwvYXohzW/8GQI7V/oSiZxXOaVkpPamXQ+Ue4rLSikN5051/t0lpxqYPWy34iWPvDV4JJv9FRUYXpPOGV+iMf/pknlKjjHyRz/2OsZv9LNlUEoN1T4FWm/0mlFjVm1FUag+qrzcSfwmN94QGPIaFjyANTLzjJ/APP0rvF6ZqD/ve8qqX4H/klnf1v46soZxNag5kuzWvocvyvo7SaMiUosmFKtPILLdu3Ctf22ZVgGTpwLMe+Cm4A+25UFHb76tU5/hfvaP8XGTaaQOvT1spof2098GWtg3Qe/zm/b4j4U1OXq/2lzLp+U01SNSzKUGJdectZfnHonQNEhcyeM9zYkciUfqX9FnXU+lsuaEGP036q9R/3v6n496BTy9CyscS/FIT1R51RO21/dbqWU0hCM5+Zl+C9whc6ONb+hq+k/nIPRA7Yhy3gckDxMo8qS5UYKxixQVCjm459+Y1s3Syqp7UoyVTFQs5cCFl6DpNKhY2JrDSr50LuTDXxbAE4DmnBOSaMPIdpmTVvtm9Xbi1A9UKBmoAN0VM6o2WFXhzJvvxYs6ucxplnHzFXYl1VLsrCaJV2DKizXOggG5N9XPkbcGVYC476TdUfdxnhKmtNzVIaB7iNeWhVDJjqBamupGTwtB/KoKGbIyVPNxSqDWMc/m1Kn7QvufDzgiUDLN3NxKFeaUhKOUQA5Z30v8ipfVUiGa0yAosSt2I8VAd1ukq0P9pPFpuL2o6CY66CH8RZ2o7qAhBwoOIFCthMmokcwAVZdKJ9QcXdAPjdWrf0blEIl6pKPUPCWhz36G3qFwT5wZEgggY7JK19UvQLhIReD8EErHKIQ2jDgbsawEETHdpgRppsfynXxtT/LP5Qf/MU/GGH1NCqZfEndOMY9PqrjHFAM8zqj/FvlfoL3sMhPkL9xZZ6Gu2PJlRfs7GoI9u3douyQZMaNlIpOoBWQ1hQuan736u4ZP9HZdkPwkNaf/U/PCWOBkGieDz+J/8LmjS4b+FTHX+VTzL0CaMAsNbf/5JJ88tpFH/gCV3KXY9/bHlkRAAVwwqEp4Ww6uKuJXqlHRREfAWVPI11TUAOR9Sttk+TRR/VWo7K/pUZf+hYYR/trx2ttH8Z/XSskgqIj7lItnky1w0+lflXsoOKA7BcbqMzlK6n4n8Iqf+lQTQfAgh+R/+n/5N9cxw0qLTKKopmWDL+TRQ51XSxj3CP+oO5ge1LPeAzG79EPezKnz5tWsyN7Iv8Kpn/uI6xAotj4ZvSPub/okzj32La+p/VwSJYjOAQks456CeiEGcURP0saqTlla6CyqEF2OCB2IAk+gXy44AO2EbaZlbQXD+xIoOcZsPKjuqg/Tf29SfHmE0u/rSdxNfW3Ggwa4HSDOsT//pkhrSjHVQYuuv+Z/HhcaIxWooh4wQiSWNB6FBZiuMQUcb1Fy3ZiAAO9NjooLw58QeexbXon4w/9E2sD+UZEnlKWNfbpXo2/ggF/YXjj2iziytWApy95Mi3MY4/bHg6kSUuDl+6/VE7ycj6Q7hq/1L/ortuf4YcmNIKbOT1sm+jBfdxoUZYNRxQKkL7lgJRDyPPYSah/PvblYlSjJMlHCqNoR3YxmdfYheX6q3roc2aWmj555CVt5xTlR8JHO4t4+q5EjQxnOeQQ2AdWJVfEhhfpDdbGSxhNkAqbQSmTk1CgIguTsBD3nKgChrjVEVGgpCxAcyI2i9Q+Y3QvjpBJlhWUesvxQYFD8/hQKDjBysqjilcqaX+qCf8oY/2Ffkl6180wtfQoy5BftiQP7NvPFscghcHpJjOeEhgsaClUo2gw8YifnrP9T8tmUYMVGhz1J9jG/WrRDE8bR+12Ejbv3i5dAb1ow5hUmD4lH5lfOt6EnURd6JOOLTOOIkv8TqN5VEquSKjl9AUQqtOrsiJ1Zw6FJMzfA4o/UzodilXflhKmq9abaYXRVg24ciFE9yZsPKhbXA3xnRYyTCda402mf4X8S9F17KLX0v8aV9Q3HqHSYgb4CRxvPUVUL3l1H/qQ6EC4liQxYtqAEHzImV9DE0CGjUoG2khwJvKExw28bQR+WY/8kGUB2SKWiWhjmZfEFUgUBA9gyQ04KBFnTeV8RcjJWqLimglFHj9QZY/q78IaP0hZ2NOtL9l3tjqjzav5z8pYamDV8WqLWfWwdofr+LpYY1LtpBAiPZHNr2AL8jGVn8rm1UDZ62zngoeLMfAXouYACKHjlnagZCQP3Go9z/EB4JAfQI6bj4gphgh0ncXY1wrrisuLAZEFgfy4ADV4w9psac/UjTYhKANGv7fJPsf3MsKS5X0kIprP5MEfhzSO3rRQ2RRTz30SRXWX3JgaSNzGTxu/pe+KfL6M7SMv6attBsMS3tqrEouUPUCB2Q1tcT8p6VC66D/S07o0fZG4fBEGiol+dW+nFAnkMoBqfn9z4R4mUXHbyFpcaEIiCqAQtZfMNjQrFed/lfHv7XRxlp/rIvs1Vj4H2W1+ENz6evM1jjadtpq2qBY96BfgylQgkJfddGWRfxZUwv4g7c/4m5R1uA+amHNpnHOQqGvyayJkMTTw9IBtQ8VNuJb+5/HqzDKof4xJxVfkRMQ7Cuz/uuyL7+/tbBoOa2iZLAmXt/+B0v4q4912bcBRaSk42v4jFVoeSbItaElUlo38q1yqajG1XMlyAzmj43NvsyjZSieU2gtPkTKiIpAXp9KhFpOI+GQCgvBinxF2GeVaQj1US+gd6Mc6TDeVP2VU0si8liPoseSkLUnMDQ4Uy6gG4t9vR5Syq3DjRRMB2Tp3Pm9QnuUDlOyHAiIcnCQ02dgL1f9Qycwba/lxJ8KSwZCaQzs2YB77LYIssHB3gkthRZZsyF2dHWBFA/DcW0R9dfmLfLJSBEGQyptDW5QOJaEno2s/bV9MEWVA41X4p9UVkVXbvC/jPFOYz5AqbdeLHMHiZTGRe7zgosKLA7toCYsDSTK5D1XbSg1AhlMHCark0tqf/BmsldHHGgf6BMdOlaJRm1MyJmmjc7/UXit5brGH15AYTOhVtYapkh9pf6XRiosSpgfS/zpqoycVAg4CiszW7UlhqA+/lsbqv5sRKUlr9qXRFJ/VfH/eP5LlfT6i4+xENJYVGIdf+IY5LJDBxtBQUP8K1qatOgWoJj7X9iFVjlZFZY8ptzPlf/VkrMcMRuSBNIcxitaBNT2p4SFlorCavEH4YQJU0JBSeX52e2n/m86khHVIOmNNP7QqHhcHeOqPt/lY6BVXYdd1l/a2scyqc+CjGkWUyK1Fq8nYrREDqmvB40lQULAqXros6QhYJQ5LMcnmJoN8lqQIl6A8SwvdYNWchWpkr6y/I+yS6URIiijVqigSNqBNhDPqaxQCAXVYiNdHsZb0AsnSZn2P+SRfqptYh4QCojyV9LC0x+tagRzD/6hzU3OJJmX+UzC1zjUV2JEm0TvbUo+yQKyacAZOniAAQFJNywtAeKP7c8sBao48m2q7a/1mK6/uQM13DTqj76uTa8ntJysXEv760U/VLO0G0AeK3SBVgJgtoj1EWLQdFxZ9dcyiu9j/S2dDI3CWiBmseaTQy/olnorQU7aizCn6lUjk7H646yKiigcI2mQEsuSIGyc7a8jhI4bpdylNgRaleXM/7Lerca/4prl1l+dBmNTx6T9KcExjc2BhomxcCzn7TbB0nkOlRNlMf5NCY5pV5T9iVd1inFOFgzwpkxRoIaBJJg8Jhuj5BYw1xFXoH29KwjDbh+FlIQA7cdINfaLhHAwENl9RdCCLgkeqJ4ckITCMjXCgslrBxfeRmYf5dOrwQJxlXhBRmrUFXXQAUzoWiuUWyj6TzoscIjhMICzSays/mKNnRW65cC5mFAcQy1omS5JO4qg5tRNqaL89nSDiGn7S9tgsUCNkkHzgK32WTdbqkISEled9k/thJqj8uo7+t/qC5qyTFxx87+2PLPYwI0sqqPk1Q7likEUZSJQ9T8Q5U7LZtgEFplFXgwgF89E9Q6McNbqLKE9TFgoDf6kPGg3TCKyCsS/lcWfqNjI+j/qj8dd4TdEpfnfvAOP8VCfoexyqHckj/VEQLhDztoclkYzqJzSccEJTLY//St5VCXyamaxbzx7NQvW0qGyRW9rX8T0zljV/qYT5UIVcWxs/i8llJKh3vP8D59KyxQfq4em6q/OtAWk/phBnJqDS/3hf9HjYxOsusQS9kUIx7r8rwKiUeS8/WnfNNT+FzlrF5TCyjIv/lJBFUVGjT/LSu2mxe0XMoH4b9Ns/1Q9raFUWuovS3ZJWX/T3iu0BTw5kuuv/sETkuZsvZCMd3K0p1sPHvTHEH8QsU3UkPpYMDtgjofoZVuBpOMrf4G5fRiH3dC5yflf3aZO1KpYwFqFqvFP/SJtAf+LYzT+6S9d74gAnp4ETYNeO6H6DXpw4Iw2Rf+GP41OJUaBlErLSVVJHpMUPvSCKMbtQgxsSGFkbNXxGVqRFGpeeUhSKabJ+EZrziwKNIgtLSkgkiJ61R1/ix9K/bUFWH/1/sZff8SFhoaOAxIH0miIDm23Uj3g6NNGxwUBe2rDqJJbYntB9gpRCQhXuZGEEWEUYJEqiWxE7TMvIA8TEmkVJ7WCYOohsvoYK4IZJUZe60+68IZ9jEU4UCaIs1y0jyxSWwCs/ZAfOLUZXQj5oEpTKPlQQ/QDK8HGFP9Y0+mSSktn5UNV1nv8Fb9ZOyIzNECXeGeZ8W9lsHx6pv+hY+KgLydYbl95k9lLbgFo7mkRYSpjkutteWXZl3FaL2ea/Qlv1KSSIpFwVHpjkE3oYhUhJRJK2ZpUUiQSUtihMcgmJBuTkvZBDQ7hovVwVIISdhJ0yTsqoAsfBlUlrxr0pHfXRCfY6Ors5BYJxR4YG5v9tVLmFKe4cIIFPF6DMOdgAJPHhaUK5fdTqaFUhhl1ABR5VFO4UEfoKcguWX/kkj8xQlcLQY6iiQoJjRlnlEFsYHGDR2zXStpaw5490bK6fTylAFtWZi2warKByH4gCl9UQsdVq/1L2+pTBmhDqThcTyeg2uIWJQkTg3lV/9w50si7/vEvNqSN1kob4EIW7GECxQWCuv3B0Zd2RE7aV8uLMiE+0VLAbWrlJmRoZnCtDjhbbCwdfyImOTae/m+T/v9P3nt4WXJcd5pRtj3QQDe8B2FoAMITIEUjDklR0mg0c/bM/qm7Z/ec2R1Jo6FEoyFFkATQ3nvfXV2+ar/vFxn58lVVAyAcG9zorpeRETfimrhxw0d6/1CVv+R1fPiMg7fOOy7/dta5wZFHV/8yOEv6KrUqKCVW3yMtCmAym91NryTVE+iZdOAPxKDcQ4apyTMrrxX6z0D+jbMN+q9IFFLvLKdap4Q0+o7655mAVmCB4rVX1Jrp3aN/kqo9lCt+G9/tSXTvohKt/GtXN1oz1JOxdIOXzptOo5jQo6GJUV5j9meQtMeP5+7TPyWAs77AQWSZ4ocB61/CjYc/3/1TN6jv0acm8FysOFPhzUYws8QpmeZv/Nf2b6BWAQBSfOCvk6hk0uMnoy4/PV8q+SuuRnxkqkwiFX4bWzLHm+0d7Ue6vu09McBr1rodJ5+Gf2Xc8LZnpYK3Tv4pwLHy7wmVCF4g1UFwfP7ID62iBd+xMojsoVr5izdqJmz+ukQfBz8JPg3///+0/7UINsq/N+ut0D5M/mRhuVlgmkztX9UEwyz/kf2NmcikCRgs2qSsqdfZnVa1pfaOEh0g4j8MfwUkr8+q/J3YsU2UrlonPGZdP+NHWGxch7SChA/5zPFrbSIhuho9M7J7PXyNb7+fSv4N2WfG/2b7O5T/8AMMHYMdGynpUZcgoR9S/ylT/+k+Cf/tREEv7OTUZN69NAKb3NuzRffPGtGi23NDdPc6iB14G+x4UPfWAtuzAffPGtGi23NDdPc6iB14G+x4UPfGY7TjpEG0Z0u5xbOC3Blwc8zGkNF772ue9twCbwuqIHcG3BwzHpLVvE5DOlNUNUQLNbJygajxRptHdTU3frFstYHs4tRbvEkTQ1Hfa+Z9jewC6/tdgd+GOCxAkza4t9R2vaCwmzhpM7XZ8moCLfuqdybo8KdzXCWVNLKYSH98aW70fkf+AckkDUn6bLrkNbUAyr9D0VHhI3k66WNKRv6hKEx1NAx6nMkrA1INPG/h33TidejeCUMDHv4qvjQBwa/PXHAdocH/ZSj/yEGyod+O2gb9D082srhaZxS5gxnhFZDSaTICTrHR0jf+e2Mc+dR8ktngvZWRsavTUOJEHXQoPj91O3JCiAs45G5ZBQ+/+YQuBWdsxcLN7N0/gnD+3Bl/5R+QDfybouLwOaKl5sbvF1b+7MbJV4ugCJkEf9ip+q/Jkr/s/iDSUmkdlTTSTY5J4wRTu4+myjCXs1H+KWo76haADXEVAKvgwjPhYqEkLRETDOCAy50NTTYdfGQWWMBNkp8g9wUXDjqfJRWi72L5S3EYGVDP+6byByr2xqgR/JDfZBBdVB5y7sCo8l/lr2wbVC0fXz8e/gACa9l1+PPwp8s0WY/eq37zbtCd9N9o/loOgobm8F+zboO56KBBATYhL2jvhCPS3tUcfO19zcOz8Vr5Vj9GYaM0AG6Sf83wM+f/0+BP5bTuIDFto/UKDie6zz9Ka4TrDLUiEj5lR4xR3au64RHT1DfrHYD16IW5dRLryi/8W4XNzAz8SYFUWbZdDVWulZ7A3an8k0MyMrOAjsq/4k9EzZB48vws9S84PwI/kyFhETBlVfH7ggzSf4HObhaukiaziNu7suTI7PnMjnru519rXsq1yoxH7wRNmsQN+O8gvHB76tPyr57w5wJAFatYq+vxQ+TH639WXWk8hLfupfIn0+T9ZS5/RfOn1L9Pi99638m/9ptqeY/0r5a6aBITve7K3/6r+safwVVnOoWtwKQaKHByqO9fSPmHPun2n0yi11UJeZcHwwl1h51ygI8JP1GMt5IdJnhxrOEfaUzxsfU/uZtb50bvfV4GfQ76X21+mIJc6OZvbPwZOaTQgt/6XHmGHmhq9IXirgirGSNN+Idoy157YV54jVeQk7YxRqkUXaaBMbB7b/kr0QSNosxkS1dB7gy4OWZjyOi99zVPe26JuQZWkDsDbo7ZGDJ6733N054fhZ8Oj6AD11K2YEsLfyvQFl1DO/GbvEW0p2GbXR8bT/82AGxhPnWfB/5qYKJwQ4UJNitxh7PhVzFdpUgwP27pNCjwhvMSMQagC7VqNx6SUcINy3G9JDbe9Dw60JpqEJB4Ij9P/JInCj6prgGe5DN+CcJ4ra0yaKLjMWXHl4FtmMbrpEKGqmzYWOXGUEUwOUUjn2/8NZY6ppKbP5U77VMVl/EVl7G6cf6F7GSxQf8MbebAXNahwU5pFSTUaUT4L23GT7WOKi9CrfppeIjWYEzwmRgvJU2O8DwxyTZHCYwxmk4edlUnUnCk/nMpfyQzZjCRQJXf6BkxuONgFdgUBwNsO5bLufKQMkdW3e4DG8XoC5a7yhWpkSaNRSssxScW3q0LPt2uKJ419E4R2yBMEaEuZOu6Je3WUzITbpUMeHRVQp+FqTZapL5P4x8O1AiK6/QPHLUtEZYXcerFjetfDYlyf5717yPxw58ydQs5dIzmuOARup1WWaN8piZZjYHkVWAmEaT6Grl2vFXZVvkvLy2j3ivATaUMA+LALDkwtQK+KQSaEAyD1WPKzymtoQsR9DawWsq8WPkj98ZISp+cOsSKMa5KVzq+XPKHj49R/pVb5KNc0OFqeZSJfwlNmDKt7akpalwA8qbUDGuy6/wfYv9q2pZXfQ5zaHn7NNbiq8XV0hhT4ywhB35D/HYRJuHfi6Ett+lZqCbpijph8fO3vMRdUtNMo5oU3msO5q+DoxiPhiNBNZMN/AsR2kKpaclJZB9D/oFNurtI/xSBDCGT1ZWVsrS0UKYR2swsQkQbDLfE7WA7maLclKvPSeDW5Jt4FxGWV5AxMFMzhFke8pr/TGuS9/TMdFlZWSrT0zOUgzvz+Ef7rZuybU4ZOPQAZ1BLHHlVFIGr1AwCEg/c3S5/2Gxap0yrg0nbLvUHt4aM6MjE5tHMG4L8gE3foE4cZ+LTvg5pVlaWqQu1PzQ9MQuowrZuV5GZe4/TgugFaWlVSYpl5GrYneqfcEP5V7IrLSP5A2Sds0gCbypcKo2EVTqUgD6dfNQB00RZoeOzfRu8JHFHM8DRPxgb4q+puzzv9vJv/Hfl0DipPLTfTv7hl7BOVn2xGUT6TfY3wlYOAERfgOjl38nn0+K3PHBtUtPyWNPg4uyO1nZWAvij/2l/bJmH4Xa/kk5bAh32g6NjST3+Y4l/uvbXDMQ30L/IxZzjybP5eImzzghhtdNfi6nKOmEEra1mRJE2ZW1tSaUsM9MyT0/Do3XUKq3XOH4yk5hPK3+Jw33W5W8ZLi0txQ7Pzm6jjYR++LL+ZZxF4XktgnbaspuF36Z/hqtvxFLk9tNsD5wOBxZZrZD3tpltySeTT8SRgLyIW14pO3buCE+tDbCvF/WxLFQt5WZA/vMT14WlgAwwnDD1vvPmWUObd/DW0hO0hetj4+nfBpAtzKfu7sTPl6IorSbASumWv42d8cgW2p4bYs26qyUqgyKIA7yVQ8rnC8avUcpqjQThr0dSOlOTcuoKqxUaYJs5TGJiBrDwasVrejZsjPw6SNgUHrncVfghzNWspYV5Ol8zGC6MlTxTu9ZWa+dhigo9QeerFwT8WIU10rm8iucEnRKL2w6d8tIQWP4V0qduwL+vgRF+aOqFHJf4+JsJdTVUPHEdLoPVLwipRpoA569XaWmkV9tl42F/yQHh6jodC+KnuGNDniYn4N88HUjS2VztJg0yGVMx8dshyLPSIbIvVfnLhcbWhtjRuIKL6OWn/i0tLsQ/vX078sQg0wnTEG/fvhNYG/f1srSC9JhAmWbgJP8RHfmsKzdlgn9V/RAfP469iYrT4M/MUj7gz0AAAClxwqQ+KZMVgNW/5F4b34wnUqHMaKXmbTpzlRcHCSn/muqurn/Iv15YDC8KKcorX3nhudLppY0rr3KpIHHq8yoCzSB2elsaxMBYFwMxzr/JFJvtZgC68rd+6DV79pLEYx2hqpSlRTvepWx3wLbuJBWPSQd+Qitxn2bclZD5CEqGff03TjDhIPCusn8fKf+eQxnoHHxEij4r/yk2wzpx5J4o9LBCqtfWjsp/lVVN6mR1wruct3p0WW6IaqHtWYtUIEO2tr81UlqlS8CG3zRxvacFmBd/FL1pWjNg4r59Nw1lnIkTvLX86fBGJ6jZDEaDzyzNa4C/KkYfG1wBSopKTLI3be+EV4t9drF3o/5B3+rSIvVlinbH2oTDjtYt2soHrUD/ljGIrhA6ab+KPXUyVHjLUBt7GzNsP3h+kckTwrfN2IkGChHQZKeNunnzdrn33l0gWC+35ufohM+UbUzQeCRvjYKh6x30kwxCrH8pK+voZ6T/lv+frP3LTrmwV3+sUwhngtVu9Sv6uIwdQ3Zp0wkwfo3JqOkZwrBnatHcsoM1JIWotK1Vt3jg9c1BXi1F3zboHxDhnxjz2rr+ka7ZP2D01/rX7KiB5k148I9ZWCMHrtWKpv81pcjDb2d/V5aX4I/+DeVjPTZ3B128ZgA2xYLYJJPuvlfUFX+jQ4QNk/7qhNnA/5+y/CX8y4w/EwOUOVJNE6vHAsmCXi00g1ohCekAOf0qwYB1XFwVj6f6u2X7W2HUD7O3CGv9T+68dHnwGLrN5W9sC+2e5tkFD/XPRR17aHbJghPwzMMB7ABe/Op6+tmEzWCYHHesorfbZu331fzdtejn2CsWMd39+qc4tLOt/7m4uMQ7ZUV/1omQyhrtwaD/PTe3kH62Ey3yb7vgAvDNmzfL2bPnyvz8PP3vHeX+++8t+/Y9gNyYNMeuXb50qVy5fCnlvm3H9vLA/gfK3r17I78VKr79j0yeN/sTESJ75F+tjNRu7boS3hDZQttzPLrZP5lUX0UXB3j0z5ck3Tp9Ba6/W0O00PYcpiDr2Faxfvb4GbeYfafw4u1feo+hH+oapBlZMdqzT9QA+oBxz1h0/9J7xoG3eGuQDW979qANoAuIQPEbrBuL5iU8DEKr6AFszCVVlZkK4UoE/+FbxaSQ8FtZjMsv4YHpcJm8z7ML8z3uT4EfgmOcMsMJFTydNJnIDKUdXmNZrZA/6vrc3FK5cnWu3Lw1VxbZeWDgFLB2+tboqCyxsqMcIi/lmIEs+Zl15bLL0zfzNvcuNjKsYaldo5gxoZly2s4AnUM7mY760tFxdhoTkRX3CBVeMNx19nYqdHq8xK2Ay8saqHmMz57y8AP3lx07aHng27RZeeLOFzsUYYG86o6TxkEjB9gva/nDSi3bqqmVUQNtkCqfra44pbSCbOrOHeSXVnCd8qeRi7wmy63ba+XKldvl2rXrZZFJOHOdonGYZol6bW2FTqoNvV3Pmr07lOpgIQXVrcjOl3vv2VGeenx/eej+e1jphhY6CFl2CZ7aeUVNoFPN0bFzQmzqDk8boSnoa5Nxht/V9U9Wwgc/ip+XylkNpeahv9QpOxgC5Ndm0MkO/gSDx1XPDvf2x1SlLCytlCuX5suVa1fKPA3ytm072Tg2m8GvO33Md5myqTbThpzV8CkmRaz/oFpZmmfgd7vs33dveeqpR8p2o0JCq7EOwpS94YTh+dLZv0iUHx18bJZ/grsIma8uYrcsOv6zU4O6gymq9p+MhFH/squKgnInjwO2RSYka06tRAP5sfGb1i5kNGIg/1oQlVR+A6FuGN69EdbHxJ+1k/gMtwzNVwymEA+aRqfVDrADTe2jg3a4YffJbNm2g0FXyh5YVUj+fSZ11aME5H3zzwgTcUkk9lFoT3cIq7maSw0fyf+u1D+EoTldY3LTyWCbwtBOVVXu6s48E5OzrBy6u+QGkx/Xb8yVG7StVy7fLBcvXc5CxsOPPFqWlufKvv37KYv1cvHC+XLjxrXokxsk9jCr8thjj5P1arl85TJ2+EJ59tmvlGefearM0i6vUM93bNtGPOULEbVkQ8pA0rwj3o3l/6WQP231kPb0P8KZ7Y1MUQiRd62DDjrTb5nyyCECxC2wyO2aEZunytUr8+Xy1Zvl1q0bLCgt0cNggp++hrt6sluPLKkGhPNDeYhC/cuxR/Kq+l8l17c/whiEqzHVT3AtDz36iYRUXFWWHnYUEXy1nIgFTMg6+KqTadTQNJsTa4vlHurn0089WB7afz9h8I1redq+Z6ewYTXDL2f5K9vYIOoUfsumWbxe/vDYibjnX1kYpjzieKliGIU2WQ0KJqA1/DOyPx1h6XrRRiyxc8w+VnbxOcCGSidf7ZcusQA4zcSpVLumdP3GYrl04Tr2YJ537AwLfxOJ73joHrY0/lNbqr9yoJx8jxQQnnpQ4Qjq4iLL6PkojT0sw9Xp7HjQuDlRqX0Rljjd5Lo74LSBHjiuCy85RmL/3QpHnJMkS/O3y31M/D715P6yeyc8oKqmMxd3Mmfh1gwtIB/8Vfz8gu5uK3/7Adp9xx8z27IcFRnUxQS7tbSn8OBfYGbZJUj7eujwkfLLX/0qsM88/UzZe++9ZQ7ZnDp9uvzh938oly9fLrt27y5ff/HF8vyLX626gHze+/175b33f0ffe4kJk3vLE088SRvwTHn5pW+WvffdiywV5lB2IBY3f/2icF9svScwH/bTIK13Zt+efZoG0AeMe8ai+5feMw68xVuDbHjbswdtAH3AuGcsun/pPePA3VsmToYxFXw80VjYJqqGqbfyj/KKj59UtK1ACRvD1cGMhX0G+G0EVR//zLt3lHodKBLqwDCIMQ4a5DSRNT5x1NSkbxnwkspLZn2+PYCePrue/x4usfz8yfDLhMbHFe3qd12lDo4dnMk9cqCGX2QQdurMpXLo2Mly/uLFsjDvagYDYLeYUXmXMAbT6aDJj9lGiHi25l+jUaWjZ1z+raiVv+Yl5j6BdbCGZanZJ+/OeCYbO//+uYV5KQMVt/xNMYCfonO6yMqTu2tWluYw2HPle997q3zrzVdYrdsBLXZwq4as0SpN0bGqBtkuEvSpFN2kQc8eOGXiS1f+ytXi0Um8rw6YXVdDfomzGgDnnh1F71ZJV4+XgZumVXOnwxIyvXB5qRw4cB7DfaicPnWqzN2+nW2JTrRkiyAoVtAPB162htOs6s2wS0X0bmVcZAv7yuoCndL58vzTj5Qf//AvymsvvVB2ziJ1gGxAlbv6mS/q2G8gtf/UzlAmgbzVcrIRD0MJ4ye4UkYEJ7suzLi4DfrXEvzp6z+dhnU7UvLHHx0jfUgltaI+KTUvd7a8YO7W3Eo5c/5KOXz4WDl58ny5zZL1FDtSJlYcJCg8yplOjhfyKiZlawc6nR6k41bPRVY21pZvU85r5ZWXvlLe+darNMi7qRNOKFYZRuQkV6CRbfVW+sy3KxNBUjr8SF9QdmHGxd218lejuvovP3eo/zKo5cjYJOpnOh3yZKC2yCTW8iJ/rKYtLiwkzsFyBiwB0/5V7TVN5EX5GFbtn/qPrlsPzLUps2/Iju2j+CzbxBLGM/KvYdaUdHB7+cNVX4BmyJ8wcOFgsuIgCBqnnUSn7rpiZr45ngnczl0700mb3caktIrU5e2EQPKTXnNt+UlSB5OIDT/iTEOjjEMANTwy+Wj5h12Sy0byiSeZmGvLrsffwyWWn89D/2gPU6KUf6Ok+rCk2ECsYWS5vDJZrjOxeeXKjXL12o1y9sIVOsKHywcHDiLjPeWN118rX33xmajepUvnyweHDpQTx45GVs8/91x5/fVXypNPPFL+7Vf/Vi6eP8Mk9s3yyivfLK/xd8+e3Qw8OKLBrsFJZFkpqlYgcvo8+f+i6r860piJh5eMQpVwjVhGBlPaToyXgxrrCksBDN5qPXD1/tbCWjl85Er5/e8OlIOHjjEJdSErvNaj7UxuTbG9Z5n+A6BVz0QjBt8HvDqP0wPEU6mIjvKjTiRdk71PAswiNd+6ZEB0soaZIv2zQT0zJvZDUNKmPpMJXv7gcG2hfOP5R8tPf/hO+frXnoNjAOE3A6UObtz+SJtE/JnUv0GZKJMqf6VW/wzr3edR/z8u/hCH5LPoqH2HQovBo++WRVqWybQj7rpwIevqtfly4vSFcvjQyXLi5Nlyi0nXCfqrM7Tzt9HRMGkm1VN9vGainzBlMCAvYL6n/POrbml/lZC1ZaR/yTUZGIfDn7YFTx+sx/EAu7kmodl2jw5g+ijZwQ6ku0ycXFhih5yLmF955rHyox+8ha170iuHcjtA8DM2MeeQks5HJNTR1vBHZOP4BwyGZn5a3e3plH7dZ1z+1j8nfJbpS82ywOCuL2W44iS6dRhClmhTrbMz7CxZWFwsv/zl/yr/9E//WM5fuFAef/zJ8s1vvlweeOBB8lgs59ltcoD24MrVa2Xnzu3lK899pbzw/AtZzJyfv1UOvf8+Nutgmbt1i77GUrnNZMsObP5//a//e/nud7/L0Z2dd9T/ZgMio6opVSb8joUpT2n/2K6mFjw+fpr8t8qiQo/S9OkaTXcJ/g0TJwOCe2/vCZ/Dt8ZDwoYRmyTyoZED6AFc7+09nxl+6Y4GmaP+zlkta3UkoEPbwlLJrFhCEKfu2EGO/lulzbT2T1t2fa4aPSDrf8CCw0x03UNvw6U/4UlSafp88UsEVK1yFs8ehdXbWWvPtOAWmeG+cnW5XGOr2B8+OMmWsGvl4uWrrIqxfxhjMIVRcKvYKp3kZba60iMc8ZVKtoF/a062h5H5R/DfC0K4Jn+9/EmqA2rlGfkzoDM/4yJyGxvucXBHAt0jjCyGmsGLnf+VlQW4XMI4L5Xvf/9b5TvvvFIeuH9X0tqdMnPPXDpxUhsuykH8lNufTfkjj/QtlVcKwnLCGRh9VpKVb6W6wqicneKUdYZWNISszF1bLO8fPl8OH71CI365nDtzhi2Ft1CLidpY2MjTYE52KwspNORoXyCTIBmMLaZhmJlmpWXxZnnh2UfL3//t98q33/ha2Q4c0Pz4V4s19ZdwG5w07NDozhR1IVZZwJYm/g36l7hO8boHCQitdU1/xBGUX0T9C8YPwc+g0Z06TGCkY4xX3V9xe7oTe92Kqbt95xbWyyVWqk+fvliOM2Fyisb2xo1blNlsBmArc04mknad4wOUs8XiVm07Y9YhjwysUMhrFO4qHZnpiWV2mUyWN1//evned99k5WI7dYZBn/JHrGP1j7DIrbLT/VoWX3b5RwORUach2hjY2mj/M2lFXPQzKkg9gXVXkZawOwsc2XD1eoUJ2f42fSYloqv5VdcUIk691NbgqfaNIEATbBR/6oB0dKDVpgqkM1AnwEb5CzOwv6E3WLoMTSdM8PuibjiZtsQ9G+x8IkF25BHmRNr9rGRz2gFdrHIyhWlHzh2IdcVtFLaVL5zELjRGe/4B/yj5R3RDtEGxBf9Kr9E3gG+4kqySMpL/J8WPjJyIznFWyoIha8ot8qftXFhap8O8Vn777vsMKKbK9p17y7Xrt8shBu3vHzhazp27yI6eXeXxRx8t+x/YjU2dZOfYTcIZKN26yY6SJ8s773yr7L/vnnL81NFyg91+N29eKdtoj1984bnyAn/u4BO/u/9moCE7E1rZft78K+uN+mfYZy1/bKJtStV7mEr+2h391f7Edloa7Mxzl6R1I4tCRN+aXy3X2EX7q98cZuLkYjl6+AQru5dY3WeiMM6dkwxoQeICUbgyazyKMOisU5Q3/+suwKQLpBD1vyAfl3/4CazwIvGdpzmKN/Y3fBAgi8FvvLjq5Mg0fZyvPbe3/G9/893y1pvftCb3+Fv7m12fgW/4yEt8ZoMn9scgYPImDR1+n9U+EdjLH+DeAfBFlL+C+TLjV6ZIcoUBr5Nz2gnlvUa7r4668yQ9Wcp5njb+LIsiHxw4WY6wgHnm7JVye24RnaOXy64oFzKdBKxOoVT9T3mC52PrX7IggQl9bKF/tfyBSLyZN7wVu+/rdDgMZUiR/oLUuEvcnVsrtIse/2azBQ3lPBMBT5a//sm3y8tfZeKEoEnsloN68ZiJk87x6iejip/4u7D8U78YQ3jvlMduPJqzyiSK7bblmXIiXtu8iCyOsNPkv//jP5ejR46Uhx96uHzjpW+UJ596ouzZfQ+yWioXWKj2zx2ru3fvYkLlofLggw+wKDZXTh47Ut793a/LoYOH6G+vlvvuu4981+j73Sw/+cmPyo9+9ONyz733dDuxEd4G+1snTixEpYvrvb1nPFgQosbKv0sawLGf8TzGosZeBnC9t/cEcvj2p8SPvol+5MYII/iOsugkO4Qf5bLZV+FG0PGNXvsEw6ChvwfoPTX2w2F64DFqZdl0XTtXmTQAl/xkGo82QP1SqSIlXkb6Fsiaj+BdGvPY6GzIaoW30mseK567B7/MQpTHKXhmpwlW0gMQ87eX2VnCbOaRC+XkmXPlzLkrWdlOx4NRl509Lzly0JZJFzowDK396XXnI/kPND+Socwjy1pGekc5SSRvgdNvLLl3aUJ7gkmbAnM1XTPNgNxOP4OWVSd2AJxC+Nsgc2Z6mQ7F18u3GKQ/tH8PudVBqh0Mt/N64WaMijhJU+kxx4rf37TXRviyhftI/gdpk2v3/rnrXwZtENzjlyv4Z4CuwqeMIcjVYvf7eNMG4yas/gRbB9fYeXSx/P790+V3750uZy/OlYUFGnkGiR5pmqHxt0PmasIKg/BJd5jwp25p0JdYEVmaX8zsu/K1UzrLHRoL81fLs2zV/C9/+055+80Xygz4cycpZTbZXU6sXHL0AUXwn/sv/DfZL/NRx6IUtTDuWvnDW+xGL/+qQlX/a5y6np0FyJEalnLJBCAJ0WLkgPaRwFt6Ll2+xUTJ1XL42JmUzS3PylKUU5PcT8NAaoaVn4llypGB+xr3+ngh8gS79yeQ6zJltMiKhw24ZbiNCZkp4rez42d2cql88+XnmTh5rdy72y20yBwdIGUE/KXV/48t/49n/zU5zUJodVawNcpz2claOj5OILqLTZgcUwv+KsMoQvfel38gCYyFMcdOUSx7FCNv+KMfJNLmOcFRG6kOtmLr2p+Wg7pDfeny6XKKXcxZbJVOE8A//9uJX6G+rvKnXcwFd+IH565du8r+/btr3ZTU5swjztzJCRvis1KQbEdkdpA+aip+/d+x8qVuf+HDkvCIYi2RKeoksqf+3L69Wt5971A5ziTnwUPHmRh5rDzwyBMcq7tZDhw8Ui6wQOECxiT1dmVlEXktlF3b2TlpO4SNffCB/TmK8/AD99H2ssJ4m5XG+ZvU5flyP1u1n3ry0fLYow9SZ3cEJ4aXAYq1tcrZFV1l3BHWh/f692WSP0ykqW9cRP+sM/zZ5qBQtgNIPscbsn4No/PYvQsXb5cDh06UgwdPld/+/kS5fnOpLDgQBdqdqrZjroyvc9zJOu7qeepG8hu1/8aJVpn2cu3w1trbaKjy39T/69I2+WOdU8f4qQmSeSsugIPfCRKjfXdFG1OfdtZjstj8iYXywhM7y3/66VvlnTdfTj5OIAm9jp03sbqUf+BJruSn58+i/sGKvN3d/W+IxCYodO+rsPy8BJ79GdEnB9g0G+U2uxWvXJkrvz9wopyh73X81IVylck+lwVnt+3ABm8nHX00+q1ZqCJ95f5j8t/KvJV/UvvzYfonhgrY9L8GGIgjTy9ltr82Q//Oo2yL2LLb7GZdyd1p3v00wdEc7hjCbj37zEPlr370JrukHiMttRXGZ+kbxnX6b/5b1T9xbeX+VOXvBIZ/ucYAu5t2kPKRTPsDczfnyjx9A+uj91OdP3+RnSYXkdM0EycPlQceerDs5C4TE9yam8tEuZMs92Hbd7DTc3ZmNgsyJ08cLf/6838uhw8eYCJqsTzAZMpTTz1FO7BcPvjgA/pt3y0/+elPOfLD0ffYrmQ5Nv5VRjppq75xfyLHfirkEH4sesPLRui8b5F4GDT0b8iO1xr74TCjVBuh875F4mHQ0D/Kqfr6iZM7A1nQinUA0Xt7z4Z87xQuWItrz5p0/G2Y3eeBHyqs5WlmKuaGRfpUHEOrM0alr01fIjFq1ZbQtNnYpJVsKUxtY82zhSfDliv5J83dh38VI4WNTuXyEjoHYidOXiyHj18qp89eYiVsniMaDHJnqLScqZ7AoNkgLWkgGNjZAHu3hFZeHtOJ7yU54D/a9FH8K2/dQP5pyQ0ayd9VtKygUZ7+s9rnPg3fYqSq8VphBOmkiVRs58jODIPFab5SMjUxz3bmF8pbbzzPPSdMnNC5nGACyYGNl5I60Lf8c3bZxNrwP5fy9yxqp7ZVcvDm4LybbHLywW2Fa3zRxokPv9Yyx4z5uTPXytETl8sHB4+WY6eYBb+8QEPgcS0G23Y0kZP5uUvfiwmtCw6+HCwoQhuThQUG6TScrlzv2LGzbNvORAtxC/PXyjOP7mXHyZvlL956IZMikrRGQzPjETAyoFhTDuscNTE/kXlMwbP7AbDs6ew2f4BaXSTubq1/VTrQV1niV1e1epIv28RmRZas3mdSyk7IOlv8F9mav1Te/cNRJk4ul0tXb9DJauXGZAmr/U4v+WUsVD71MjuxENHa5Arlym4I7vtZcgWIgcQMcG7xdGJxBxMnS0u3yktfe6L84C/eKPffRxkwiHBF20k2t5z+udi/O8v/49n/6Dx65pE0Numhy0yYcCSn3qqfWKsCMnN7hmXrmppOGfqrzg7L3zRaWCcdKDj1O51HKFXuhMU+1axrXPKwkIE13LYoYfp12kBtmuHir/GuAGYnCSGuaKV5JEyd8/OWa0w8O9Dyjgc7bB7XMo+dtANeSOomQ538BXFY8YX8naDrJtMrUYZLXIB5bniLLU9wYr7s7e8KE2by6gB8hbpze4GjWgwmzpy9Wf7P/+v/KYeOnik799xfHnr4KXZvbuc47HWO2V1mGzfHQ3awW4TJk5XNKNSDAABAAElEQVTFubLG3759u8tezrfb7rkosJst2/vu38vfnvgPHnqPXZVzTMLcV57/yrPlK888Skd7dxUwcnUMEvuMv+mbtGllPq3+1xJrZc6z2dwUd35Snp+L/UWdxBC1Uueif+oYOhz/FJP16CHtvgPRBS4yuUx/xl15Hxw8wwLAB+XkiYu5p2tigiNNKHQ9YuokM9nSDiKyIMmGe9kj78pL400CrMvqvGFdeB4NBjlbd5NZCKtpNsjfGmOJOEFa9V/kgpJPsje+1n+ZdfJ0SprRCydM6h1v0yw8zJUXn9xe/tNP3ipvvvb1LGSIX2efyCOztY/zJS9/pFGdvHXy/yL171Pjt8R1ttveT0M/1v4ZlXWJBaEr1+bKcfpch4+dzSSf/fNb2BFKkJ1+O9ilwo4GJ1hQ1lV2cNSjIFUDa5uhriEX6Kxthv6R/Y8up9JIh7qla7+jkM36B2wqmPlVDpJOb614MQPilLdVjo4tcmeaF5zarmxnwmcXNm5mmhaIfsZXuIvnxz94rXzthcey28Rj895laIuly4IY5aoK36n+3S3lnzvBsBtZaOh0UbpvcZTm6NGj5cjRk0yIMBm+575y/7797Ay8B75oKanH27nc1UnbZcryhhMs7CrxuOwOdprcs3sP/bSVcnPuRrl+/Ub54P33yoH33qU+L7CIsa88/fST5RF2KJ4/d6H8+tf/Vr7/gx+UH/7lf2CRgw86ILWt7I9FZ5Ftdq1lUP4dRO/tPRuS3SlcsBbXnjXp+Nswu7sTP30lm4WByHoOes+QizH/OERjsAMZi+RFxeG/7YXKM3JjgCO59gIeQW70jaf8+PhrPhKCr7WG8fuOx85p10EVtpErvqHbhB/G+smCQaRmKIMLjGDfpsrf546/Urw1fjjp8Ntk+n9piW1lGOnb8ws5Y33ELausXl+6cguDzMBrdhemj4bWnSY0uO4yUepp3O0c22iLkplxV1NaOW/CrxCVb5xPCelc5G/D50BhEN6iefahIKiXlda8shWXNA4EVxkA0gTlolrP409yIebMzHZWz2dYdaOxYZ5+Yn0eG3uLiZPny9tMnDzEVmgYpYzYOsjewXUuPq2dEp7QG3J65PI+oCVy+OLKP3LfgF+Bj+tfJXaT/BNcqc9XdZSoYdYF/9hx4lEC9WIZeaRJp/N+6dpCOXoSY/ybQ3T2T2VL+TL6sua2BeS6jl6kibPh73aeOFhwp4k7Grxc1nJxF4p/GnAblToQs2HlHOzC7fL4I7vK3//0VSZOns8XXiZZQVliomUbDayOPkXKQhod7MlfvZcG3WyNtQyZJe6O/DeehY0f4OifslEWBnZljLcLTZg/vleI+majXuUvXTVS6rbCb1jF2QEGP/6Gf8z+EI4MhAxTNKCr8H2b7eVObh49cZb6epWyOctFk0BNchZ/issmKRMHXyuuLFJ+s048sfPcM/0TU23CxAueF1JnLKspVrdzCSIjrCkmvXZyNGCJXUBf++oT5Yfff7M8cJ9lwAAMOqWnE3HCRvwTY2QnnK34D0DjWcD4zUY/icf4J9yo+uh/BygSG1mnzIgZRH5s/CbaEr/EjfCbtW6AIm/u1sDocKSRCWbqjX9+JjY7TexXIl/1M5NNAvbSq/lHCMFPVPjnCRY1yKe+IdYa0sIaRNcZbnyQIhASC/58acEQ7Wvwh7Dkm0kSXu24x75gN/kPKZQ3HV+f6kgdbNFBJ3IXEyd79+5iQoU80auhPgRxqBaXRx6hEToU08iNSxHw4K78V8AGLuTQjaeE4lb/RDKIzHE+5LG+qf0VrgMUiV7g6g4eGFeGOItCWeWhF9eBVha7EPG7iyexA+AlVliTnnbn6rXbtKc3mKhcY3LkWvnv/+//LFe43PHRx56hYm4n7ma5en2OC5vZ7cgKsgMO27Xt2ybLfRy5efzRh8pOPid7/PgxVhPf596NqfLaa6+URx/ZX86dPV3OnT7OquNtLoudLd9665Xy1lsvsyBwH/jtkFO2FIVPaUw5bMF/GP4k9a/JX+nIfxUfD8pGuWySPzB2PwWsQiZd5+/wt2zMSn+fKS8tLsG+id/GIQNEs6opzHuNQdoCbY8Tiq7ynjl3tfyeY8e/ffe9TJ7cYEV3ddVPPrMCPstAlEHLqvWXSVDlpN63CYaOkE34h/pn/+dj8T+Uv4x0/LtCr988a72vEqjWoGZuHVV/pcsdmy4saLvtE5mt8ZOrt8rXn9pZ/o5V/Nde/Ro8OcFS87KeC2utrXmN8I+Wo3s2xDbmKhUtiBw+SfkP+TfDDeVv7lKr+7zwR75VJAP8g/pf0ff4h/Q0f6QdG9llNBDOuP5bXh03Aa2A0VW8imPJnU0BmaY9Xyln0dX3D59hR9RJdvbeQCfJkV2k6xy/Va/9aIO22N2n7g6eQncm3BocG2/rgT23X8R/ihwYEcdTA4zo9C2Ijc67RITI/qms5MdwdcbYCgOO8J+Aykjw1D66E8ZLqxxVpa9Bj8SmKJ9k9yJ6++OFCZX1lVvleS6g/+H3Xylff+HRTLV7LFjtrOSEsIoONJWKDh9vf7T+mfST6h+kDPEri034O4hW/h5lXaI8//D735d//cW/MHlyggXDneW5F14q33xVG/4oEySL3A84x8SJ98JM8yGOW/nYgth2M2Hu5Ni1a+wqPnyEI55+SGEvx7RuMqniZ8anaY/dAbovR3UucIH4oUOHyve/9xflySefxj7Uz5BvpX9VJ4ZCHXLXZDz+HIeo2lD1YZiPaYBMveaBt+JveY3n0olsYwYNeOw5nvKLxU9fKJqzidARGSPfGNW+jFPeR28Mbu/1Ofztk2zKbIR15BtCx98y3hCxMbi9N8w+bcirAWiJG5TvxjhcpHZrdCjpfkCqZgiKa7am1uRqRJK3cYHwGbPV5WTfofJT8WO8ekdKlStOUwF+DFHDX+OA6Q1a5w09Nn4Vp7mbK32UpPWRy4iAkA+3+Urp2lrtYNcKU2m8dmu5nLu0RIU+U46d8IwvW3+Z2fbTo9MzO2ict2UltQ2CkxMN9iSrkPVLNGw7o6MxRWdIyitNG/gHu/j9+yj5y5U7Fswt7OAbCD223wYncgLAs8fLDBKXMdDMnVhyWeGzQdnBzPYsq3nbMDwOBpxcmVybp6M1V956/UW2sX6Fozq7yYNUbJf0s8yMy3HgtHNh8fgqY/EMSZE6OapRRld6fW7gf6z8u4yArylHqT5M/zLAER9V1zJtrnZ/qlSrhJXbh+AHfc+XefFXZUnqTJwwkUHn8RZ3wpy/eJOO5pnym9++j25cpLGmg0nHfhXj7qBca2h6Ze6/aRpwV0ukgHaeFdNFVvnQD857MmrP1zgcAEyhO8p8mbJTv9bYrvnYg9vL3/745fI9dpzMWL7ku8o9ADM0DFYRXvM0XKcE5NIIQyyJ6LX1NjFVH2u9EEZajanpzcNUNSe9xng4qdbDVv/rBGEPlSStyppCOszFDqlqUiHdlYXaAJhn16OW9B5/+Ki0ZNAW+oSu1Ht22dUZi1p8y0xwXr3B5OaZeiznCJNY51ml3r2bLft0N/wqhGUyyc6qfN5U2shePjz6ZB6enfarUgt+Ncc7aGjUs/PHyRWod2JrnVWN7TMMHpZulJe/8Uz5wfdeLfvv9TPErAJJjyTLdH5G/PfBifkQ/Yu0TK8bT/Vh+t+SVZ5MGiJ6XR7PqdqBaknhjERj5V+TkknVXOmwDozjd8QpYNUvPLDc2QSCLT7vQ8qxNJ4escjRFGEATUkClCWKTEoQSvnU8hdCN6S6YurbHyD9F5juYYqe/y6uRVUOa77+Ru/InyFgNDr8J0dxttIjFQR6jM6BujrsJI+p1JudTDrX3UXqNrIi3vPau3btoKN2LwPNQT7mCeLGETlmt1N9H/4GfffToEmHd1z+cJ8yJqIxKQ7hwmYNzG/S1vqON1w6CDVO/j1+5pute62RNa5RGxtITIW2/JQHsSwckNpir+UNkbaxmacFWnmktGFcGZqD7+qS+01cNTzDly8uXrxWbjLh6STJTb5Od5xJT1Ykyk2OxF667MWwXpLIoMJBMDt71rmLyEH8ww/dx3G5F8qePTvZ8Xeu/ObXvy2njp/KZPJ+Pkm5zOXa169cpPOMrnLJ9hSLAj/k7q4f/fDt8uBD+9ImY0aip96Tov2zdEeuSqu+V/vX6x88fWz5k8F4TkrwDvbXgrY0QgZQCs0/ftQ7JageOCCb6svfAlDw0FRNpAmA85PpZKT8ky1prX+8m5PuJjvA/GrRAe7k+s27h8r7Bw/Trl2NvKeQt+3ZHJOe7pyMDYgiUq68u9qrbqxwJGoC21fphs4QX62LpFdtQt9FiDNMbaicqA8j+yO0pCrb+KBV1+yPUlN2Hoeu8g97ocOEQY3ddqeJixPb3Ak26xdMzI82l636E6tz5atP7S7/5Sdvclnwi5Fsrcdgtw+kQhNqKTVllocsSoBf6tXxUfkre4Jx4k+xAG9g/a3RlZOW8x3Kv2WU3Mx0lOoT1/9QMszpQ/QP/GKt7FT8ffvrwlrjP5RV7gLFjxIb59+wWqOSH7picVbJ4uc9drjrV7W2QP0UTnvrnWXrLD5SKuwYXWI3wWIuff0dO0mPHTsbG+Fu7zVskYsjHuNbsz+LPqonKVf6qjmCm11utYLIZS3JLdo/eKuu8l/96o/2X56tf/Ja+YehroAH/JNUHEbpAs9TO2ibsuQEJDsWFz2GxGTOLJMC27m3yT62OrrMF+a8i8eLjJ9/mh0n33+1fO3FJ5hqR6IcKa4C8kF9jtENmg3y/5j611Ga/l8KqubVMSX1wbdJ/7r2v/IvjP+tJ6Yf1zEpiZhsSykC73704vQldo9c4ijOz/7lX8vv+DqO8nnxxa+Xr3/jlbKH3SY7mERJbvBp/bt9a54vl16HZ+o2Y5cFFhVPnjxZ3n333UyIPPHY4+UHf/mX5YnHH8MmeF+Mk1J1l+JudpfcvHE9nyh+59tvYxvQFcju+z+VwuDzp44XWwm28gwX5Fj56YGbR/mNkrRQJTgW3N7rc/jbJ8HToGpY1ae7Fz/1WQej0DgidpyPIavxj/HIS96TwbjEjFCz5P9D3BeNPwwP6EklCvm1OxVZoPBVJrKnSYzZy28aLAdA8h0eeTTg8OqPkR3/qeyjik1E5wijUgU/MHa+YqzJC2/tpAFZFbfDD/IMGENEIqmZKna6bYibUkTmuaciuzaYwWawKkVuAfNmZwvESuRZSLd1LnFBnVuHf/nu2XLw6KVcSLeAMWsXTa1gNKQzA+WOp3bfhzrj/8o/HhDVamagcQY0eRBUQ0NPfQNEegGpxkpY/BCYKy3MAmjvEXf7YnRU+Np3ycy1nzTzaM0Sg3O/5uLWeA2+ny/bsWM3nVDPCSI//jRgaxxdiNwx1BOsyHzr1efKt7lP44F9HD9CGWs50DiZh8SQxk9J0sPN+OmzK38zh6wN/EcWln9k0viv5e80EgmwdK07GAEBTPn3FUmqhbckaufRXLIF39DKlNFxnqf1c2ga9hl3KyB478Hwxu8LfJbx9wfOZmXu4FG+IU87Njm9h2bO/OmwbWegjTe7fICPPpBv+1KBxzpYc+fsPZeRwqfnN/2z0XS1K4M0YCzj0MNXXJ58iK3Ff/VK+d63+NSaPCkf+YZwscalHpCG+FoecohM2kvPHDB3rH81q03yN7iKLwDSlqYRGXvFiw6S/K1/0uB7JRGJd+EBUWcIIT9BUv0EtQIClzLGnyf0ty+T+Onm5AnUCh0ku3HLlNMCMr504SZf1TjDysWBcoHB2Mz0jrLnPm5epx5ne28mWuisgIIqTnmx2kzHeppO2SKrFN7Afpt6oh3cxoBhB6sU2/jOsJ8mzmc6KddMfNnwOwjjONsb33y6/MW3X+GcLCtahLsCG379kZXKQGUynFb+Iqg/Vv4IKvLyWaWETC2Q+rup/mn/xvADGidtZKLJM6MtnfWd9OyyWs2Xiywjy8s8azI7piGIjOwEehG2q0dEk9ZL7ujk8HWx23O32b2DhjAQtgM+oknIDn4LGlRxYZv9q/nW9NYMfGCWgar/6cR+ZPujvDqyzVBXEVX/4NcdSeqCdVi9lG+ptW1wR9ikFxAzsRn82lDkNckMnBcIb+Mo1yMP3p8jArXOtLpcKVf+7vRoA8kB2t7byKopw/0fV/4RDem6Mq7syj8B4qesxCFj7kLswGp8YBKVdw+gJT18ps4DvMaAZo2zV5OTdGxhRFEwBUy9Qw5ssczuRuD9Spidey881IDMsKigOly9tVh+9ov3yr/84nfURVcYv8pxuuvl0PET5fEnn8Q+TJWzZz3ffiX3EO3Yzs5OBhpLi7dZlV3jrpKHOKf+Rrk2dxW9Y2C0vh3Yq3SeT/JZ4mvZJerQaDuD5kXq9sP37yyLc6fLD7/7evnBt19n4uRedBbpIpJFZlV3bK+lUeWkWjRdl+fOD2zT/6p/yvMO9a+TbQSbTP0xkD/yvqP9JToyDi7os79iO0bdqZ3SaoalL+WCHczKOXUzdl6SgkM42nLSKx+Plnqh5rq7IMHhRpN55PmH968wYDlcfs1FvCdPn2MgxyCOvsEEEyOaY+8imvUeKBh3UmYj/8oYsK6JwedLh78HDv9KSh0QBuMT+0c6Qpy8awPLpAbEwVXF1ckfHpIOAI/8oTSUuUBKoeYTInjZsWMbNpzdB8mMOHDVewy03WBbuc0dJ+7gfL289dqLIUl807EfCAYBeRTQ3qM+LU0IBQakyTdFaNsKXCgMTfw0pD6bsMK/PwYKg//ztv+fAr9J1ZVe/pCrJDJxjL9yYclHIHlXdEngIzx2eVDWrX/tgl8mYTr+3XmoVBSvOukku0nrkTvkSqHcoG6uUNZz2OPDRy+W9947ml0J1677IQN2fDNpMoF+rrEryvYHwxaqQgL1JsSlcbQc1bSqT03+oteF/Oqt70P+pVJCZRdI+2et/jf9V1fUDW2F7Z+LBq38PSqmDZS/JS9D5zi2i2IuQu2gr7iNv2nslNl7rCj9IGQ1wyX0ayzQvMgdJz/9y1fLi3wgQG239+huZrajgA27RUFtav9befSMyYAv/G3UP98JFqI5y6OWf+UpkSbvsqne+ivdGSOoBCYKDh7pM/hW4Sx7ry+Yt1/ADn53ilziYteDR47mvhK/srPv/n1MiD/CUSX6045ZSFN3X0+V69eucX/N1XwJ5/59+8o1PjH/85//gp0qv+DrOudpNx4rP/nxT8rLL7+UCfZLV7x7crXcc49HNndB0EqO9/ixhjffeLPs3kO/PvSG4vAmpXq0fTRVod0yjatsDLw1IL+DuJ7/KsDkWzPwF0Dl3WU5Ch/3bZJ/l7SlG2Cu0r0L8KMDkr2FGyNuAydNxJtgKlynfpFbZX4AGO/gfQvUCdoEMgyolTkVaFOhdHFjJA/S6qXaSuOYs3yJSzRxqRytYgRwkEfHv3WnrTg1kBhd0qOHMZAj/gkRQWxMpVFsDuCD1YGI3iRQi4Hp8UurabroIX6j7EgAr9cKmwv+MDQar3Twycet/fnHIMNK7CymDeg8M9vvv3eEwfHJcu76Op0z7p9gIODstVv7JaF+ehKzSccwjbHowBb5y2hzELgl/6qYhl0O9JJwk/xbHnmSC0agftWG1SZW3Mw3l8+al6mh3+3jWANmYhdyZnKR4xzCuUK0kzsapv0EGDQT5H9svY0JxNM4RRZrbKFevVneee358u3XOaqzb1cazHyBgE6K0wC1w2ZqpduYrfR/+vIny+bIXnmKSVx30j8xuwXTbkzwm66l6dLZEbCLl8ttkV1VLJt+L3isDboTTTZwunRL4des04FkcuTCxUUuGL3Oxa8Hy7vvH8728jn0YhYjv42Ovedukz+fUcsKK0R4JnOWSREbxGUmseZve2SBBhE9dPXU3SXTyN5txeqSCDU/yruTaLZrPvHQzvJfmDj57lsvdhMnSgUEdMAqJO/yODTK0anaWXDt1392uKtwTA+vCnis/hHcnKBm6zvplH+Uv4ZUvFQGuyK5mJZnlX9SJBeyiI6p23ZpxO+qp+VhXvYfxaH+5+tTwGTbdHChyja+/jNL9J9IcsqhsrJI4OmLC6ySnioHD7BSyoBpna3ls9t2UZ/pOGuIHDCQTPmHFvJLn4OKvsLfKl/Nuj3HCgb4J4hwW7cN+LT1K51b9EqakdW0gxTLhZvunTh5/ZvPlO9++6Vy/710jKhHduT+FPofQUdKnfxVYSc5jEhRKMHBex+h8IXlb7hUbTqdFyIbiWyiN8orEYnkx8R2jJMhtgh09OmccPYCWO/s8cs5dl4mOW9uJyXlKHjKt6aTxpH9C8Fd3hVDDVH/0J9mM4GozHV5bOQ/8RXE0lNHFYto7bx54WBsfnTQjm2VkXla/5w4caDg/VTi1y6kX86TJHFhwTYKuGiIAHT0Z3ZMlEcf3Ef+vCtCfoLf15pZqG2dtoogETXjLX6Di59R+yfQME2lf0v9Mx3Q0tiSGOIOIKuhq39D/itQV19IZQ03tW1FHdwAz+6Qsu7kkUcXk1tWhmedjJhSS0yP7NjF6VEI8bOpA51YYSHiVPk//u+flX9//xTtKhOV2+8p9+zdi25NlQcefqjse3B/+fW//zqLGB4jXWKH5wpy3b1jhq8p7KADzQB/8Xp57Y2vlzfefpOvZi2UU947dvhcuXD+Bvh38CWe3WyBR/fQQzecPP3E/nLPzK3yxktPl5e//kS5d4930jjYQgaQ66C5cgyhzSEb1CK0y/+d2p8KXtNvKX9ySPkDaDmSlT7+EIgIrGJN/wIgNOHC9WUuKHpkHSLcwZv6rFXNpCHgqLSFmWydGJ7gAmvbRfWY6RD0lwt1SexdJgeOXC8HD58u//Pnv+ECxitcbE5dJS4LQ06akFXwo7DVDFT+q/LX8pTG3q40P+hDmOl1YUP9d0lB+29ABEAIT3nRVlmNeLGe9d1v4mJ3QgjYTGobSZ1zgct7K9KuA5fFB77SMQvtU7S/mXQ1Af9jxslXeVmv3VH7wmNMnPy1EyfPx/SJ07mZKEPqgxNpEsBf8BvHX+iUEF/kykTCfEblT069I9svXP9ALmfhKWzBV3g1gt5fyq7x7pMyEzypmGwiKCIjRPvvDr1ttKcK14mu2L8O3rJ1V6diVP/to1TcNVela7v9s9+fLYe4d+fYsdPlEjsXvYdm0ol4JkmWOTruheM7uRNOJaoT6JWiDg30Vf2zt6dKpMiCibKTtw7/VvW/J0guSVzpq880e+Csx8A80o9OYnOUl7bFPkQm7GhvFrwQnf6f/UNTq43T9P+yYIY+24a6wGP/WjdJpZti4mRy8VZ54Zn95Sfff50vQT2SOKYOAeDYHPSQMvqt/KX/E9kfF1mCdyA3MflayenwDsOM8E8g/kAs/y5uOeMwyZ8xdtlyiTqgLlJdY/Ljd++9j60+lJ37e7nU9cEHHy4PP8ylr7t2ZzErCgRNytDPxXuh661bc3zh6yptzUp55JGHs8j17ru/Lf/+23eZKL9QHn30kfKtt98pr778zYx/zp47Bw3bkn4nC2Az2AaPvPv54kOH36ff9na+zqP8czSfMlO3K7v1qWqOucZuHzgMqPJXncJ4D6Oni6veLn4AGO/gfSzt4GUTyDDgT4+f9jHVKwzL5Th5m/murA2g4q3v/upGZTCCa772rJDD3xozjO/9vafBDwLire/+6j4Kf7UoQFnyGLTKtcpUG7K0KUSbXwZnish3lC0dCmP0E499TD6+6dUZ7k+FJQZ/BzaQrwB0GmJda/5gqw0miTXII/w1vyy0EOq/Hn/r3dNhszI7/p3OtyHpWWDEVjQ41Ipl8GDemOhxoMtnS29zyef5q+XY0bP8nSinuYzudsEgs1112i9qUMnSKNA5cWAcfhyIyZuu8S+RyHET/waH0iqn8E9YwHnWDHkLEtODB/krAA2iJhcR9A29L/kSRdIaQQro8tPC8xgJP21m+Wh8vNhSQ27ztM4AUAqyXc7CsjDyaUykyFcKysocEydfKd95/Tl2nOzBoCNQ8rXTYpqZ6IfpKBv+Ptvyr/zLu5L5OPqnwrky18uTpDbEdiojMd+lWTbhIzqLnkV3kK0yUVNthtLJXKUxhFdXSpx2uH17vZw6dbP8+rdHyr+zo+HypblMkuRTd1kx2EmejJkw8H6pyA6BxeZ5THGpK8usNFgea/RupclGZIb7SdzuLGFOyuGR0NSfNkBxS/L68q3y1IM76Oi9mokTToqkPILE/h08SL1l6j+ITj5615CLckx4yqp7M64DU9K81t871f8OKA+ZBWfyp8HXu2q6ioV342qjLg511/wjf+IyEcGL2eT+H57Ge/u56bywNeVPYCYmzZc8LC/zE8xL4P7wwdFy8Pg56uxlGlUGTGxk3bXjXjojOzOY8+tXrpzaGU8Hg8bc2+qViNv45xb84ga7TPi33V0mToBxVIrSoMwcJFJW6R2p4zbkrGwRPsH9DFNljnuAvlK++52Xyr57IIw6MpV6hbcTrDLXq5M/fz7a/sGo+gNNePB+QvsbmddcOtSVCMsmhJi79HTP+tYFSK0ZICnNj37SqWPprNHBs0PsSna1/5QdoO4sWeQ88gIratYFUeWOE/C1+xCkqGpjrXUdJimSmh7/iH985JMOIdHqTG1/kGynQ+FBIN7NR6+Mjcs/gYR1HTzK1obBemc9N12drIRLsq4XX2rzgKEOzlgPO/wIIfG1/pGW+i3SfMKaSdhZd5w8xI4T8g0t4IyzoSDAmiFCcfpXKQvEhp8ak1/TiR+IP7r9JRGUgIy/ECRy6oHIoaBapOqvNayDJ411wTB/dbGSKX8TMxVqVhKV/JFlBupOUJOresy9QqL0UsBTXKB95TJ1dW6l/Pt7x8rRsxzRucWXFLjI+Z577qXz+zjHbLbREXaH58nsLtlGu/UQEylPPfYw9W+SFef3y3tsy36Mrdh//bd/VW6yA+UMx3Q8oneDvNbXmPxky/skbd0CX4xzx9hzzz7O5dr3lQfuWSsvPLm3PMMXyu67h7qs9JlYdXA2W0dBnYzUHJlS5p+w/m3SP7KLrCx/5VQlnSKpUd0vFQmBjfBXnxpq+WtP4odm80heKVOTV7qr/rr/h9JCfz2+oM55h8yFC7e49PV4+c2/v8+E80nujuFYIuXoQNTJTb8a58B0mTvNHPQpxxwzFnHwVf2rNr7DCX7/BT+PkDPGf+XHgavchGaS1sUsmyvlQeY+bU+CK1wCg/ypX9FCYaiHtqvZEdbhdALSyW53eno0x3qcPot6FxJtX/GQv+1vWb1dXnx8V/m7n75WvvXaCwoxOGkecPYRlS0vhGsj5KdLThkQZ0GYM/7a/xFUIOVf8zLON/HrWvovzP5/Cvypt6EaOaZc6pRGHdR3pdcJxTfFFv6bX/lEEFWEZlXljzQoXMvXbO0/Wcp00e2WA6S+KX4vX18pp09fKge4o+xXvz+ae47m/AIkO9220R/3MnF77x7XnXbHiWWmHSftnfRPaqJ/XTk125VilkiJRm6paB1s3/8mVrYyYRkcVSnE6UWt1geP/ztxYnuR8QYgi7SDfpnP/p/5u4jmIH7GHYvyKzq0RJqdOHEBKvWO3dMe1Vlf5iLjp/eXn37/NT6l/ijAxNsfh39FZr1REp9K/+BHti2y/OCvzsAqz8S1KqFSpxB5Wk90ZOBi8jL85zgljF27dqvsYqLbccci/bWDTJb8z3/+WXmXIzlOej7x1NPl1VdeKc8+9zyXdTNxjiIsexKAyXbHgO7O0f6Y5zwTJ+cvXeDy8LNp/yxHy2KWPvTDDzyQ9uB+dqt4/8mli5dZ/J7nSzqPlu0cz3EBYxo+bty8Wt777e/Kr/7t53ze+T+Ut/hk/S6/0ENecfLal786a7iY/K2u9/eeLmIIlbgK4K+upa+46lvLoj0r5PB3lEdL38P2ngY/CIi3vvura+k/T/zYtjQLFdmAnlCghiNgXR/VexI8+tkYvuV7C+ye448Bki7bzwm/28siXvBbF9Bb+IfP4BuxVN8/Lv/kYBZNUj3tMonTEvUrnR3jMYCkCH5rajwAV5w1qxH+ms8oun/Hs86ezgx+6MxpyFpbWIfYNuOYHLJaZNb6PNv7Dx0+xUwos9qXrzMQYGA7u7vcpiPsWV+/I+/KqbOnVtisaIgMv4aurboGf/vRIJH/H8W/YujZNnHjv2bqN9xz6RVGSqM9C13idxeDW9ucnZVONdgmbTudCW+jdiLArZGuBpvOd1fP3T6bVR46I6GUu1DK+q3yNkd1vvP6C+VBtjhr2B38+zk3B452VOoKmc1eGOyLp7Ge5yfk/4/Vv9RWZujt1Dcn7xncQV9kIS29bIWkg6RoMfDptDmRQpocfaLhs/G2j3Xi1PV8meW990+WU6ev8Ik0VvE4BjLDDhMn3JSDg0k2YYJHuVOPEIkdUcdmTmLZANiwZ8cSAwFnubPaKY3gt4gje58djekI+Ep26xydeuKhHeU/czns997wqzpGqFWm14sfnPLXl4cwOhpgy68qlUAkCBIfvnf1PmG8mwcgkNvlFU+fxI4fTJMYGH/wKiva/E3OrO0wJYnJ+BPeAVtHETpV+Z9wNxgNJ680lnY4ZcxOs/BwZRhx7vz6gA7/BwePcAHseb7Gof6yomVZeNmmPLG6qq7STaE84R/EtSPMrhbOOS9xZ8zysmXiauwKk4q7UiZelEzNoDy968SBH10zGHAVVnrtoK1Tf9bZlTXNjpNXv/ls+f63v8EgzHqgNXEWa+Cg5ZPU/y3l37IdyL8TZy0b2R46CE4HfxjeJ+gA8+6PjmcltqqKr4SuUAmkpxa5IUbYYbWTYTp26HEvxRJy8ZI3PynowCv67iDHiQf+6g49EiSjDqcZfKj+kbmgHR5xVWWqAX1U7wnE6KeXv6VHmsCZVo8Pdaoexcr2ahQs26TRF62nK4l+NjFHc/DXNNX+ps4Bk+NZ2mKyhc16VGebR3XYcVKFJqLkl2cXJgXRllYpQlOjsVIopX2w5Z53noM8hjAVYPDb+O/SZidnn9YOofVWu2cu5i9+60oeWpcEGyu0Tk60KH4lyUkHL2I1yoGqdUB75qqgu70mOeI4xznGY1yefenSbdLRflChr91aKL/94DgXN1/KZNuePfeUxx55hEnLHeXksRPl4vlz5V6+erNzxzSXAz5YHn/8ISZWdmYH5ZlTZ7DB87R7O+lIXyq3mPic4ysM61zGPUt77b1j7iRbQm8dTO/ffw87LSexn7sz+fzUo7vLk4/sKXu5VLaWf530m6WD3nTkM6l/EdYnqf9VhuljdPL3kUrJr3rXl79lxZ8TC8KoonUCQnjspvYTZthgwhdIbucizXf/cLh8cOAoF8FeQ24cK2MLuxPOblRxkBhbniIlb3XfjG0IRWumQ/y81rDmEWDgmv7FekJf2kZTkJnO+pf6L+E164Awue2g2LasQsOjfJqENFTI9F+0KzM53jqqo/JvXyXwisUMevsPK9ZV2tMXHt9d/v6vXuMut+diJ7VXmTihTfDLTMFvoybSrs7gwxHne2yENQi/NIlno+v5TyaVweQ14v/D7Z95AzvM+xPJ/4/ED77W/kffbJSbsxzgQTY6q7qZ/45kSU3/A5nKRN1hJ0+m1NEnddYES6MIzZTxdbnMXWWHuVPwD78/yN07R4vLFKtMNEzny4/eJ+ZOE3Z4sDA6zcSsd134NcLqzKhz0iG6Tv/01bKUMFzKrXFhWIAlpxO776YJdP3p5J+xADZQG+9x/3wlBogsqHbPRSZ6lllI8wti9re3OWGS3axO/tGGMnPj57zF4YJkJk5AZ//HftG0d5ks3uSIjjtOXitfff5R1AV6s6DB5CGwtiGQO06j77qPq3/KARrsM0ffe/7Ng7/Gv0/5j6v1sdobIpKFkiYfZMKwCVivMqDsKOOTp8+Ud5m0eJdLYC9wPOeJJ54or3zz1fLsV7iglfJTTsLNMnkrTo8ztbHNqdOnyu9J56WuHu3ZuWtXefvtt8tLL79c7tmzh0UvLgxnUn2BrxJdok1wHLSb+1H28FW27F5WlvQHLzPx8suf/7z84z/+t/J3f/PX5a9+8hMucfdyWfRPpNCd4qXPpz2kGMb5l+8AGDGIUiQ1KOH9z8bwLd9bYPccfwyQdLnexfgzcRIyB0T2wsCzgdVhVO9vMCPoPqr3BGYE2IUPAr5I/OASc6XXqshbh7/pT9WOVmH6aNsQoCuUDYkQIy7wxXq1mCBJVsm+A+xxGNjlVTNJ0zRKJGUty+4ZuA6iPWIsqbh2iM3RRNoHj6z47uDNr59cujxXjrJqfZRdJmfcskpnYn3dbbxMSGDsFukQTmW3CisxVGxXIm2wXZnODCmVNLTT6ZDzxqU4DFcy1ddiEjjGf+x3F1zhN8ifLIRZtTZnIMMsNwYpkwOG01GdZ1vkwgKXuCHrHCdKp7/CZfeEvUFksWwjhtF28kQj790MWeXJjhP5Y8cJl6e98xp3nPD3IJfDms4OxyTbpu0gS10d8MIz+D7T8o+8NvAPPqVY5bhZ/2rZ2m1vWthJvDUI0mgjBe/ykfJkwKAeLbuKAU9+UlTdsPM4vzRRLqIXhw8d5/LXE+UIF+f59ZwVztFv27Y7AwBvRJcmL6zKKhbb0zmmykCr0urWzCW+5uIxLvE4655VMRrOZqgzOUKcdAcoPNrwVLmGL7Vq+WZ5PBMnr5Xvce+Mt6qnltGxoyiTllKkHKr+JTulZedugoY5Cqp+Vg0NGsKky7IzPiCJMOAO8keewWsSEti5qqB2gJO404WRX2mYyv5xwyGsfVIDap8MGMrATpo0u4bixIUAk3SM3f15iwmSU+eulNNnLvAp8FM8L1If7Di7ouEZfMoOgtao03b+vYNimskrJ04chHh58zJn2xdz8SurQtMMF6jDO2h03U3GMA84JswoL1QEWtQPdIXeNLUo9iITJ8Zzx8nMJBMnLzNx8p1vlPuZOFGYdf2sSqgTKzzgU1gD2UU6BhnVAfIqZPd7B/l3EkySlp6neSi1+tvhJ6zJtuFvGIKmpU9mhJg8zoAR/hXru7bDOpQnMoYXZb3IqqA2ZI5LPhcpJO8Esgw1U+q0O3CyFVv+QVAprFh8uTP/HTEBIB35gbJ3Idn06hRPowzTNb+1cxzj6C02y84R9twVQieH2l1HNSc0EBvvxEkmfszVGcDO1YnzmOLwV4+WUvcdcLrjhImTh5k4cdKtMukD/DACyTjKyCz1hZkEdj+DgI5/IzKA6bjr+ccT/nluKn/SVIo7IGXVSwl4ElZa9FdY85Xe2oE2U/yUedMjY1wRdS+DnFgvHGyuM2nvhH06+9g5B+uYVS59vckOvUPlf/3mQLkxt1b2swJIZS1X+FzkGS5uvs3l1p5pf/TRB5lkXi7Hjx4rt65f56tVO5HdWnlgP7tD2C2yb/+ecg+fd969Z1eZ43LA3757mPb6HLvM5ljUsJyY9PArZkyUKifpm+JT7vv23Uf+u8vpkwfLkw/fV77xzL7y7BN7y9OP7eXID3YFRZ3hyJEDlpl0niMAfkb6r4CVY2TT+fQr91Y80c8uTmh1LzBmpy/ALSaBEXjSd4AjHMq24bfQtNs6fyuU6erORcQ5Qzx10cl7v84hllnsmsewLKErXKZ5mP7Ne1ymefDAkXLi9Hm2vbsT1XsVdlBuDnJIr+00e/sNLsagu1kIINz2RjLFX7nDB2zqH6TKf9ofoBpExxZp8IVsrbq81ZwM005kh4Dx5gye7OiLnbEyAcDkTZN1cvcFez1LY6vd9x4qd23Kv/Bpb4PVd/NVh6Gqk6MTfl4O+8ITu7kz7FXucnuhkoeo/UhAzhryiK2zgkhLIMwLB7M1GE7AV/s/SiVM8lslJOqwJRFhoMUkl0RKsrITsKY2zoBW/jXtKK76kgRv0vP8bPWP3LVVcgmCTJJAjzgbleEfvKhGxJMy7KJDoVnw7pcY3Qlsn9RW1Kcy8xPuMrzGUb9IRV3Dp64eOc49JiyMHD1+hvssbrEchTTof2fnreVIXbV+uKtjncVNbbHHQyaZQNHcNimL31zVzMi/jyFIB2zkJ61JV4Mr5VvLXxpj0gFVT/3ClOqRS5LxZJci7aETx+48Fr1FX/t+ft3JiWN13L50XXyw7qYApaerIdFV+m1TXLpc2G38Yo7qdBMnys/FTsYm9os8TpIuoAwRM86/PHVEpExN2znxdagNMXWNFGKcf6PiZAZXc+G3y1qcMmqfIHUCouaZrb1x/RrHKq8zNllg1/Y8Y6u5cpnjNt5v8tTTT5XHHnuSSW1sEBNLy4yjHG+lj8yOIhpnjlDPMT67XH71y1+WX/7qV7kH5eGHHuLi2G+UN954nQtgn8yC1zx3010Bj7t6pMU7Y+69by801U8X1+Pys+XqlQvlF//68/IP//Dfyn/827/mPpQflb1MvKSPIl/KJEKo+k4Jha8wPfhRQkqiPQdRvXcUN/L1kZ0nMZuiBwFNwBsSNoj23BCd11HcyLcRLjGbogcBfyT+0cTJENMgvw2UDaHwbwTcKmRDEl9JlpXRFMnW8WNRm9F0iTZHbA7ZnH/O7tNwTMYwmQLXtEOPQkx96gJ5tR6pbGG5T1ANVVY7O1nYXArWg3TZ9/kbSTIvV92En3AJcVDY49dvHqbr8TcMRCQ/6dBAU6FDZJoCDM4UlXq9XL9xm4HYJY5gXCknGIRdunCjLNFzmJ2hInE2TtvuduNVZn09ryfCWj7s8rDRppIbP++FsZmI6IixNxHrWgmzIt+Rf3LtnSzRAtrA1MpLTM2ieuDZ1Tp7GzmXR2TDv8SuBgd5rrK7FXAWw+EMa3bBkO8ydFp+WR11NYcBoi5pSGfj76qTjdyEjV3uOGHihM8RP3g/K1LixpC5bU7jqEFJU2djESVIdvxE8JH5Z1b+lrP5WuZ30D+jBLHzZm+udWaQUNepgF6AhAto+AUWWbQtv66QcrCD7YU03ifOsaPhdHnvgyPZibRWGFzP8HUhGufVVTvclZxcRpZRhZ0DGjroW2M3gls2Fzgm4mesnTWfQG4a78gftCHCRwjvntJGXq7xyYY8dOywis2OE76q85//xh0n3R0nrdeS/mXVGGVeOwukDaP8OLETyYgYAoflZVDgeKJaW9Y/40OIeQnvj8+aFTtnI9ca2EUAGLF0gbKZDqJ5Cc8DcXf5qUuUmY2u9T/dAOsenXlQLVJXL16ZY9X6XDlwlE+Lnr/EoMwV7ToxYgdiQiKsg/xZRfL1Fuh0AOcEzCLbP1f4W/PrDwzIHNDOcA7eLxLZqXFnkHc1eIRI+SvNWlbqDLXXCRWYsKNup2+diRM+ilhef+VZjup0O06A89hfBImMv0j9j0DBXAWKAMSfURB+nPKueqGcugAjmt9nV/52TGsK5ag+Y1eQa8KQG685U+4xCHeX1J0m2AyzEw6PW5Dz1M+f9s87fJpOBxj4MSfsR9i/Wv+7VMBHlTvaq4KJFwKhM/rWgSoFweLwpK2DMHlbcXeE/9A3dzV5j4120/Jv9lObl/SdaGJfCJG3iIsJGCdFPEqyxgTCNnY4PPrAftKDURh/dOI2I3WTh53e3glHHtbgrVx2wiS/ZGCmI/677CuXcLuV/jWYLrk4QntDRny2ibdA4YHtJzyxeVZN26DsEgCH5Ni2ZNKBCTbxrjABMs+EyLnzfHHs/ePlV78+wIWvFymTHWXXvXtzFGSe1UR3Q+zh/f776eByAfHVi+eZ+L9VHtyL3GDh6JEjqa+vvvJS+cbLL2YHygJ1+OyZS3xy+Hi5yA6WSXaAWdYz1EsHMfMLaBqJd96zq+y9794c+/Fi2LNnj5RHuavrWT7p/uo3nigvvfBUue/eHYGd8i4kGMMsVPsDXcPySkGl4BDIp5F/k/NA/inqTs6t/tn/sb4YHPl3nkzOaVRVIMrAr64pby/5TRj+utPUQSQLPkyIHDl+lrtMznEB7DF20h7jmNQNJv+xsLRjU95Rw0KI99PYP3LihKEg9QYmKeesoKOPTtA4ISxBVZ87BnzvUFdiO4JlZBChPtf2R/GZiD/+A9UxWOuRPORYqW2rKMTJU5vizheDcgkwNHuJeo7PZbGDOBOQr5OT2WFsctFQQU1nH8VBu86J9AmOImfHCTs43+KojjG2AVPTyM52CF2OLTMmxPgEyPLH1dIhb3H2vBqTnCS60pRE8t/ZjwZiXkN/l+wj21/xhx6ZC5r+dZRhlf8nb3/MGnumcZMu9SFy8KEcKK+O3gSHpgRXEjrCBPHLdBE7L1FTZFv7j7YnZu3ExwTHxRY5gnGlHDl1oRzhAtgTZy/xdS0uD2Vibzt3l6zYJ0XnXeCynlsQs+wOtd7b/izcXmBnNZfERkHlX/lUIrXVI/kTZjDRm5xJ7H/Aby1fIHpYPMqjk3/uruK1Ewc0QRf1JF/WxEa13ZXuhrL/l4lIgNVH/7mztlrPigATRl5WOuWiTFxEYpcKd5xMcMfJi888UH7C1/u++qI7TqScnSi0U3qlQbKqMz/5/yP1z8Rmxu7k+nWs5Fwzj6zIV/5TiMDJh22iwf7h98WydMLoCveYHPjgADvcDqaePv7Ek9xj8gD94dVyi88Kz7CrZN++/TzrRLeLlu7s1JYvs2vVHSPzt7iPhE8Hn+NOksOHjzDpcaU89MhD5bmvPM9Xd/iS2u49GYd5fYIX0Ht3iiT6RTt3pMyyuwd0qe9evLuNcrh+7Qq7mH5Xfslnj3/KpMm33nmDz9hzETYCRL2wpfCU8scPmxvb5+gD4XECKO5NbnPE5pBNiYLww9r/EDTEd8dMN0dsDvl88GMzLILO4VM5epqNUmPu5AZUVu8gwDTtNU9/DFLR8ZBti05EjfxC8NdaSIWQ10qWVKGMXVhHUAagrUfYiFQcpMmjZ4C04SzcJd8wJ5x8Bq4mympch9j+fow+UbFTY/ibQU9Mw16fptd1+dpYY2orXTS29CaoSOucc1stZ7l9/8QpLpPjkrqrN+Y5ky+tNMbuOADW89jaArcWrjBIcjykcdOJWTh5sNPt5avOcNbOdOMLIOixIshoGzDIs0Efxv+Hyl+aIg8O2NhBZWual8AqeLcB7tq5kwEK9GMFsgonsQwONOSZdLFEQW5HRJnbVbKbVrlCRljlNT71OLHi5bAe1XmxPMBRHSWZc8ZMxmTHDeXvdtaaDuTKRr588IyHn09S/h/Kv1l3/Ach7zrTVNmCs5NvJYlJL3VAl7IQTt7dns0geZYBMfJw59G1qwvl3MUb5f3DJ8vvfnc4g/Ub6IbD4ZntbOdjMk24VbaI+mWCSXf8IAdXWSM3jn6ssOLnV4ycOPGiUY945PIvnsrEFW4Vy4bXjl8aSKUEA0g0MCqaPFZ9onSAX6U8nuyO6nyfozrpRmrp1Ub+q2fyJCkRhvzion9M6EQWUboAk7dyACDKqF+c4ief+kh6QonrwpJhTZLOYvJvgb5U8pMvrzWtz67+CyrJOshIPa8Ep+G1wyIDTl542ZtEXuczmYeO0PnnUjg/kXmLT2dmI4odf2QaEVC84U8ezNg/O9s8/Yzh3E0GC04c0ii6bTMTi6xSWwYMm2vdYOUq52pJkzstTO3gH2cnGnNgsTEZyUCNOu8EzPTE7fIaR3W+51GdXA7LtlRKpsrVlF+c/oNqXP5VIEjeDlrtpFVbBE+WE/Gby5+Oy1bl3+WirO34rLCTapHyWWR1p178Sv5OMig76oOlrX44EUH3L/ZCWUxzM6gldGf8EVmnfxCZV/Ib6p9hebdsKkwe8ovr2O48QBDQ6pL+wJJM2+mEqfwmkbTjdbDo4KxuKTfK+meHXRlWZ362JVWGLbTGk5zLSDn+sN2jOnXiJPCi7kBNbt1O+xL9lywkQ7hIukeHzYRdWhPiNvGfQH6IFkXAgmuD/glXswjt4q8LC6YRP4ksf3juqqVUJY3VttHnijxH0Al30sn2wstd7aSu86nHm+UGd9zc4OjWuQvzDNzPM2jniON5Ll92ZxidVLetWx/37t1XHnzgPtIzacIng7exA+yZpx7jCM295STt8r/87Ofh9amnnirPf50z8PvuZ0X6OrtS+GoOnydeK94TQMeb/KzTfj3L8trO0ZN79tVJk7mFuXL21DFCl8vubWtlF3X2B++8zC6xV8tDHOGRKWvtKvU/d5zIfxUDDBLNv4/UvwoYOSn2O8o/kQKQr2KN/H3RP5B/B9frBIB1Qrm2Y9EnMyAjd1moxvb8mz30KOP5i1fK2fPXys9++TuOH58pl6/c4P42+gnwOuP9ELRNi+yMcgftDEdHbasQRAaoq5SvtrENZpyMcdFEWYj1j9U/6UxKZQu96tGQ/9qWyL+2Iz22wKiHwQnwJOXrQpW7v/J0O7+ClB4z62RWdzQYU3Gpm761o9S+riGwSS6Hff4xjur89SuZOLHQMfPkbUp4pc7XSeBqL6VRFNXR/iNv2/xKg/j5ky9DRBngDfUvkTVJwANXE20sf2WkE+sXr3/w3skt+NNQV1pSOUJ3GJTlyiuvScJrnppn3AoXr2tL/eJadlo4YYCsLUfhFpm08zP1B4+cZ5HqZHZDXfez5B69ddcBddx+6/oEuxHsF4HRSQemtzLQtq8rESvMeln30yqEPn7Aqz6p2Ck926bUs0RVOqOM0gwEfype95D8pNtK/tp/6dE5aeAXOl28dOe3fTzbEXdO1Ak+sTsx6U5W+iDWVdoYj+nUPjm08t9FTxcm6WJCMjzSx+AAHWdWbpYXOKrzV91RnUninByWBvsl8pyuiixLD894+FGb/RW/fInHv036R3bRORZOo+tdJsp0jH8SVvuTBiC45ME2MpMn4LnFhNc//Y//Uf7pn/+Z41ML5YUXni8vf+Olch/3kGSREhncd//9mei6zTFL0/sFQxefb8/PoQ+XynEmec+dOYtdn6Wd2Ju+thNQDz/8MO0DRy+ZNKlHgPgyJf1tv3bq5JWXEO/yglnbA8sCm+Ii2TSLZLpb7IA5c/IUuxUP8Fn6H5Snn30sE1thVwHR8LXxp6LKxAme6ERy4EdgBXgnZ8IuunoHAaZpr3n6Y5ByxUO6Fp2IGvmlwY9uIJ1NHGxkagAw8PYMf1JPy6s9B/mMBw3eBt4B+B/llWXLO0oUDz/2oMy7c2NoLOmmUdon0yRdg67PYVD1pxr34u2wBtj47A6Jh+xigBsBFokmU7wADvFrbUzjnzThVjQCYYbOL5VyiftOLl++VY6duFqOHjtbzl++WW5xlnKJTpP3UUwx8WDH38+ELdlRwL9tB8bJLXfgirEx4+QpGoiAb0wGwyUa+c6QCjJ0IWkYMBBSjRtBxBfeSNBV5PDapRe1nwB2RnaByy39xKMd/Vm2rnqPiTPcdiTs7PhsZVo7EDUTG30HOhEWRktDGN78HDHcrHnHydqt8s6rfFXnTY7qZMdJNY4TTpyQRrOZTzu7c8XEn2H5R7wRBD9/hP5VPUpRhVF5YQYIvw2xBOKQ3yrn8smYxhY5sAo+tzCBsb5RDhw8k7szjh2/UC57YR6jAxuyWc5fuoNBnbCj6YSHimn5q8kpTuI8JmVDoVrYYG7zXhkadllp8q/6SBJlhvM3uUS/7Cjy3pU/Yg7/PtdX6h0nf88K2fff7HacyJ/y57+NVquK6QCYh05irAdRIqm19oinhZC2oyLg/DT8d9K/Wnu7vE0r7eBPNuITd/eaPMxYjF1Y8AFXL0EzXGrQQ2Ti38LCOhMkS6w2XC5HT50uR6yr3D/kuucUx6RcJXVrvRNXfpcyOyIoYvXQOisuV0+XMjAgIkuJ9ciag3c7NR4DEOc6cRQpKZSJe03U7yqdKn/L23w5OkQvZcoBBmXtjpOpydvl9Ze9HJajOntMT70ALvzLcudqGdeXvrwBUkztvSXKex/B2x+h/8msZkjOZqICDUtXfPJWf+tzCDGuf1WhyAd5ODhzksTPC+eiOzolqG8NKwAAQABJREFUmRbxEjdsiXzb4RKjLjpEUt8rSWgdRaPfn0ZmffpbIfP+afg3/w2u4TK4+d0W7OW19SiO9yVU3TFeqJQOfFucvETHjGmk6VfV1rUh9q5QItjPgmE+R9zuOLFNMFHNMTBdNz4cp2Nm9NA1JO3ZxWk7xoJS18if/72g9QffKMNhkCVfJd3yaiXks5MPPcfYJ3uQ0T/kgSx0Fb9tC4MAbOoKuxqc6PXI3DHr6onT6AF3O01sZxV5uZzi8tfjJ7gM9iarx148Cvgaq6g7ts1weS6XvdLhvnHzctm5c7o8/9xTZR/nzZe55OAqFz3/8//4BXVuhUsD93HE52EmXmb5DPyVco1jPsvUfz+B7YW9U9AxmdmDKb7Mc1/xM5Xa9Ru3aN/nmMi5dpULAJno5l6L+auny3/88Tvlb3787fIIR3dkyMsmNV8zrjbKY2USj7zx4nvnWlReP5H8zc5SaPKv7w1Jw5/JhA5ZHaioNdrxqp3i1376NRptoW2ZafxE66mzlznK9F45cOgEu/Qu5/JdTCH1jzsWvAuA8nHV3pV6KZlk1d7jiPK9xiSUA7ypGQe7dQeqXzMSrmpH9Yk/7R8yUkxpXhuIRA5ceMo7Eb7kpwPikfYIXgzRetubijaaN/J3Jdpdgd4NYXvqu30Y8Ts5m4WGTo1bu5q6Yn3hr6KEbisoLn0jj+pwx4l3hr3pjhOyQRXRJ9LAP2+ko58Q4v2paaUxu0x41i1KCbAwKkjgjRy5YVD1K8PKb3vfWP493j+B/mUwr0BwrYzTpwjV6EuioFyYFD6AyE6v5ZWClDFcdIlSdRHCxTvbVPur9iEvcBTaxcsD3Cl48sx52vhb3FeG7L0/bgftPEfJlkjjFw+n3AmBktnHTvmDxt0n4jW/aQx1dhLF3oZAsPvs6rTERLskzL8KU30b3o1KED+NH4J0LUpdUjbiX2R8sMQXqZRb+5KOkybZsSgcPLT+n+nso6iK6YtheLKY5j0n2jLqHFmCh7pNXZyeYuFvicthu6M6L3LHif1/RjdpbLyPT5ZTJtLsn9EDNwyq/o36J3DFGr1sTFqgtUIST4z1KT7ekLc8SIpyz6fn4dOvKB0/car8wz/8I2OsY+X5558vX/vqi/lajgm86Hs3R2O2sTvIC3JvMdHuIu6OnduzY+Q4abz8+wg7TFyA+PpLXytvf+vt8gCTLtZbryCoR6mx/bQhy+xcvHz1ispAvru53L9eRCtHTpw4UnGCRaEsssB88yb3OoHz5o0r5S++8x3aFi4RT5+x4w2ebAukVd5QrTHX+K+Bg7eBdyzBJ3lpebXnII/xoMHbwDsA/2Telld7DnIZDxq8dd46cTJIUL0DwBa3RVCiEj6I7L29p1fCUUjVeZV7azeE7CC2CEpMwgeRvbf3bMKfigHyzmYmG2lpKVStWuU63Dyq7axQDTaxVLp6wZahuGRiagcsNWyUL/lUqBhf2y1pUaHtWNpBUIs91xc9JqzBW981SsI7s5jZUrHYGcLYrlLBWSAt19g5cP7iNe6sOMvEyQUqkKteDG65FNJZSVfOTLve3QYUQwde65xGvtJd+W/4JNmzjQ6uXcGUJnmK28A/5CUOqsJ/ZE1YZq3Ba1qNqrHW3MQT5gpKGnwyyMwudNrpd7Do3wyrMLMYI42DsDZOuhhljHE+VdyFOMChf+R6CgA2XmKFN/jLSg0TJ5nJduKETubb2XHCUR22ONvipRFk4Jgv8YgD+WSVqDM8H8Z/mA+fH13+FndzLc/6rEI0OjLxaRD5WkZZb0jaBBLpdkYDPNbBU/3gucJXFyaYKFtCzpc5I3/42HUmTU5z8dTJcvLcZQz0dNmxe28GyjZ63g9j47SCPC0pB0rK3jP9rl7ZdxO/M94a3W0Y9x00DPRTwWXnljiMfpU/NJBH6GBQkO2J0N/KuRZJLX/1TL5hgoz+P+re9Euv28jTBJP7voqkKGolJVlL2ZL3pWyXu8p2VXe1qmf+zpkz82G6z6nT05u39iLL2kWR4ibuO5mZXJJJZs7z/OLizTeTSZmyytWeS+Z778UFEIFAIBAIAIFpfJysa/+BUwD+uhtOLBO8S99Anjx7cQs3yQaWnaRuvwIgz8PAuFpRUpi/KQuOcSsZQbnGvwcG8Izfr3w3TbjKGPW1FC0Nl0PPY8bwn4MAFWV/bdfh0xF0ZgSYtbl4cYrZ6mvt0OFjLN+9lEHBBLNPK7J9bkW7yVJOjVc6dJ13+Ql5Wj8qJNLfWY2c6oIRS2PXpk2bYlB0u50Gk/zDYGKdxXEY6eqfCjnEpEKtL1dXmK9OpV29ZV6rcXaZPfRu1cE57Ov4OPk+p+ps30pJwoN20lyU90+Vf2HZyiWU6fUjzazbRfTnRf7ndxQ3SVPfGjekC9/rgThD+wO/ym34ZA7WrXQcAyDH3GYge9cVJigpdVqOK26QG+S7mpnDFSh7goiUJDzJ4b+Atc4H+IU/9eV74IEBz0EtIeCwBL4x+/dHK/8Q24zgSfnBq/svEZrltn6j6Cq/M4td9WYyt0BEXpAux0wmA9MVxcyyfw+fk97pajl9gsH3HLRai4Fu357tafMkM0HqSZ6QThCBO7wlvLzxwzX+XCELv9UPPWr9P4z/wE844Q+5KW/UeVVYuMP2hFAJLhDEt+K7ofwq+dDW9nx/Fh8a+Lc5jW+w3/zuDxhOLrBNhtnE9TvbDRzAnmW7jn2tPKLBEnNm27h5NYrwNo4IXtMuswz71q2pGE2+/OWXaXMznHLFUd+s7PrN797Gn8kMe9U5pnLDVozZk/ThV+jX8UeCjJ1m//ta/HswZGpzrH7agMzduWsvK042INdvcArHpdB8tduvKN8KTlK5N3m2/fRH32w//dtvcaTljhBcwylZUkrraDH/W1O9Tsbrv9eKbWX59vcw+lP/phnon36XzITj1eH72Su1LU6U0ZoznjUmUvY7Du7VcaZZ5aN/tg8Pn24fHTrWDh8+wYqTS23NBk52YnWevlwcoKxSRlpSB/oIN7eP3kV3Uc6Zp9xpX19+iZCDfNNJ4/q1nCwItJLXIlDXcuVP2SjIIvkn5gE5yB+SW8QCa7717tYNzdfSAaZBX0Ey2z7hnTJ4Y9gmdlGCKGZi5ZkbRPNfck0GC+2fwLRF6zinLGI4eZ7jiP/p71+P4cSBJ6KL/oDco4ACwb6gEKk8e94AFYrA059B07zyk3BflpY/aFnO5cofjE01qv+88NPpUndL/kXb/x+Hb/1btqGUKZQSwCJkVWgVvlDkWayifynLQ4oUNnlYAmWk/CzfztI3Oyly+eo1HO6foo8/zgQmhtB7ymH6+GybgF+JrGFU/nbFrr5S7rN6xXpS37bdRK/K4Bg5Cn9Io/rhHvrb/yU07Q1tHs6v8o/3OwO2pl5Mf/speYq/Tn/HJMYXtqvNXX3pRKtUsC2tZkuIW0Ti+BU5aVuvvhK+ho89paWHURLyrb4oxjrgeQKgtBL0fVecMPE3wak6X8Jw8nc/eK299PwTwQktR0KEZy2L+RaOloLrc/EfAKlzt7hpcCaroheISOfUIbg6trA/9B7H+9Di1vQ0Y6vJ+C9x8s6tu2eR6TeuT2EsWdce38eKDoxI08Szne5g9ci6jZvRJ9ThHHfdp69Yn1Xa73K08JtvvdWOffJJW8dk5Zde/BIrVV6JA1hPBJ3CEO5qM9ALfk7guMJRA8lGTs7xVB4Na3epE+VUHTAAjeib7VP0qzLLtqd7GOImb1xp3/v2d9j+iZ5OOeTu5DuIEsvs++KJDUOkx9i1TFC+Jnzs4+hx9JD8e531XBe+jsEYPS7zdZmg/5XwkYcMW8J8hfVy+PWwfjdmf+73Sr3c73iM8ece1wZLQxoounyMqsbxb/2533tuD97HYwzPColRxIJvh2qYyoGdhFehVCpXrO8EL614m3FwH9Ik4fDDF/JUjA45KaQtKyFC0IhRjdP9gQge8rinMKdh20fWMuqeA99ofM485JSSCCe2XmA4mSXP2zTQC/guOcJy1aNswXDmeh6DyUqs2itXcVqMgwUyzQBPPJydFxEUXAVmlR/YPHzu8pNNlSgPyS8URMB07BV69rvlQwBhGSFMKstL49cYlPIjrO4zUM9xtu6hZ+ZIg8maLEEjDunE17gZhKMkFXWrzgqDjg30z0fLrRqlKYXnkXNYDSdTnKrzfPsePk4ew3BidzMPDAWmacw9sxHg3+tOGEKreqRceVgM3zjWdC9/3pfUv2HmZL5L+c/OymxzIpC0S1qjl3rPLbkHaqbBpGNBM8xY95gln2KG4NKVW+2td07EeeEZjsm0816FU8IVdOAqmh7PbIdk6nC7vm6AmeNsMXrcyaw1vjMYKDmQ98jnVXE06gCBDpE6Gy8n4HkXA2g9xv/VU6VCEqNTcLz899mT/ZQ+ToYVJ+h3FBko0j9JzbmSpxsQ/R6u4SRfDfDrYvhFQ2ulYnHL0zj83v7TritClYBEaScOpAJDLMSFD9JOt+9cxef2SuDr7Kj8LeP7TrwZVo5MsU/1ODOkh1D6TzBzfRvHuquYpc5xpgyyastSlVOFyn38bhtIiQCnUmbn6GosSM9AamNbz37X0ChIOCxATohz8DVQQvEHTqEA8qBTInztN/7keMuZrTq2NQZ4Eyz7j4+Tb7PiZNiqo9pP9il/6C9DLrnMf7z+Pw/9zcr8/ZWzO/9XWEozwBdvYkDn0D4wCTMYimVWkXveuEX+1eeUM7LEwRYxLl2diuPXNCfew0MUTloaIXQi1Ni2FNOEX4xqnFz9rjHqs/jPeMkh+S5t/+G1ELbgF48XTMH01MpBFbnyawO1ZYhcYkktoSCq4Hotpn/PgfTLyJ/Er1Qpv9WbFNKZcMTAyDnsvj27KCuBRuAnN9+J6ZuYlGKWHPwwdhX8Kl/PIjkEjnl8dv1Dk8CuND3jvPHjdgyNtpY9g3JoZXTzlFb2pZLM1kCLSb1KT/teB8Qqn3fS3mbbB4dPtd+/81EckE7h26ThM2QG47M+iDylauVKFeM7GN6mOR5yY3tq/w76L06xO3ORbTTnULZvt53bd7ZXXn2lHTjwTHyGHTp0uJ3Ep9H+J56G9ybgwWkcC+poEozor+U9TzGb8MQwjiPeiDPU3Y/tZSJkc1aknOPkBcviEeMun1cir52Ybbs2rWjf+gq+u75xsO3ftzNkkSaucl0FLRaoVfRfyn8mSBXya/4Pb3/L09/0bjW4jy5jHyadq3bJVZ41AledzgYpI394tzKpkMh66kEn1pbKVSPXbtxiVd759ps3P2jvffRJu3qN04uQY9k4qFHTVXJp60KS+4v/AkiIfVadR3mBH67SJ1J+gOpIttp/b2sPL3+Vw0zIkDIpf3J547ngh3pDuKWWDuIl3YjDVK9Ov9VvNJyo+Kl/ZAICHtDxtzxg1rbBAYIQ8+y9wNsvVZ7VD0Bj+3b0mxf2b26u4Pzm688TF9igVM5hSeNqBvqYKq3l5ymZml1BieEfopT8Kwzqi6CFKQ49Ec/DZZ5/fvn/p8MXN1eTSabIr+BNnViX/MtR6NSi/bElVSarG9su1aVTupCDuFYzIepTd9Bbr7DF7sgnF9rb738Yv4LMgcRB8Twr1FwVbty56KHkAg+rbwnFui6H+Mbw6vznF/EqOvsbsvOrnEqlJkSdQ2xtCcSgDmvlR7Dlg/CGsgDTCc2qV+QM7c5ipLe0UonnSYm3WGHshKnteDOTM+p/Vrq6uBOWmRQV6NjV4af+KV9hK571WLLWbMSFiVG2KLW7nAD13GPtJ/o4Ga04wSio4cSyk7YP8IdsDBWV4DoGPo/W4WL+02iC83xWsGv4Sd8JbLdeOv5wlblGTMcf+iTxVFHvk/ggOXToQ/4O42Pmdo6T381Wmn379rGCxCOAV7GK5BqG86nQZiurAV01BnUznpvmeGH7Gn2SnDlzvv3il79kC82x+C559ZWX2ysvvcox9I/FsHLb7e/odRphNN5oHHFLj/37VvxkebKOq9KyygT6W3gNxeJ9h2OJb7DF8x4rg9Zz1P0sWxbfe/8tJry+15597kmMwp7SJLHCBqkKn9MnWEVcna71Vr89rN/H442HjadZeB6PMf7cY1j/Qx0StHyMwnr8W3/u957bg/fxGOPPPeafBh850Dl+yDS3cQDj4QIb/9aBP/p9lJoH6DVU41i+iTCKtQBvFDR6eHSgYzEre4lV8GNp5MV/waYi1HN4rMIjm9JCB5WZb51yRrMB+7tQxWOP+VTfjSGPFvyBYQb4WX5PhPv63rBBkMS9k+a5OqskShF3+amzEu5Xdqb/4o27DMSu4seEPdY4nXJQtRIHdasZHM/dwzLJoI7uOVInGEZYUWIUiXQYCI+UX8kUBBNrqJzheegY1CrArEjlJ/7EU+bvHW0EKZ8ctBhX62hWG/KTQTZCxHBP+nDPtnnoR8Dlb+6fzGk2KA3u49PCrvCyE7P/Uo3s8AuJB+EnQ+GDUqgNgioRGk6CrB0WuOnjZCJbdQ62b3NU3+6dOEW18wDPnKqDuI5pwKI5CAYnu7YO/4vXP9gV8SiXxARPC0k9uKwxM8UoVnZY9u5RaviWYgX4kAYBqv8SRzPiKE/Mkv7Kjen2/kfn2q/ffA/P7ZcwgJAW5X7lapftsaIgxiHoilIaBZe8V5uPBgjocI+ZgFmEsQY7G+tKljqvWecyaAbpCO/s+wdWWENSmzbYGWI+3ArFqgzy768+yDLFf0arL3N0nk/u1XDCqTrfcKvO0NlCf3XO3hGadTIwmRcBOVWn3vglfmhb3xJ3gLLQ/h6kf6EhNqYnbehMjVNO+39MlQm2jrrfllJ6qKNCYoBCYuBjdzIT/jjBCCejZ1hl8v6Hh/GJwFGj+DFZyeoS1AY6WOKyDS+dCETJTAf3KAp8cwZVY6JL8m+7XJZ/q0jrip91eGt3YOLRpMIsWlLXtO9CH/ipEvGgZN6kJ3E7/5lfyEW4ivIqVpzEc/7928C+015/ta84oSRkmi0r5sMVMg2//W3swxBUsUzyKPRPIpOEdtKfJxEUT2HxP2XLO/Ir3yr3GKtMSpjs490VY0ax9Qhf3073yMDl+frqmUFJ0meCpxdUXgWj0pig0paMCyMEM38SR2aGbnV5B88UlEc+RRgN/L+4/HwjetVZEEw+Knj+M28zYA6Pm39BxKCUX3nhqQdux1E54xXZ78xkrXzILJONLEvsC8qjwA+AwLYs0BE+LrqUAS8iSR8na1aw4oQtI8IwKleg+Gp6frw7VPCq3OohJEvo2Bci9DYeug0pHl7/wZRYAUZGwKNe/Wdf47WKlRj5ilyXN+xFVTRVnL0cAGSACkIzHvmNwuqMn/6brrGi48IlT0vAseP0vfaH9z9mW8ilNgM977MVZBZ+WcldJ6Qq3PdnMW5sWMGRw1s4hnIbJ7Vtbx9+9HF7+833OZryKo5cd3O095fbZmYNDx05EmX74PMHwWUCH2QsscZ5JDvFmEVcy0BLIyp+dljCPnt7qm3dtKE99RTK+trN7dr1SWazbzjsz3bJnKSF76kV8PULz+5pzz+5o738/J72wjM72ratKMxDPTjDvRYGsOjSv7f/4gwCJEn4mGeJVgGf0f4q2jj9K435+J/6MA+B1VNgyrcakidSsfUsThpbEH4kZEaatM5y32bUeYL+683ff8wEwPvtDA417/F9DdscpLvbePRjUqpswRFz4Qs65RnBV34YpMwY4CeaOAJ3zsGFlzh4TwaE85jGuFj+CNNotrtRXPKNI25hkK7T1uQOchwQ0YHSD+vPxP5dow+UcgBEPvYrGawTr/xPkVD4QoJe4WYeB/bNQ+flGiQXztnSoXNYTtV548ceR6zhhGzIayU4qkulLVjW4O9dGJXejiK4ByxPg0BbCp/PxKvf4FiZLTzmU303Z6D8q/FfUHkI/PjY4BtDZ6IVXe0bY8QrKocU2cKl/si2ZDtkexDpYpVIY2yD1CEromhcV3G4/+npy2zLYSvGUY/CvsJ2nC1tx64n2CrNZEcYxckJ8gBYbXklt1QmecPLXeaHpIEikiIm9R6s/6Js/RbRieY1BKUOqVdlnzlIfEsh/2UiEjlpnOiZRLBk6nw6P9Vg4mTZanRF5dFqVqWaVt20DEqUHT79zPonftpW4BPTVwyi0l+DRU0KIbvcqvMMxxH/UMPJE6QSW1ZTu2U5qrv6yqOUP6UclV/w4WvKJd7F84ytBqNu+DGw6BOo51nKbP+pUfMeE9K//vXv2n/5z/8Vx9sX2jPPPtm++53vtucOHkiV2E9cw1nr1NQ0K082YLBgxZrwgCK9PAH08JHDOAA/Ab3ux8eJ8Dbiu+TJp55kG+fu0Fb623ZXA9eVvzr6v4lz2WlWnxiuY9f1OIKV/mWQZhxEuPBXMEaaZDXkHfQYL17h23lgHm3/7b/+5/bjv/03bAP6BitVtqRMIQfxVAmiu0KuzvMhQ2//yU1aEnG4jT3k6+f9Gc8momaUwTicUSy+jocbefzbKPEjP4xS8/BF4I8ZThbDHgHowUsA2c6rgfeijKXoj7mHJYqXel6PUPiexXiS8YJ+EfgRkgqSZN7ZqCClTHxTlPilX/LSnJ2bqfxPnCp/lc8GHeOBCcxYgaCwSwdlurGLMJf1aemjHWdJpUnWotEUNlrBa7sEUcgXGOQ1Z6dLh5t3PlzCX8XZC8xcH7vE3skrOEZDwWIQvEYP3Tgh0ljgzAUin3+2JpqsSAd+ockLkC2DF8ioyBAnCoGfuFIcfh61/M7wST+3dwQk7zZQFfvCh7NbWKIWCzfluovRxOVofjeeQkvFYz2zaBqQDJeW5d7FehGxYSAUYVwCdSn9S/mwxii3ApO/zMbQCShYPBXGrTo5jhilIlt1FK7QLUYI4qvwZ7aSjs6l0/+S9V80L/pL57qsa+0kdhCgIUzrw9k7AhwUibv8oILv5YycM0gOvh38Xccp8NFPLyHw32knz3j09HXichwt23LcBhLfNs6YwktrMUr1utZIZHntNGdwZjXL7KmdrYNyDVwrsWI7MyZcHZvWPnHoxSAknaeKp1cYhjsg5LVeY37KNfB/lb8H1t1TdZ7as4HjE7/SfhDDCbFs7Coa2ibIzVelvk1MWL7qh8Z2XQF8gOceDf6QCSm9IDVZZD4y9E/mBGksmItvAHpwMg4O3gMEmlhOcLTVyq51Ao4ywv2w99uF8zdw/spxzydPtbNspfM0jnlXg63BWVhmFeEv+Y7/2ecuvYGk4uKyUA03t1FkZpj9iXMwOlK35qzTWSW0uMtstjOuXfkeyR/4ougvlXySr/hjNRJYDn/G4dtQBmOs1LhIR59Tddiq81W26nzve69yHDFlgicmqIzQqlICTpqYH5egHkH+Ga1wqKe8DfSXskNuFYsoy8sf5ZvgKoWpbBsFv7Iv+ddzq7IyHkXR4KQwjMwaolRUNAyuoE2QOHlW+ao4Bgg/bzwLR3r55IBiVP7wqvCJ4WdpalweRxdhWdY9Clh4qHjOXFIuZQBI+KejPR0AKxzSF8BjztbLH1Vc4Az0V5nyGHKl/p8K34FYL7/YAQr+t+6R6RYPWszT1yCiOY74MXiYQDUx6WG5iZt6IEh+ys5QM8oVwvSXB+9mFaotfBLm8vVf9B3RnySQSBM5Cah560EywB/2I9LUAjhgcbbU1Vvia59pgSWh2ztc3XD4yMl2joHPY3ufZhJiazty8nz7Az41LuE/7K6KP3mtZ/Bunzt59Ur6qyf27mjP7d/FKQZsa9s40T4+/G67bF6HP21Hj52lvW5rW3c81qanWEVCfT71zP62aeum9umpTzkJxtlC5Bqy2lOtlAeuKlvBMvb1nIq1H4eyHml5HYPJeY4rv4tVdjU+TZxKuH9vBqV7bbt+6XL74Xe/1l5/+Yl24Iktbdf2CXCkfPjycEbV7Rtr6FsX+K/qrFM6dQ4Rvgj9qbqh/fEAraWzq2GcRc1FWMFXz7H9YoyjPmAa4qvfVP1du47xH/9s73xwpH3wvicWnUbHmYoPmbVsadLng4MF+z1XR9puc5lBlz9BhmzrA7/ihMGAWzUXHrxMk/5LI9PQ/utLFYfPD+M/2589RtQUfmztqzk62KNjheF39Rr70QzIGNnYRl1lKGK2by9TBg1hwVuG2v/zGnwfhO8XeiBuqTGARV6YElLoj2EFE0PPs+LEFZw5VSd5O0FikWklAHQAVsQgUFyIY87pD7hbh9GX1EV893tg1QtfF+Tf8H1Ef9L6Pen85jWq/7zwI7S6/hz89zD46ckBiGQAuLHUf+wtiv/lkZSbNkMJ6a+dcOMbNIOdkbFVLvWyu7fusdL7Unv7vY85YeVEux5fIBj/aGvzrOx1C250Aynhf2kI3OLYCHOC+cfqNduB34pm0g+cgGmXLY75v5T+xM9qSsko3qw0lDHMx/ZnXup5uQBu31Bv+VJ8Iy3gXycv76CT22bVRVxhsp4Vxo7IPXlKY4IMZh+Q+hrv/wRgxr39Wf+0p1Aq8PmYxoehiWjF+1lThuGEU3V0Dvv9r7YXnn88ZVa77Vt1zFODn/n/afwnLhiFcNK9Gl3JiYBqs7Zd6gL56DYk61jZfnfmPtvaj7bf/ObX7diRY209qz1eZUvNiy98Cd1rMzrDHUiMcf3GFHS5y0qU7WU4If08uF44f7H9T061OXzkk3YOPfzxxx9v38bo8vKrLyOXd8I/GH/lD2jq4RdeG1jBYnu85goWtga5TV6nsVuAZ/vP0ew2eOivEcXyzNCHXbl8JZOlmzZtbBsR+HdYnfj7X/+2/cf/9H+3N/7xH9sbb/y7OJu1vjQYWx9y/cBRdteLLiEsCpLmBPQwqzJ1T7yKO5aiP+YuD6b2x/LvEcaCljw+EIOAvyT4yIVqTQ8Ub5wySwrl6wPxF8Xpxe734WN/7feB5Mvm92eED69R6ZZghMhCeTpnKGzgjMyygYvBxu6Xae0MDRwPr+/GVzDJmpVhfo0bgVLw0Y0RwGXcUF9fhxMgL9ThCC0tvbW9xlwWBJ2D4zPnrnEM1icMxs62C1dZHnyPpZ4oW/qdcK+kXOYqDlPqSDV7fQNfgQo2CqBgPpRhoEeW/dmPUn7L18svXuNXL39xz/gXxD4KjcvfXF4mfBUGBVIaq4SQ/sC4zeDcM88VDs64rEM4ZxsI+DvDLooRbhlEAHEAFmFLPuLXrd6LMSj6Iw4BNAwQqXS9dGt5mHcE7kBrDu/77AHWcOKKkz2uOCFNhGg6r6KBIqboL/J19fJLQim5+Hq0+jcP6ePls/Txbjkte5Sm8Bhh4G6HPaIBnY/dn9tBMkgh3k2cjR7hiLs33/q4ffDhiXb24kWELatE1m6Mk6o4fiUfFSoy578GLQZZ0hH6uPJkFgu5HaaOH+3u1qCw55hh6JYVDODqCQfkQnrxBb6VqdI7lKUoYthy/D8evLj8pp/3OGJXnPwY57Df+FJyTe8mHP5b6tAbHJxNDwiylGDxcVJfyYkOKc9F3/yScLz9jX/3OeJ9qOKSEZaPix9bX/ziCDCDdHmC77SzqP7KCw1yYGypTXGXNjDJcvJTDLiOHDmeWagb7NHfsGl728hyzntzHPTLDLezwPIl8wf40uCXUWbGoXSoccTrthzqRBxVvNdidPR0D48KhwwoOXWsnVsyYjgR4VBOXHv5CTMyBiZnpaV1/2J0lywrmlRyFA9gQqBtYYbnB0/VEdsvKv8+i/4qfY8mf82FAthguKxFy1n1L2F7Kakr2N46uYsvE2WjxhIdwVrXrj7RSaRKizLH7Dr83uaEUFC4E0F6GRAZmS/CFmbJC/lkAf4Qd2gWVsuo/MlDnAtXgWfQlfoyFH6gvVqXzlop89JekYt3UHI9yUEZr5JbM9TkDGiVOssT4yvgg+sjwg+bA09lJenIzzKbsfjo40TGXcOKk8d374wsCAlSKusOPkpLsOVERCUfki7c86XKLC16+UOXR67/yo5cR1fkp/VYqjn5grM0ZQAUP2KWhX+eeOOqTNuwWxo94UKD8OEjR9svfvFrnI6e4oSbfe3Vr3wLJfYuSvDbOUVnDbOKGrE0JruU2pVZLoPeuH5Ne3r/3vb8c4/ji2Sa8Mn2xP7H2c6zon3wEe0fp9zTnKQxc7tWgb74Eie5sVrn3XffxpB6K/hr0PEIy3l8qrgs3NnPndu3tycxmuiI+yJHlF/G/8kd4HqigrLffs4yq4C7zePbX30lK06ef3pz27cbA+tGnBWiZMfxM/HWuB0FeoTOqZCiR+g/sOCjt78H6W9I139kYbed2iY9OUL564ROBu3yIm1E+emflnF9LMkvl67cb3/4w8ftt2+9344dO822JPy2wBMTbInSX9ssfHj7jmHoTQzsqs5HyJsxf2HaPOULaWIbQa5Z/i4ziJj0fqTHS1rbz6PqP6G99OdP3cG8NcIK36G4csQ2qrFkApltGXX4XW21cLYu5D3/kQu8qnzhDT4LQURy9GCc+i59O3xjCIv/XOpg6FX4u+nOYb/x2gshiyskdBA8n2VktlUHYH4qGWNq69/Xkm22eSGa8QK+fxnyf4EqwTs/C/zX8Q3W0kqhTd1aR/KcK3WtJWkuOyo1vEsQuaEeKz+n0awjIc6hf0lrDlbBV9lke/dtVpEeO9MuX7lOG2fCaQXOO+mjPTFHw96c8hD+Up8PbdM39BU/dEwyHD3txDx8o+yCj4r/BOc3keY2XOP1nyDKlgpLJOpLOJaM8PCkn4fya7SU9x10qwN6KQtzWg7yw35DkPFlglyLIR7+lc9mmbwxT42+rgjv+SeT0Y/IFPwgPfC15U77E4/0D0VDRzwTjdXfrjjBcPJjfJx86QDOYcklOh38KTta/PS5hltPMrp58b74elj9W9/EV4440caz4xTLkH6f+y22t3jc7wz9wS0q9yoGjFs3b2dl72O7d2PEKEfbt1mJrcP9S5cuZuXO7uGbW2Ru4I/kEFsJ/8fP/0d8mWj42L9vP85jX2oHX3qJ0zt3ZoyWFWEgnrEOclzdzjLemLyObxIMxJRvPatYNnF6jrLCCWbrQD0vW42oK32f3MCvinrbug3rWe2o01hPWJxqb/72N+2f//k/tf/tjX/f/u0//BTjC46IgWPdqsvL99ZTfuEHZc/oKqKMXpc+SP9F8RdFMEfz6vfhY3/t97HvD+T3Fw6fyUMwHKOXRRyVy2c/y6BLwhMw9jOeJsFLApa8LkRZ5sN40J8DvnlasamsAZhM4LvXCD7FXjAgSIuKNWIJv1eSpElDNwMjhDn794Vux891IUxNa8PllqWayDCf9V4N0SOcXL5vA9KJ2R38I1y8eDMO0XK88NXrKFAIZPZaz+lwSiEIQg7wtJS75Ev8yuJcHYFlKAT5xj+XBopvUHaLBmmN4rs/KT8PhqX8NLBCmM/5PsQbcvWbCnaUI4lA2gzQEbTmqRBXkdLCGkMAyqzbjjSa5FhABHRYUsQti0JuDH7GD4HxIPzQn2+RuIBWSJvaADuDlUgMebl8nCA4MJyswBmpRpPvfvUFfJyoRGuhByeUGxX/CeBHd7EzI0Nr0nIIJuX3gauDzWsKynfhDwGhdUUdfgee8DvxO/+FN8nYLIIrGeQOhNQd727jUXFet5FlyqT39MVj+MzIaTl4bj/26fnm8cJrcU41j+HE3PWdo6KkAK6ZApADwQnq34GY20DcKuVycxWpdSi5egaX9xy0WRZhuX0nijD46D8hvmBAVqWgsK7yLM//Q2GJ4pW3Icjy+uiKk/0cR/wfmCFzq87KzPbwQfobKbHqFrkki/mq4Id/pTMPZCTO0NhMwexR6Z+8/LEG+J8tXuYnT/PPYYpKV3E2dJV24MiuhQyeShFb0S5fvpWjMo8c+bSdPnWRgZGntGAQVTliYDTBvvY4aIVusxkMQ2fo6ULYDI4d2GN4vM+MKtCpRxRt/J9oLFmjUh4NAl5FfoiLxYwiLtPzf0T/vFj+Un7BOLIhpxQRjaKknPxyl0dQuHlcDd84IJybvc1A60577bXn2l8zi75jzDmsZCn6DvBJN5B7Af6fSH/LHPpDX3lwJH9SBwVnAT6loyAWu359gBvFz3y4PGlsNgYmj/WzLuBjeYsILlHWgCh/ByY/kleeKiPqAnwTpK3yXSOzSoiReUq4cldE+JQ2s5j/BgAixJW3Icjy+uivOUlY//nfliWMUvJUapFMfO9GlMxkM6jWgW0KAHyNo8o831XIKtcBmGC48jYELYXve9CQ/jzAZsS35Ci5PCMeWXGC939mXR8ftupIS2OlnfoMHsqsPJJh8jTKcBmeNEs+jFQSwv/k/jf1P9CQ2skMPjSLgiyOIJeBKvx+h8mLlRgrJ2/Otz+8/UH71a9/306dPt/Wb9nZ9j5xkHrcwGqRMyzVvtz27N2TvvjGjWvQ1YHEnbaZJdQuu/bknE8Of0TfNtX+zQ++21588el27uIFttGeb+dYZXbtuqv4VrJNZycn3TzJaWQbczzxxcvnKCj9E318nGhTba4qk5DrWAmze+eutnXndk5xONduY0xR/utLYV6/JwwAJnAc6wzpPVajPb3/yfbEHk5zYKXBV17a01558QkcFaobhNrIK4xdbMOzNqRCp394mnevipkofzr9x/jfNiCsGP/I1lU0wndW3DbnjL1G8BmsIJeu3G4ff3wCIzNHaXI8+2m2PeTUMOSm/kxiLLe92T6gmUeqO4i4z8BvJIzkp8A3TpUmZR2VTE62xS7wHy2euEoQ71KDTB7Cf0TK1eWPMMRH+NV/85m8VlIv9o9ZZeJkELhG1tCfCjvyBjztM0TZsPzaL5v5OPx8MZ5wePGz330d4PNKGHEUXuZr+3erDitO/v1Pv4KPk4VTdSA96egL4IfSGy27V7VX87Khqxuoc0Wno0yBwaegwIvwvbyZQ16NFPr37wOvEbxw/evxX5XJOh+/lGRyBTjyob7BT4T4rByPfkFbs2QO7e97sp9xKSjdCSsN5tp77x1mO/QRZMM12raFZvIkE5Ws7qIuNNSrb2pEsS8pfgxUIHWdlDYMvECex2gaLYA4wCoDCncH+gkSHy/qmfxDf+rFtFUGESQi9dYnuoxr7vJUyjSk11jiSuYymrBFMLofq9OcwHQC1okZ8yG+2xbtf8zJq6+K1cCnbDW04A8P4/Uf/ETLWPwRkdFJ6B+uQ2/DBI1gYMVJnMO+jnPYfYGls1y3UdK6SY98pGzCCTzz8WF4fxT+G2o8BmkpuYr60reQxpFjx4+1t99+t51mdchTTz6FTH+6PbZrTwxExgV6JoFtBfqrun7tSupnF9ttdPyqA1mPC76On5FD8MTxk8fBc6Lt2vUYeT2FH6v9bfOWzZnkiB4OXPtQZdFGdwjwb3LyBv3EjdTVRlaPaDTxYIycaMREz2rGShsw3NvGjauB5Q7jxQ0bt7AKZmNkjOO/mVu32qcnjrffsuLlp3/3d+3Lr72SVUPFCRZm0FmgqfxP1zO6QtvhTX7tPDMePorc43Efy2KooIVYy6VN2DIfxoP+UuHDy531LOQ4ykvfl35bJrpJhjwqNr8LUilfKwrhVNyD11IY4+/jz0PKZYIeBb7CCM5MVLEYzRiGFAqgzl4DgESq5xJP1fyWkqvS2cQUZDIcaRiYKmhNLS2Ss3ACn3ej+I1L5cd4gtOe4QMilUbq3snbOBa63E6xBeMECtTVa3hvRtysZ//kfRtgBlb4OfCYPZUuOr04VEUr0Zu8A7JYl4XHnwpkANhwfAIWaiAP1VCqhAM24jJGG2P6pX5MXM+WDvGKMKkw9++qkAjK7R/O8Oq4yK0HWludgaoVJi5NhmbUS1aQkHms5ENGCokS+qU4deOG8DNYHINfqwKK/oUlJSd9BkCUQbIrpq2b+WGrzrdiOGGrzg4NJ4ZDA+gZYwsliNIpH1sW/i0qv6/D5Te6Ed4etf4H0qUtUJhO46FtqAxKj17+Xn+u+Jihw1ORvHD5NjOjp9s77x9tnxw/367RAajoT7AqwRUQ2btPx5RVPeSbJdu8u9xSx4fz+MBRme7bRFTs1rDku4wizLAAxzK5KqUG5nT3WNqlrXUbXzudZry7cqHoYNksk/QbijZ8ycBoKKOkI1b4z/LP4VndrTpvoOh9/+sv4ONkiCH9rTyu1CvhmTlJCD82JAey1C7dAHfgGv0z2x9JiLJc+682SZ7C9b+8aNsQW+5+n6NtWdoYSwiWPvpBcF/z8ZNn29mzFzF03qAzhdcZAK1i9klYqTvr1dU+0Nuuk9zCNRkUYTS5yyBI7+8Ohj2OTgeCLuGvJe3Al1fAKXeTk0dmDjNAGvhPXFH2iv48c2GmSTtTCbCeTJpSpd6gH/F99zjimsFmxcmEp+o8277/HQ0npEMhz9Y/swxscw4GpP0X4H8yTR3zGwBBsp47xikNP2mWxiOO3xLOXRmXevIYWf5P4+BthnpwH7rbYDQmy28pB3XiDJ0kNXIMNcEAeshE/K9uUgi8KOMCyWcuvw8Pll8pUcHCMO+ASsSUTEBhziSttCYJAgVDZbRyLVqMy0bD8528bX9epnJQWfLP8g+8SpxuQK9YYP4I8JfSP/JBOg355lQd+iWdn2o40RhbBbENiodl5M8VcfU/eOZnDP5CoE/EH4i5FP44bYxnzB7dRy9pIP0LvvRz4MjNIz6p35GfG+DbB+gbw77k4jVPv7jKCoaGoeNaO3H8DCdg3GAAhIK6dhsy9ka7cuEaMnBV27plEwN/nAjPcJrCupVtPytMnnrK/fgT7cihI+3d99/Ldhr9lqxesxJnfdOsZrkV48wq8tvCdsltO3bhK2o9q0euYGC9TNumzdEf9mXo9zFiyw7bcAao4m7e169db+cxvqzVqTe0nqWfn3PFwGp4EXm+hi07O4i/edP6Nnd7sk1eOt6+9/Xn29989zWcGG4lN3nGOjI37kN9CWc5+fdF6G+Fm94+VpoLzveIUO7p29EN5FCl9Qw4TbES7zRHtr7//jGMVx+xdelCnOVKsxyjTHWuYABadUyqFKH43ZUVebfP5l/FMXcSWV7bH3B8DofIwyN+rW9SJU2OGGgZFT+/AAqRijfNwRS+qSfk1WfytP0Fvis4CHPl7WqM3GuQ3xPoOsrugCWu7VPDbfJRQHCZs3qZxjNQFUCFLguflEbif3QWy8Q/213kACld1r/i3q324lOb2PpaW3XsHhwAszjNCNQPMGIFFR5/lH10pT8FgImCi5zj9yp/4ovbcA3U5U36gU/aP6FL6B/586/Ef5/V/0Oy0K/3v3KPcqbXsbyS/9YrfKxk11/gpWt38HN0noH2efr5czjvvI4c9oSZ1fFhkr4VY7UD8vAudITkoYu+JNQrq1sRlnVcfX/qE8MJEhuwFW6qWOHzLroSPNTl5rN/0Bv8rJvif2NQ97Y/QzMqHuJSj/Ke/ccME2Xl4BVYfM6qKPDWz4YrSsRRGhnXeEYqR9vmz9/A/31gLayCP9S/sWxzZCQmFqFiEA7PJbkwoA+mgbYCmfrisziH/cFX24ts1QlvKiHUkYhsCfuBGcIPEqTvl/CLbsvzXxlEaXMU6q4HIADV00KvYejwaOA33/w9xtqTyNi17Rvf/DZHDL/Qdu9+HNhMtkgr+wsqchZfUjM4mFU/2rptc4xKR44caUdZqXiXLTu7du7mhLTtGFOceGSsAf+7pXMHK010EOtEh8SN2wLyqIlJJsaZTL7BiWqujNTAsoHVI+r81rb9v3XhmMm6cbX+dfxvWZY1wNHfnVshs5KEuNb9bfyjfHL44/a973yz7XxsO0YidQXyMiN4wlvqDBqWFiFVxwhK3IX38eeB4ssE9UqpT/wOMmpIUbcAHofTvy7NcPx9/Pl/PfwYTiRkZ/4quIgNBQu+hbS/Xr3IC6GLQyrW2G8S8mNLXOb6V4dPedOQZQxQUgCUsix+JUTqA2xl3HQc1RhFP8UgqTwxshSTUQlBGNJsKuuCI6MkgA/+N20AG04qOyX+zxLPNhWmplnfRtG/yVGHF65O0qAv4VH+DAYTl4i51cRl+QxwaQB3PFZ3pcum7dQQMg6qaeylAGApd0aNjiozC0vh8y79qx8TaZQxswl+Y+VHAY4y7BfSWByx7yLS+BGKBMZIgbbiUj+Fo8cKu8wsnqApq+VfiXB2C4hWbXnPLFUkJEQZCsgZvMksrXoFAzqPWY0QjiAvpWMViFddFnzxIFUu7zWrbxmlBzEtKNsQ5AD9N7Q5T9U5yKk6rjhhqw7KqPUR57Agv8q6g7YuI+75jspPVl+o/s2RPJbyX7VFDBoYm9w+s8KtGwxkjetMo533LQYtHx45j4X8EMcMY0i7UY7H7uO13RUN83RMme2N1ohyTbmQt/ZB4FxGrFmPX9XXBrRVeXemYQ1LM6vDt4PQsIKRBP5X6bNjTT3LrzC5fE63UKtZKIuUT10QPhSt3h/K/+Rj+SFDtT+qY3aKrTrryzns13AOKwPRsWoAUAcXgmkMkhH1/VEB3lCguTIQiCIx8MbD4A95LaV/4RRg1VYpv9xVRpmCLxw5ynbr/Q4nDrn//uiJi5y6wcCLJfXTHDHaVjELgMFENlal938UY7kR3G2nlkm2dAVBTkfRYEZbkCbrnKlktoqFmMXLwI1CChq2gGpzqAvkZVXHACByXGRJDOhG+VUo5Ft9QbiKyjq3HRqp5G9/Jl/C4hzWdfD3ODJ1JVt1/upg++vvvMxxxEoVDGnWh/klbx7I6HPLP7AKfHEHufAOeRtWH8jT/D9D/uZ74qtWSRGxg660EQ3Id2knKn03MSjOWh60mMhF+KOULMpPsPBh/NBW498IPk9zzPaJVuS6XwbCqfCJdeo1KSgPkRb+8WphHsZ/Ah4+d/43RQ9LMwN38fE5xhDwtJ6FYztV/kE+6r9wERT/wZHvA01TtuQLMP8njoAfDr+QWKC/TsZtf6ZR+bc/cVULK5PxccKpOjJwysmNx35ZfoNjFJDGy1xL+3+5dSl8gVc4X8zGMnBfjv+s/5QNwNIGQQ+tpFuGRuGNGQxqk5M327sffsTKsGkU4Mdox2vjePSKTlfpX2+ymvMSPjZuslKpceLNaso4zwqsiYmZtmfXevyT7Gm7trGnnFlB+5kT9M/v4uNAJ6+bt2AcwXjt/nNXWKjsbtuBIYRtTaJ/4cLFzBaq8Ma4QGE06rlXXuV3M3l6As8mZhdvsOz6DE7f59lKtAbfRsoJV6OtwGiychXyg20Xm7duZDn4nhx7fHPyUrtx/igDkK+1f/uT79WpOtCAViujZxtgtTWIAzLLyb8vQv/Uf6rAmrA1AAO+sSqkgy3VStSh6xTt8/T56xwtfByn2Ucy2zt5yxVhGvSJiZxxm8MEZc/2AzKx/LY7ByD6LZtFXpJp+CI8EoAJgmYWUMj8+cxf+D/MTLCfI3+QG0FcOWkGfljg/3H+60ZCJ7pQtlJGeSv9qx0sD9mOQ0P1RLT0m3mW/yxA8WVWyVAfc5QzbVTcBG1bGuCHvyP/LHFdJf95J5rf5fG0oZQRmpleMA523arDccRORHzjqy+aq80huNK580YZoAWtg+c08IAefswY/jJOhz/gKMwBfvES4ebFr+FBJTdxEXP+LJz/TevDEOnPxX8Fgd+HwJdXxNVqNK6ljzNSXnyvj9yQHTOzDLCnbmFYvYGfskvto8MnOS3nMtXPNrh1W0hc/Wx4yJEyf9FVB5raS0gFC58+CwjSS9Nh1R1wnMyI77DCJOSx3aykwkJ/6UsWISDyEHq6Yrj6eNLyrShtfbm6WDSsW17Nh3v0DXQLde0Ztpt4ybuuMFH/czyh/qvuKV2CiWrfIH+dHLIa46CUCNEISQ8k/gmDjwMidTOyUIbL8vOoqSSxAeCWl9XqbllxshPnsK+Xc1jSrcDAMc/kgAYoSyprp43IQzwvJ/8fxn/SK2MYGsxtxiOrWMGrj7CTJz5l5dAHnHRzAl1tZXvuwMH23HMvYLhmQoB6dLWHEyyOn5TlU1OTwXn7ji3t6qULHDf9Idst38Vv1RSG9KfaX+H8++Dzz2dS+PrkNG1tIj5RNrAyMYYlK4L+IobNgR80mkxz2qLjFJ3M2qeoc7sCXP1nlfo/hZdXNNy4QsYJ6I0YTTbg/0Q/d04+W9+WSVl0n+PVT5w82r7+2msYYYatvNSvjBBa+ABN1B3VI+oSOa/hPbjmx9i5xmNaJ6O4i56HyN6SkJ9U3Fj48Li0/6/gv2z4tGElI6jaGMOaw6PYdwr5/Eeu0IY4SbLo5cGED3wO6PwsoGKyPxN8FVAvhUKase+jzgkBkO8AD02IlEeRGUjV0/mdVuqXEktGVWgRkjwTscpRIAmQ1cxwAOkDcW1Q6vXufTXqJIOucxevtuMnzrLS5EqO3pvEoRyHDWJwUFFzL6SDedLivBFvCDyhINLY8GoSOMorZ+VzzNaAJ0HEq98RH0dCire5WR5jiOMI6SG03vlC1F6AxeVPuYXFZ5fP32dlg8vXXOEgq2WJH5bR9Xqss9zV64/oD7KhvwLLtuyAIA0LDJzBScaky7J1B4Ep7zj9RS4YNg48IFWl0WLbncPWMXCUNytO8HHy2oH2na9zqs6OTXQa0ACFo07VQTGVEvCGCpzXQJnQKfQZ6LpQ/0Yq+AODiUQlHMjZqZohnoQC5jj/yZcmiqVf+lBu/9mZTeOE7Dyd91vvnWiHjjLrceI0hhQ6MPZ9u//bVSgqnKbRMdkcQtYZDo1Yq9zXjFCdYT+9+1knVNg4fcmZ1AkcE2uccelz6oUSphjBBGzsOOGT8JP4go+KQSlcGOr4Zzktai+/Aj+rQnr5/cZV5S9OS4qx8t93xUmOI/ZUHVecGJs/4EtPOTSXdDdfmZzLDtGZ1/CuccUkHU099mTGzUU2xhKLpfSvmQ7CO//7mJxH0AFL7RE+y5rda6wGO87S/mP4Mjl3/jIDHRwWMnCX/hMrnSG2gxaMihJwaeNRlCmbRk6Px7sHL/ZVP6somCdJrcZ/RGYn4MI5mFmlzBlJ85AsXaGyzOE/PmSwPPCfZLAI9ZpUYj0kjiowKr9LgS1j9mGT+SoNb7Y9jLKrdA775WcxnLyC4aQGK87W9MzN+ZH4n3heQw0u0F8e7/JXnkqdk+s4/XuZSF8lqe+RQ4RZF7nIyvGCq9t0HHmXk4x0AKohJdkCNXVO5NBGLPzATYOW+QQ+dO7wrfXSK6Sz8Ac+AOcR/CFX88zKBpEZyjQgbEguonBp4jEnAAY+oZQxrYJ35Z5KbuQfoQFsGo0ltG1XKyH6SrmFZtK/5GSFyWu2SuVjVqUU0EeGn4idpmJbhAGGnA8eKGgqfmsx1u/DGCD9RDO0sa4gRJqhZeK1pGdyNVquUfYGpK5NzSU/SJeE1bf+KOWrLVYUgZhP79OTQyoaioiT7E7iyCrq9C4GzmkG5qfOcFQoKxvexfHoPdrWvqeegW/W5qQrT1a6y+TEdbbFTN5kJhRaO3Z0bZh9xjzG3V27NrQvvfRU28sybff632Kv3olPL7YPMQBcvX636bxUeXwXPnSg7Uzj9u2bWKHDKpTJa+3SxfMwKoZRjhW+e4dztRi4zNyslZhbN2/iyOGdDGRWYVyZzMk+9zgRbSUGcdsnTIG8wLi6Yobtmqw02bmVbZlr2Wp5s109f65NsH3ozvUz7R9+/K32D3/7HbYF1b78kh/QZKiBqjKotwz/Laa/VSGVpX6+pBZSP4SH/kOePg8CJ3Q3awfyZSTW2AH3wK+3WWFymQmhj49fbO8f+jT+n85QJ7Pw2TpOCnMw6OAu8ggetv9LtZKfNV6TK5CQ+owvD9sCPKlheAS/9z8Dbxbm4GIhwkViy8eh/LB2yrW8/JHrkxDZ6LYN2px1AK8GH8q1Ennvsvr4oXJrBllLnQygSWr7lLWLV4uPKQjh4CGu5BQpTLy0ZYJN27Ht8OVxoXb6i3+MOJZG+ek3Ytxz8MyKk5yqM6w4sR1BVspADugFJTOCKIqKG+oAAEAASURBVEksI+mrmMEpj6Fj0KjvyV3k+PoZ9T+gIRFEZ3T1x8KSjwP9zavgG2PI31R5rEyGL4n3aPBJETxH4EeoCL/oLQyerFMfpT9ArRaBz8BjZ3Hufghn0UeOn82hDDduzrWbDEwbq3vjIJo2qQxOfafgvJM+9ZF8wMNy8mwZvKR2LhkvxCAEuQO5uYiIPpNroP9C/ZuHdUZEPMYKs69o8blWJ4AbXbS8altxEsEVDm6zdkuOOaxh8lJjgAN7t5WVrgwOxBPV6u1ATd6mDQrfPADAE22NQbhpLJ4hXlVUf4k9oJhQiWmmhEvj8L03+9lM6rni5GZ7nhUnP+Y44peHrTpYrOk80GORlRrsbQ5VnTxIKPIV2nLwi6YC4S/lkfed4LpOHV6KDuz29NsaITREg/DGTVvansf20IbVvzzVjLxptK40ucOpSLf5k15uzZlh5eHPf/bz9s677+TI5sf3Pt6+8fVvtJdeejkTwlOciiPc9awc2bJ5W3S526wklgquNHablNccOvmNqWuZiNi4eQunp9U4RAOxlzLA+GYmDtNsCXJScwOrCzey3VNd8TZ16thK3cBt925xUq88eeJY+zYn6mwirv7zzKvo38kCv4BQjieHUOImfXNJt0e8FiVZ9PJgBg98NiBA//8Dnwnbrvr1AqYUw8tYEYfg8a9GqveHfBzLMspLNatRKlOXqOkRvY9D8NmLGnwIiAp+yMdKnLTj8Kvp0uDgmBouCEIY5JNW2TnGfL3G4BMvM7ZjwQvIWRpz5LKxyaLcI8xYNVCDT74pJGl8Wv5VZpy9U8jaQKdxFnmd5b0nOfLw+KdnmYFhZgqDyQpmwFTEJjiJwxMtsszTWWOExwQzT/exzMZxE1nbyBz4pziBrnV3VFKRS3lTfotsmS2iMzc81+CMgDAz0cHRDjZxbMTmKb244j9FZUVBangs1nUagNZZt4hkRgFrqII6ygWz6AolramuoJB+WkozmOwVTWh3Wmq/YrB5K3Ad2MbgQrBI9RngsaT5YrISzNaK+0ktG2VBEU7ZWHEyxx7g77BV5zusOMlWHeI5gHVvcm3VMQ8ySi9gWuqffMTay/vQ3PtLwpMkPwUqPhio5J7OyM42lHJlJkROZoTbyXAZQ3gaQ26i7F/ACdnRU5ez3/7Nt4+2WzPFv2tZ9r0C4Wn1zdqBQyN5QC/cqVP4iuKQIb44PBIPw4kd4Fr84qxbvxm6U6PWG/Gkk/XtSMdBe+qZd2W42LtNaKXMmgvaA0PqSushqD8lm7AQ5Rpvf5bM8i7X/ubuseJk94acAlBbdUxMltIfnARtumQhRmZVL/yKpJHFCayAUd99hgbycb4Ujc3Hd2dPipJwsQo4ObgM3qNAdbopYA11zq9oWLL09FsYN3H8yt7mM2ev4qT5NAOmKZRV8ITYK9iXO0FHdv++swaiT7uxzPzZ/lSePXXKbSPOIDhI9oudmx2hTp71u0DEKLrWrW1MuZFZRrAUI3FVociSziqsERNelKgXq9T8JUgUNvBU7+iDKfGTOrZzJYVLWKNAxXDCccRu1fnuq8zCMIhUIbcyzI64oT+/ARogBUuYhlklOtMr+hPW62iAVa+mr8gO0kOmJBZK5W1+whtvf9axoZbDdqtc1GGnTl9nuN9DxkhnmBmMVf5SyuRZ/FeKX/gx+YtD5SnkfhkiGsIr+MTjEuWF8vhsqDVVVPHZ8oUthjCNDypwdgPWRjcIl8JLnRJfpUmjSV3SmH/WkYYTceBuXWlMdZaw5JzAxIl6JK0w5DfrsnzAdJwAQHjKD914zHsVJgUiwHu/pC3PICzXKUdtJzqQ1jnsvj074D8+RuAao7KLmKhKA/filMpxAcZ4aP8Wee9LGKPwqP5nqDvqGZFFUeVkyyFg+wXTmBBkQVga0DkShjLJx3v4GLp24yZL7SfjrPXnv/w9K/VmoN86VoM8Fqejq9duYOUnRhP6YA1vrpSxvjay0mPvnl1MYLBVFj8m27dvZGbx5bb/yf3Qeb6du3CV7SWXcwR844jcVa42IVy/Oju27mhbODlnBe3mFv5Pbk5zWoKKOgSac+UWg5/VLA+/exuDC8e970Zx38TRw9evX2Wr34XIdY8dd0VqSIzMUH6vYrCzHaPJjl1byfd2O3HsKDBm2m4MNDs3zLfvfvPl9k0cgroq5j59dAY/0Mi6kFDF/9wlmUxA/iGidZbLeF5+G27EG7U/21xqwHTmB+VVaoivTmPdRL4Q7qqYeQY//Ifm02wrPY1z++Ptg8Nn2ulz19tNjmZX4V/FqW1uSZR3RcetptatVWmA/WGcKdoWeJf/7zvDity0nlHOiFiXxQhPCF/+sxFy8crVy1/y3CLKwF2XsM4Tjx/5LA1YGgFXMw4smH5DOa7eEkM3+o0raDOAoRzmYBlcFUlrDj1K204tmhWh4iSiVQ+R6WkzVY7qL/guMt6IN6I/QeZvLrkLS7qIJ5d+KSbQb158YlP7x7//SvvW6y8mH3FGHQMv+jie7f+lcYeRjtsM4Jlkb3YCIN+cIMZzwg02Wk9YL4QYxrf88Ez4Z8r/JCByYBA5CQ00Qy+/Dbeh/L2WK7h/JGaHSYJUN7SPvggSkclG4DkloK8wH08sU/9U/9EXnBNI9oEewnD1Oro4W8iO4W/nxKmzbJdzK4XbZzl21gkRjKyi7Vb5oN2LIQxxDe7UKrT0OfhZGB6q/aV0/Eh/MzJcnrBMYGd5vHw1b9PZNhJYbSww+FjGbWHIx8nK0ExazjiBZlshrYZcHU2vQwdXkAcECTwUQNqIQ+Q5aZW76hbk6hs8RfrU0xAngKSiAOsW1PiRHqBCHqQNv1kH0gX81TkJV5NQiqzKipPJ9sIBVpx8//UynKTRI0+sD1Czi7HRhP9Fk3Tyr21WHDv8GHdERZlgJyQZ8cU1i6Hr2IlTHA//EeOqM4TPtZ27drNK5Mn4IXHVjXzgShTHLfbB69CtXZ3rVskrV67kSOlNGE1cmX3y02M4/f8EY8rNthejyTNPP9OefuZpZPcm+o9JdgzMtB3bd0RerSS+JdWfjPR3JZrjn2nk9gx7RJ0k3UB9bObI4HUYyG8z2eyKsUyUIFfUUTW63CS+hhbx2sQpP6sp9z1krnKwT6ooD++wlefC+TPtww/ebz/6/vfbwS8929azujHVZB2n3qSgtV+6KQ9c45Xos9dC3Y5/9Uu9D6FLPxrBi/AH9P/AlaPkrPFrPBOfvf7y4NN2kRYgFnTHcS6M8/uQ4LEYn/24KL0vXiNq1df8LopY0fx9SPBChD/ytCg9L7b7CAhxqJdR+cNPyW8QFQQYJYMN68+0/lGAJDeMfzZehYxCOiLGnhXGj3AkjxUMopJGvYIG7fI/l1ep3Nno79IwptmWc/zU9fYJS35Pn72E4sQxVxpMUMTsDJ1RcZ9dYSYOXsARduCDTJAi2HtuiqXCNe9El1lT/sRIYVKGjAb5VMvxBuENXDuUKKd8Cy3AowSr75W+wJmvp3zcTSPPEmXSr8HJ6AaETU73oZwxtpj9CL4UG1CqjHzjj1DLzW0R/ZNu8U9iJ5NKluQ8IqJKOFAvrjiJmM5qHRVrtlLcn+JUnedxEPt827NjIymGQQ3COtt7gC8e6ZwGXAjIlS8dTUIEv/Tyc7oVi2JPYkdiAPvuM41pzzLk6/fKjt/AxfErKxqu0HkfRyn/8NCZ9tHHOClkr/scZdBZqLOQXtZDVuFA1xi+SO/AS6UujgMR0m6Z0sqtgPXkJTtPB1WkKPqbkQhwiWKu4Gb5QTnPPIwKWgS3WMmjXuv7kI838xpew3YP8t9C/c/hrPdJnMP+009fbz/Ax4nYKaJCGftGyhkw/Iy4eghYOFVHaJWnikZ1DgQRFr2a+Bk7WCjyzqwlPbPKvr5dbEv3cYI7dx/HqPAtaiZKhZ0MswR0+LfpgK9ev9OOHb/AdqlPmRG+Cb+jvDNYsoPUGOIAXmXI1TwZPAienKxj+d8tOTEsOoACx9QJnakOklW8Y7Qivryf0yRIHRlgWUcXL2N0Nm7e+314Hae/S3vTmpLUZyNbR9Qxr74ZP3yRwSfe7udrxckPNJxswSjkDBH/pL/0NdFCTua2cBk+Dh/SEp1Q5ZVPkgnukof85lXKELkTZrjB3HKPLuW78Q2jDmepG7fhODM4Q93cZRbJmbUMCohlkW2/aRdj8Ek+XAv8FyTMmCv5A0ieK/njc32r7ynBCL/+EL4kcdKRXmNjGYXJk/RUP/yBHwsVKL7H6EWGtk0HbPKPl99iqOOe2aJojoVPIhA3tOQluIpbHob7cHsY/fnMVYnyyw+gEuKXZEWA5QjS4Gt/ZTgihJUXHDHLSrUHT9WpxJUz5edVG2AuA736+xL4CQfhpJVH/AN+TnsjmX2RW1oitogUhZlE1q9oxliCIrkC3o3iiW+hWyAwPX2bAfsUzlin22VOujqKD433PjyZAburGnbvZtsNSu5NVuJduIbDPXhIl4Qs5MDwsYmjJHe3fY8/zpLs99pbb/2hPb5vXztw8Hl0+rXt4iV8oOATxcH/Sgf+EMcB2ATKunvSdz32OJJjgiMjL+NAcIqygS8EUa6VcSzcxUBmXdu5gyMtWdJ98+ZNfKCcZ3YYYwpGG6WYxi/5Wj7xGPLt23a0TVu2oCTPcLoD23MwtKxePdteOrC/vfrCvvb8U7vagad2sO2HQoCPkxae1uR9xOcSTdr5y89C/cvF0r/oWvxPNkbsF8/ytuVQbpKAPsiVpbY1Xmy8lFv+U+DeYALoEqeOvHfoNCt9Psoqk+lpZ+3Zhgi1hZ8kZpSLuwNKrvg2GHBJgBiLi1HyKJ/Ue32vX3Pq/B8+StwhkR9NUyXNb8ARLh/5KT0fSCkb40BagOpt3GNCQZdzpZq+TGIw0dBOFA0lIsbjAnzTkFdCRTqXd8MSPWkrfPhNmZeXP50AaZ/WA5nURAZQxcGBF6cGHty/aXQcsf2fXZOrT8WSRkJd0a+BV/QT661f4Jj6DxUsj7yw8N2nvArX555u7G44KfPd4KorQn1IIlMtlF/wlY/0HeATsJT/KhUxkseQVc+Jj8kxwGxjlDPEsR5NQ26s5lC/i/5tWFZYYmBFb7+Pse4G23LOnJ9qhzm18uPDx6OLs0aD+Ao+pU3xpbqBSNgPQcncOzai5lXlqdA/Vn6y4BpLGZovLn83qKeNEVVKoZGniElKmIZE4zlx4ABe3c93V0Q50bYaOSWv9EShZxFVBLjEQUJSJm5L6Z8o49GMvSj9EAPaSBdzU4atlObGtaGR91xWzNi2cA47e4MTyThV54evtZcO7kM9pt4kmCvoi/2SKRgByz8NX/RHQ3mNKRRXm/rs5KGQ1Z8ch509d7H9x//nn9snR48xRrmDb6qn2te//jW25RykH1uLHqjsElCNb2aYZHQloP7RLpxl+/XFS+h6OAFGButPZP2GtfgO2Y0RfTsrcbcyCbkhOp1GbPHbiBzfyPHCORER2jtmkv7mb/+vbL/GCT6z9DmbSb+N1S72H7U6CAOLBUO2qLfMsq1qki1CrjTRaexWVqasRFfU2GWfoDGGYlJa6ALdrlzBX9Q7f2i//MXP2t///U/a3//0xzlVJ7VqvtA1bSeUKn4OqY2w5JKWywQvifXw10XpffEaZVhf87soYkXz9yHBCxH+yNOi9L54fUH48FPY/bOxM4rE5iaxRzALhc/8XUCaiiLlQ9MuRHwwv39h+L3IHZA4hQrKQvEYu+obgePl95V/ParKXRJyj5dymZI2X/QyJrmkkdMgFRg0IAWtMFX4b7Ga4MrVm5yGcikz1y5hdQ92Tt9gAB8Di5oi8bOv3tGGedrxEVazriJfV+pIlLi8dfiZVR7Cxz4mp5DYuJWgPhMolAjqsfKLeOLzMc6iaLRxaoml08HgbQaDWvG7IhGHRmakwcheu18DLD+ZXyFiGevRaPWNj2PwDVuIPDz2xLwuLn8JBfdRxjJNvDpVB8VI3zCcPvAdjyPGeLJ7F56tracIOQe8DpW9FKbS37QFeQQ/3/3xA3iOlcPQeuVbGII6YpCXKyci8DwQHJA8k78dM3eNINMI7jPso33nA/Z/HzrFzNwUS8IZsrLyaBX7UcVTXrYDLWOJSe2eVLQMW8FMJEc+c6yaVujVKNvr6Qg0YDkwV+Ht5fki9M/gGYhVWn4HUljOavUyO1jBJymuH7wGWhF9VP+eqlOGE07V+To+TkRM4kh/Ijp48OpGpuSRAPKYYFrC+JQ7dUX4fWmTbr+s69EXiVJOZ8kPOhvdzAVBQv7A1VkR6szmmpNm5lejgNxjhkAnhhwFzjapT/E7MIk/Axid/g1DG+3aY21VDWzn8n/Qd7BC2S2//N9XYll/1pHbqKw/t1SIrTMl6TSDTg2unAnxW6+vEENcRXfAv1745b3TuT5Lj6K/hpPkw1KnzGwRkVzhde7OCkNqZ4XXqHiB6wqMRxPzrDj5yoH2/W+/wnF8lDWGk2oZA/BCQqAdH+9cI/ihKZCH8Hwj/znaQ/kAAOjAQtIs6QiSgOJmRSScMIO95C3/Td9COWS7g7NpWWFiIQY4ZZyo/cSPwn9J1wEAo1AQPk/cHoX+Lm02rvWrDImS3enPW7XXyi97zomn/PRf2iQFNU4MaNS7l4Px1FshFFoIxDD/D5j6sOhR+ozDH6e/OPbk0jb59Lv5cBU46yBASIICSHj2wqNorvU44u4c1gRkNERN9kL3ysoNnustQYt/xuEql/xK5EgzaTck9K7RU+XVGTuXVGsIiC8MZ9CZeTONiv4c326w/eUTfBKoNN9VgUYJxybSTp2/2j48fBK+ucds4662H6OIW108TvQmxmqV0NXzd9uze7e1r3G872qMGu5hv3T1EkYKT7PjBAQGV1PTKPxMbOgc0pOz3EKzDscvd1B0V2E42bf/Cep0ok2xymUKY8g9VlWoQKt4xwcPyrMrLZyl3O22HyYZhOORmJEPkMHZTwhIE1gBr0+2LSjkT+x7Msu0r+NX6QqzofeQMSuYtfVYzy+9uK+9hP+Vg09ub08/sZ3466gXfWWxUhVYrmYbvyRt6n9ofw9+gxiL+J/6oSKqr5RWGk6ccXcwSia0VctmvjqsnmEb0zVmYN8/fLm9+fv32scMXPQXgVTkuwNP5STCd4Bv3Q/VvXC3HfDB+q9v/hqzX34wAu9jn+qVAHhIubyI/4ekiS4B7Bz8TyTDbIPO4oc2KT9yhHC3KToo01jiShOfY8AjnXRxdU8N0M3OvMS95O8i+IVcoTs8F/xexqEoIrCI/iI+FFJ8/Zf0wiJ86Gcz847h5EVO1XnjJ19uX8fHiW0HcYNh3ITqj/INtI91mbQSuF8Dsrb5AgctwKNi+CvQfvFufIPGPtWrAbZh0o4nGZImuuFD/Q/BudU3Pi4pvyUdz8p4wi3VmK8QpOhC+Xh29XBWCxpNXjI/Y0i4fKe/ow+/dXsWw+nN9sGhE/DpGbZ0XE/f4qqH+Tn+QNKVk/VPOAK2zyS9+IuXeAzh3kKx5erf6FzBxIhD+StNPg3fCFlS/uRqwuFSf+jlgdKRg8qNGXx5uKLIo9PV/9TLLbOypbYB/+nwq7JHGJARePp/rPxlNIG/KL+Vo+FE+dw163mOXy557cp0DCfP7Gx/9zds1TnwOHTG8Gc6ZLbG7YG8sJIU4iI/ZWIZ141Hya2bLK9GBmE4TD+BbJyeuokOfbT9/H/8LFuVnnnmmRwP/Pi+x+PfRSoow5RbruxQR3ZV/B1Opzl69CirvI+1T09+yvbCKxit17WXX3m1/fCHf8NKwq3R1zw6/hbHo7sFXvpvQD+PnIUW+juy/DaxrG6iLt1yM8lRwuoHGlfWx5jFWA/46gT2Efb/yiD7g1vEV6d0Rdt64CtzJHatloYreZ+9x4RRDGTI26uMG976ffvvP/tv7X9/4432T//071ihsrFXkdSzqiy2/wmnjnxY7rKNPMB/y0V8MEwYla2tMS3hwUiGLER88PtfIPxltuoMeAdZn5en5meVs2pkyGfZ5ENqbmlkQ9RFtz8jfNitAI/giyQv3EYCbwn8IQbxqHy+leVdVhiyisS2BDTmtGGFKl8doJo1ivEsDd4Odg6FyiizLAW8fP1mO3L4As7RDuPJ/xqhOoDbRCdfq1CypC2CwI6CjPjz2FXTVy48BAtgCW/RxXsKJMNW6vpsPsavJMXOKD80jsRU4bd3NYpRK3agG2hewnevnw3buDNYtT0q0aVoFDI+TNZjge0nOuRo0yEzraN5XAJfeAOQAux7Auu3UFmAn89G6deAa39NKhu8IleBjZBSoZlTWQOQhpP5wXDyHVac7N7JihMFHOVRmVC8Z7gCDWvwbV6L4RvyANghwJjS39SpK/Iu4V5LJlU4HWy7tQPqB5prgKXhhQu3cVz4cY66O87e+clpsFnJ0WTrtqJ0s3+afxpDCtcaZAUKMFTEtUL7rmMrLfAeY+ZeS4UuwRmo2bE402IZco0XxgrKh/o6/slnO4eUb0iam+Uej+hLGkPFXKDTENHyh+eGTIg+h4Owp1hx8sZPOFWH44hj4AizSKdOa9KRhaaEEfGlc/YE8yHtQERKmRFTr/BsnmiHWgcIcPl+Bl6yBEGJyU9W6cAH+iPRuDJ18x7b53D8yokbp5itvsSA5a7GTQZM6+j4dHass0OPhpPlJnC2a1u3s7Pa9QdhJ+7A3sGVyoze21fREWp8TEEEDvppe0voH0NKeFlaVvsLrinP8CNZF13EWNL+u8FJA53PkqrAwvOD4UQ66QgzeLCdjaEPPk44jjg+TtiCxLZAZ1vFpLBZBHRUJ35b4H/qh2iBJU7CRqHRQLiKTl/dvT7yCTkZY4n1bVQ0jmxHUf8ayjiDwcWBs/5LpvGXcId6TBsdIiiX8o8BgoqU8mcEvz8Rdyn/hVygEoVaZIPUCDUDEtL5v2iXYBLVXdhelkP+H7UUwpWZUkJF1xURs/w5ayQ8/8U8Ju78ufqoFGL4iXezHyn65sG/yi/gRvCHt3wfh/9Fyi8Pp1nxQLUUr7DCJ6fq4ONEmR4E+UkVdNpxl58Y949dAybcOv+NfSSBefBRQJVp7qUgGwztqnT8EkNektZcmXkDoHS9gZ+w//Krt9v/+99/Gz8i+588wGzdJvwQXcAh7DVWk21qG1Amd+/aCX732kX8gzi7OLdSB6w4vWULzIGn96EE0ydTF88dxOHf+vvtN7/9sL2PbxSdyN6DN90+u1bHiijbzi7epS/cuGUjJ+LswtfJ6nYVB7M3nKlk205mHcnLlTPOaLqiyNWYezHcqBTfGJaEO0O8loGOS7A1tG7AR4qK+foN6zBU7cbZ38Z2B/8rl/CDdodVVvrVmGMF5bo18+3gM4+15/Ztay9wP/D0TlaJccqa/JU6Gugq4Uf0DxXD2H+c/+U6faww6ObP/p0mHJl2h36GIXmOS1f2aSS+Non/p2Mo8e8eau9yROeZs1cw/mMIwmH2KuSfM6pVxVT2gEYq0vr3vX4WfTJ4rFUletocScYjimlNauUpYAJikD+Ly0+OtNXIvHwgK+kVJlQQ+V+GnGhrs2VawwmTENA9kxB8t58QrxhLSats+bz9n3ByfY7yR3YYn//i7D8vjbEr0G9eGFacfJ3jiLWXuEooK8AiE9WJHNiRQHzTyHk2PyIrZZThSp88U36fbdNC8c+ki64hIHiQtiSV9yH+svTna4FKVmn/Zh4I43AKovAN7fATtbIA3QEB5QiatG3MiAnlZ5b+twfMa+SmL3OV3xT668efXGhvvf0h2+TP4RQaKDmAQT51BbJl5488xE+Y0kKaZRWLIWljfMjFu5EJL/r14EozJM3XYGeGRveyDL7XT379lKIhg+UtJxyUJ17ybR90O5jWaKLssPzKJgfnrqowS40pflPvsLpHVHwE+L0svf4DvP+I4HgevGgskL5FJwwC1ofjFyaiLCtHWuTbSlaczN8vw8mPMZy89NzeiiudWV3dt+8mO3IMN/FSdLXVye/IJL9RKFdmXMWXiaeluXrwHn2VOoI0O3PubHviif3tyf37ORJ+fU6zUV92+01oot439NOnT59pv/vd79oH77+fY4Z37NjRvozz16989Wv4jsLogk6tTncV+e74x3QaVpTpThjfi7NvV3hrdKO8wFGO37rllnnkOqtHdmzDBxVfM4lim6VulSnKEXV3V6WoK2xhdaFOYy2zer31ljEYMJ1sm2Y14yRHGbvF+i7HJX967Gj7xf/8OccR/237B1aduP3TKzJRngWh2LnJZzBL5XtF4mOI3Sm+8Mkny7L8l+GjkbyWjTSkFj7fl4/ylw0fGaPGMaA+lKdKXL+Lg4a3Htjv4wnyXB/6534fRVsUMPYy9tjjLg4a3npgv/fIo3t96J/7vX9WCYRTqDW+dCHux0URYUu+jVEnEaKcExhhk/RWPs3YfCQjN30OuJ9X5cuGqPAYPqXDYQImvkyOHjvPMsBjLAGb4ljCazRsFBGMDR4xPKEihHLtCSrmWQoAgoeMssSQTr62BZmzhfHuH7/Ct2Hw7JcFIVdiJgadZcrvMs7OyJbbcozKT2O2X3WfdC6ebdSelqNil46OD1pJ1zHD5UxgZkmhkcqkDV/8Ro6mpNcy9O/wpfA4fMtCgjTwKAoWlUBxWnotLn8p2UjeKI/Gd8mlezb6qTrffu15/Jx4HLFnqZfyFB8xCvmBDkXJlKDKOsAfSL4IhcXwO/0trzXhqghukNFZ3PvQU+t7jCiE30a5/u1vTuKI7Hj7+OMj7MefZqvTprZ+8x767k10AmzfYSmgpymwvoHyWy/1J11qNYPbQBikQ/O1COy1eN7WgWGUWXAoBc/KFg8YQaJ7iSKPVbRHob+d2pC2csjvQvlrIPio/Cf8+2zV0XCSrTo4h9WfUClC4KnWB3bSSqjh4w5Xnh8MJ6ljegQHVzKI8E1n6mRh0WFH81Xv6GXIWMCMDaN+VDtcyXD50q12DIPJ4U+Ot3PnrqB4tLZx81YUf47KNF/xJvEsf8JauZoACOngTf5xK4mDqVpCintnlvevZyAUh705ihReQFgoiuPkjww7/9e9cLeC7OA+H/93/gMx6QE+E9AmNZPKlhj+BxLIij+vLClFqbI8nKrjUYGvYTj5Pqfq7NxK28H3S+RA0kuwxddC/YeUybNi+Ssw7tBYwiXcdy8B8+igMl8cyBCmclARxVXnx/P4/LmDrwiOhWUQpp8IBCb4mylR5QXyNBczSJ2oXOT7YvipoCXwkwkAQw+iD9gRnNoI/RNoxS/h/9QNzCV848RXDHFi+KpEEQN3WZXnMnFnQkMPctfI5qyXSpNtdRx+5Wv55TdLVrRTxi69xsuvqm/5qwz+/mnlty+TVvIzCIAnpEXGrF334FYdI4mdSYSv5B9t1eFZFAaEFr8M4WUEgf40Vuu/apLc4AfLr2xb7RZFnmd4Vll1AG/krOCkrZ3mqN9fvfle+83vD7XJWzhnXb8dubcKhfVm+GslR8RuRJH8ymsvMUs30/7w5q/bDLOL+1CmVzCoP3X6XLt0/izK9RPtH//hJ+35F56lj55CNv+elWZnaePsg6e/1tG4gxhPOEhdWNew4f6nnkQp3tAunLvAEZPMLGIktE49tUG8VZ7vsLd912M7WDL+TGYnL3HSznl8mmg0yQkX5CN/W2uutMOk3g4ewPiD0eQKW3Mu4Iga1ZmyY0iXz1khM3vnBif+rG17aKevvfJ0fJw8tmsbVcIkQIwrDpioPJtK+BcCg0+ugf718ln83+vCAZztK8MhnuXFYEvfswKfL5faH35/uL33wWH8CnzCwGOGrUXbkZ27EDCrkK3IRLesFUDwKFR8T0szb8Lq3UhL5R+xhvbnbeEZWlEmi2Mx5cPkkXIOQJYpv+1Kh6rj8OPTw4EfdaccWskgTqOVWemw0n/mLmzlS3gA3UJfEA7nlra/tAwTLwNfhP1kjp+n/CUbLG3RQP6z/E7ArORUnfGtOpAUvNXTiBwDgj2YhhOgBrjQh6u3PxOFmuLmRCDvhWTde/zhviB/FtN/lEg4Dyl/8k0+0nOh/ivrUAWeKPhdxvpNdOzPg734AkM8DZsLQSoWzZBw4lp+Lp2iT9+8j6Hkanv7nY9YhXYUh9DoV2uYpFq7kXp0AFs8JBT9C0aAy1XqlAP/WdMpVPhgef4z/Retf3WeTIqhtzm2iLNlCqTRQN3Pv8JWXtTIh95HZYun/OlgXEen+rQbp5W0+Pz8R4mSSVKPfsbr3zIrH6p+1MiQ16kT6Sg1NKA4acuKE7bqHHwOHydZccJWHeLWViqdtCpXwZu7sk5kw4eEQQbap/Q3q3kcf7MV8J0PMNa+h4y+1w688GJ74fkvgSswQWQnx8GvZVWged/F98s9jCwaNjR0aIBQV7uGAeLE8aP4QjnLlskL0bl3736sfflVTsw5eBB9nJNsYCa33EyyIsSJY8c+GkzWk7crkNXpNJTYP9Qq71WsHplkG821GJo3sXJwGz5NlMemt9/TWO72IFcGTuELa3JqOuXagOFrM6ftrCLf+F9RnlNmt9tbZlfb3rp9M8Z4uxh1y6kbV9t777zdvva1r7ZXX3mZlTXQOnSDCJERNgRpCW2htvWfigrE+lkcNLz1wH4fiz+eqn/u91G0RQFjL2OPPe7ioOGtB/Z7jzy614f+ud+XfB5ex76OPfa4i4OGN24LK056jH7vKZe5V5SHR3zwy9KQhffRU3/o92Xg9qCK8vCID35ZHJJ9bAOHKGrTxRrFttd7WR5t7PU97MZbXfBpDcASgdQM2OwLhgQ8DIMbOqhYhGFO8zHrqdvzKBRn8c59ijPgL9JAnS1yqS8Ngu+uQvFUFGcBdJTm8FgBbQ4yucqTCqTqqELLsWLhKPIiMbynQeRl+KlvZtMbC080uuqgkr0/Q/ltxOav4lD7i41dhhEH5DMsZ8upFTi0c7ioYyX39Lmqwa44nR6NWSu49JZmXg4O+M814DoGfxH9LS5RLKnRU/7cTWvqihBBSlmH0uVb6DEqf4BFuORUHVOqOJojM3TzcQ7LqTp9q47CDnx1DiveivwYwgJhwHkcPvh33BIFtBbDF6Wh8FDG4fjde3cQfsRajaNflV4SXp+c5Rjb6zibOt1+9otfIwjtWIiNorYCXyZa5u+xrFFnZDouXbPSI8fAz4Ekgy+t3naezjJ1+CpzHvm8hplMKabzYL/b2TpD6fcc+SitRFFSdf7nfYH+oxKOSvLZ9K+05DYU3cylnXC4p27qPbM4Y/DnZtiq43HEnALwA1ac1NaaIT34iYko2vlmIOyLF3nM6zeG8KirZC8dpG2l0lhgOn6o43u0szUOtiik4K0PSFkKBqTnNXvyP/gIPyYfHsoS/uzHtfNl9smO0cGTuucdBsHOwOrMdRWdk/wzh48Uj4W7w9JfDSauYJLuLu9eu66Wd9uxOjPiMXJZmQIiDrAsgf+l/wpGqLY/DTNeBOf6bPoPbdqYSeCPdOAXxcN8s+KEsGr/tk+oxfIex/EaelfTKWdgz4oTNjfEcOKKkx04h3XZP6oYmdE6YoQZsBIEj53/AnAJ/IQFqcJHfsiT8Qb+S/kJNNz6C748e1KRA0lPBSj5A42ZSXIUr6GzK08xrA3ZZylr2rPGCAL/CP8t5f+UBXyHEgZ9n8fpX0btfAodVWLEOQaQCLtKof8HlT7vzviFWPIFeFn/Hreef/CJVKlyF32i8Mrb4W+TAoM6IjVFGscuuSZtMMqngp8vf2L5ERtVIcD10keCvLoOB6V74xx2IDj4pc5oaAVaStVqxiTMT8dnwNWwHsTdMnotlN88oZflNH8NZTy60mICw4VOUq3bKVaLXMKB9nUMEtcxrF3gtKsPD51lhRgnKLDFUUq6wkkd220xT+zf286e+ZT96yfY/72+ffWrr7L6ZDdbZk+3X/zqt+3yxcs4f32qvfTii5xcsyvHRR47hk8jDCEOSOwdPCbSgcksbd0l3u4/f/bAs2nrZ8+eQ3mdjgK9UnlBvUo+HcNaxPUYtHfglHbb9q0YWM6zqnCSVVT0nuHTKr+jAo0ed1Dwn9i/jxnOdTEYXr98JTJrFf2Dcyvyuf5m7ty8Sp2w9Wf1TPveN77UfvS9r7V9e3eEbq6sk7ZpB9DCq/N36D+0v+FTbvXdeAAZLp/VP+Q7T8ubCewNyI2Styc/vRznu39464P2yZET8Dp9D8c5p97w2eZA1rWcrnywD1qZ427JfBn48pKSu0Pv91H7g5/T41Ec8/IKzr1NJIE/fOuJ8zDEVf6Er/gMz6qvjLaDmor6zaoSRiM6jVTuu+LTimRMBN3lx14OZXz1KfLrCPQIHiGfs/390fIjrNW/wlDiKw3486Y+sAL95oX9Hkf8evsGK04sqjJIlzQK+/CDMkqhL995538uInda2gLrQyhPFDISDuUZ1Y7pLCJfQ1Mz4b1+eqaL3xO3Ei1b//XdHJKRiQdMCEn/s5C7mFRV9joVGfiD+plha534r8LQG3lJOE2tfXrmBit7T8Kvn3CK5WXqj1MrMYx5+tashjHauQNbeaxQECP/Sv9GBePNf9R7GLi+PnL5waPz38P5v2Ck7JTFq6+gVAbbN1jXGajT9xhDnc/tgK5qVY+1v1H/Cw3gZY24XuP6/6PAL+jmwtNQ/2l/ya1+UmdWBFdt8bTuKSf8ViZWOVadlO9z+m90ff2wVYcVcn/3o9qq47Ye+dqjyO1/lNsUBR0YvkXxUk46MaGxI90s+WnA/j/+z/+r/fKXvyLeHNsWX2qvvfZ65OwURgilyd59e5G9yCto5RHA9r+bWMkBJduZU0yQHcHY+957hLElcv+TbIt8nOPlOe2GbTk6ZLV/d8J4y9btbIm51q5jDHEly0YMG265cSVdtrISz62jffxjmqnpm0mrDNm6ZTN1xHhJfyfMxq1EH924YZOk4jS1Gxx7jK8t9B1PWdsMXK/o+OTr6Tn6jpyHTzXCX79xPVy4hRWVrniZgDHv3Jxux4590r722pfx4fUYaaoNVltKDdIWqBdSTvStOlbbeFMN1MU/FeXhER/8sjRk4X301B/6fTHIRW8V5eERH/yyNGThffTUH/p9EcTFL4mC4PQ+dvWUPVgq8kwDD0H75wodo3H/0O9jWY49jr7mYfS2TAy/ef054MM66QcG+AOIgiZbjQX4nN7GOzFCB2IMAjNBURLHy1IdUSzjNEpBTd/COREzRCeYyTp16hxC+goDaBTQ1dsYHLtdgsafAYBCmsEwjT7bXpAIcbxGZ1e9AgKbc91VpoKpwSImXqRLXQXRYMb7+GUTURARa5nyO4OoEHd2xY63FBHSGJn8TetyYxW/OCWiAWv1dImbFk73+yrMXGEiVg68fNd41PNyAOgsgN9Tntx5NciyAP8B+vthxH+Vjt/QtVbP+L3SiWXyNl/hRKJWGRxsB5yzhNIP3w2NPcDf9lQdDSfZquM2GpXLWioXddvecfA9Ejyo77BEIA3l6GUa4Jc4Epd++Vx1JvM4IzeDsJrkeOEzFyZZzXCmvff+CbaDnKezUQErJe0+z1n+fc8OQv8kCGZXNHBsb/KDxjqZUvCKdwQqHaYW7k53m/ioY6XcCnmVAeskHZuVPaBatSx1O+7D84j+fILNFspPTF+Ib70t0J+gZS5jZdvBMvwnyHlO1dm/x1N12KoTHydmMtDf8ZJxoH8pLMATtChKWxtluMKBp8o5b5aNq3gqDw41qXdZA3UwCj1fbRSUy3Z3g205p89fY9B1itMyJplpvkJG7BPGz8EKlu5aBumWASKQak87tOdoVleD3Zm9nc4xSjigHNTbUdo+nAWylarIZLUB6btvk5ppl0coV6ENH1BPJNAXiP+qHVRpioTibbhfx+lPuMQa8pEGeSFInvZ442qTBpOWuKXswSt8X00n62CsZauOp+rg4+S7HEe8RYMbxl74sso/MIOwAqzuhYnP41dx1yL5o1wxrfzFpXKUfMSJJ8ukonQXXxQ6udMRprM8oRVt1C1PVGV4X+OUdNOgZm4dJfOv8g3w+VrfCtcR2rz+cfkzoMet6M+DOHdgYgz8vsJEvlJBc5bQNugA28vjnqW7ckaekF75R1hEPQVwBstsI6sAFooMKPtBpcc900vhS7/x8gfg8GMeQvu85Xd2PXCU6eJGW3Smbp0+TvayVScA/en8a/soZP3tW3U6+vVp9DaGonXla364k6fteiC2XVGhggpOFGk9hY+wowzQPzl6IkbKTSi0+il5j6OBj528DJ1cAcnsMfja1vahRG7bvo0++RorNy6C/472lb96FV8gazmp7Ga7QHs/dvIUJ9pcYXXBVk5deAo/UbPt5Mkz8N9d2nkt6S6nv+AjUuC7nlnB3Xsei7+SCxcuMVN5kbq3jstQLQ8qq6XSNnDc8/ielO8q++avANP2qPxWBklj03lClIOg3Xv34EBwU+LdYCZxnrIgcWj6yBx4SHLN3L2FYRM5M8H2utnJ9tcYTn74vdfa3t07rHGIpfHCPpu3gZ5F36DPj29yiBiO05/nPqC2nSJ846sJfnbGU79Os4SfPY9jewxL735wtB06xEpaZKiqwJo1GzFS62SbrUwYPeOA08khtizb3uuUn4K3LPxgZMQBLR9FKfgOPzIN5Us5E5Hw5S4Iu0j+GJf/Jo8cIlflmoEamJS/biXUaGJZndCSBJn5hybGDOZkELpB2LR/LBMJX4JD2h84LEt/ixC5N8o1eARCgA7AlpTfPNW1hG/ZhO+lvFmBfuNWnTd++loMJ0EKOOkDnYQDb/W91G9KY8oBfscl+o/BhGd2gTTkUfQXIC8SsKfj7j/xWnxV2HL071ErxgA/iYf8x/ivTkiifIFSYLPaBHCmDx7SSwxCa2Qqr9JH9Gcxopw6zSlYnFh56JOz7YTbx1g9Rm0zqbEu+vhNtl0YskbHzOTRVzlYzgkG+moXGsfsekMW+kONJ0ItOhT8vI5+hvLzXuQacB2Lumz5LdOo/qljhKn6d3f6mkkVMnQyzBVt3qM3kK8yRL62Dwp/g6wyxvwILVw/J3wLvVz9S19zHS9/taXiR8csmKvoC9AxnBjkPj+PERL5HsPJPbbqPLsDHyecqnOArTroUqlX9Cf1NNm0spceHXtowUSO22RusYLt6pXL7Wc/+xXbKK+0J598uh04cCBble7gvH81W+wex2jiig1XbkhH6aRMvHzpcvvoQ46nf/dtTss5hdyabT/60Y8wpmN4xnCyhglO++4ZYGT1H3J92lWvGF509K2RRSOVxvH06fIdaNq3S3edfbuKRQbZvnULBha2T1qP6lgD/ZUzGk8si4YQ68zTc7ZgONHwLk7qlOapUUzenmH1+Q0MN7fxZahPlq1sPV3D5JzHEd/DAe6Zs6cxnHyFbT7kQcOzTYZfARu93ReurDjJU68/714DciP+G4K49ZjGWnhbHFrfFn5HX/MweluIMMrVb15/mfDhZWrVVhAEg+myP8vH6KH9vjipDFPKeaprAQLRez0UnZZPP57b8jF6aL+PpwDGQ+CrdMfSZp3wbMcdp2Yik3oaKqtXGtEWQTBaCqDKzuWKE1gv/kh41YqrSHI24i5/127cbZ+e8xjZ0+yvu4iV82a7r+IcgcBgigEVc7ykszNCIJA+na9Cz398UEzI9qhVwDN38HbIIB75QziNFIchyLLRMEznADLV7JuAfF9S/nm1HMIyO27vxks/FlP/DCI4wzFZhitwVtPI3T/pCS1mqRBI40YQiVJmUSmDDbbDN54C4f+j7k2/9DqOO81bqMJWhcK+kAT3XdRCcZNsrXaLIiVb8um/bz7MOTNunXb3HNuybHk8ErVvpERSpCSS4AoCJEDsKABVQAG19fP84t73vbUABClK7blAvXfJzIjIyMzIzMjISFdLV+IHQfBHC+6zxORev74NL9/kQxtHhOapi8+r5ZzJFLEsoVHxOnjUNFVmcHRjM3eBo4jvwjnsfShOxolFh0iZleLErlAY8r8mN1fFLx0dfkEP8Hf8L3Jl4zpWaxdQXhw/c5G98oebF15+o3nz7ZOYiGIBwyB/Iz5u3NoU6xFgluNRaIaGzGURoCNYNLhlQn46KbNjN28bNugIjDpFB5p6bgTKo2qoPCf/8MFm7+SOIf2q+j/k/zC3QAmnAdY+kcGW//LM9mSM1fwf5j9txhjSJK8sA4qw3/5yqk6OI+4UJ8Qhfvhvf59/pIHBwSWjpUl+M8BLy+Ob9JBNWGanLYWWn+n447tBJtNs1aYI9zIBO4LTyLfefg/T3RMoTWawwnCvamt6qcUDK6zSrzMuJ7RO6p2M2VzmsWCao7OaQ9ngZEr+q2zZRCepY0YpMOt23NzSycZZWCt/MkEmRP6n+RFftmbgRkLpH16+mQvvZiQRxZAovlb97/jPnSAtRMJ2BnqZ3Ftv4FtkjhMiUi+QeAOduu2gwdRbHydu1flSfJxQXxioMDwrvOQq+KXDMg0tRc5a+EOv5WnGZIYVus2/Sob8+RnRlpUdJpqelCPfa7uTfn2AQrk6SMyAQrkzaOvmgAtYHf6UDeXR1T/xlvwnbmDJ8zYfkBVetnwMLD/xZ9Dwybce/xNa+VfhIX4HOV76t/GUFVcHo5Tlm6tF1n9pyry3xQ8lNteUuwPgUvBSfwfl31ECLtPTnkPT+5Y/MI0jnR8i/xaTReYqtbks57A6GnWrTnscMcFFC7+Q1ckkv1U/lAgVBViCK8Z2earwyAc/VYxEki/mV5gqkm1XWiCdu3AZC85jzYuYZr918G0mOutzxOSJs+foZ6k381htQoyLExvZguPKnatyDpaPvvcuTlunWFHc29x37/0MLpG9WzYC8wIn7vyBfesoInbdBM5RTs5hZRG/RrWFRseuthUGsdTdWACyeLAP5Yb7yE8cO5E975ajFi4uLug40vK3vPQ3sBP/JzrrO8c+/GPHj6b+Oeh2scR2VxMgnLmSVx3H3nLHrTk15zTOaT1WU4VJ+gcAa9qton3m4jms9XY0kxxFvDh7qnnsU1hSPvrxZg/HEzvYz7Zh+Cc9Vf9hbdr/1ep/CtyCqGLKU/3YdKtXKeX/Qfw+PfdC+eTSoflMfA4xkUPZbzuIYRhSwxV9K5JOuy2DJfIb30qpmwAF5cr+J2NI4qe+DCjxbdj+pAehQBwe/C+cyIRqk5l381H+D/sfQVCvBuM/4FEGxpFHHktrmWhFop+k8E3Y8L7anVmx/QEGOJKgJYCUrcNauOr/cvxmcIifFLarAX7epVs6A03IdQlbuIS2T75V/ktG+Nn+3BupCdYaaN3ixeY+LE6+8bWHsm2LakD9su8RnkIWesljFoZIK4wB6jxKk3mDT6FAvF7eh/zPe8qI+P437TL+k5rviTeQP4Lo819cfDNxiOjuwzcCEmLE4njxqmjzK3lKehTumo1yaQk6H3ncMKGe41jh482rbIV+C0XomSnGXFh8j4xhgUD0OQSdFskuBjrm0Eog5UX6WvBzJGEZC5txGLiKEiwoVNyLkKBr5X9Y/sRdI//D8r9K/uGfbf8KYw3vInS86sLMBuqdE/Cu/xFX6KM+2zqMR4agj68KqA+B3yxWQb1/+ccqCLyO/9dR8Rx9a3mywOJlqtqClqJ+Z4Q6P8WpOlicfLlTnFRcfZx4QAZDkqSRyVrqapHhEcFnp86i2L6EjFEpDGkwfwPWfG7ntHRUII6i1JhgHEZt4Ds2tFimaOk7zrfL+AN56aUDzTPPPE1fchhrlInm0w9+KkqT3bt3Jb5KCUsjp4bxpLXjFHgdX2o5MjExmXqi30Gvbvyn7JijDp1nG6gOZLV02Y2flDFktmPFeRaiN2x0EbPGk/qf0XJSOFqOTCL7VdhmzE4+LDPLULkzhxWiCr5ZLR1R3Gxim5ALXpavvlVmsFo5+PabzWcfeajZjYWlYzrLzTqR8Zev1g8aJuo0yb7qZW03+vKr+9rdV4QGdpBQQ7sWSpy2fQRekq6dvg9t7Rjd1+7eT2HWzNufBj+yQPA9pgxeBg/LqVnjrYspoDSG9j6I2kUYfFj+sCx48DJ4WB55jbcu5vXiD0OBYzqvLn33kjz0vhbrCe0yRwKFUZeytgg4EEhvxN1JR4Pn+MtMwo7gx+Td5ihmgNN4+L9MgMLExrDAfskFV361LqBB29FWB2yBq2G1QYnGLoJvdLBOWuywzMMoM0CFe8QXODP4t3eSvtTKlnKluI3DoHyX/qKem9GrUTCw83Kw4ARGfip03AbiSpGa7Ti0ZGLgoH4jK2GuVjAliADWs79azMADgQ3chl6DzJq62knHnDV5CrohfhNyvS//iSNtkEha6Ozy1uZTgMP8G0kFFMLTGZf5UmhLNRPdZmG6+Syn6nyudQ6blRtoTr7kMbhqqw4S2UGSiNureLgcv0KoT1sohK6Y/5LUfu7YiXPNG2zVevn1dzjx4QQKlBmsFADPJH0dFg3icVCZiQ5I5JdKOSdPdpSxVtDskvyY/1FM1TdklcHyAAn4wkPip3W3fLHjdG9rspAwcyBvQNlliper859IRm5v6aNIqaWHn+3wo5Vo8SdiIvu9q3+FTXwtqGX44xyWrTp/F4uT++tUnRAMtvT5bf0icUQx96AFivXQY0ulQ6WANc4wfypOIQ1PHCURx+CLDPKPHz+D8uoI2+fwbcCpVjOzrqKiSKM8tGpI2QsvtBQPc+Qp6cfo+HQ0bEc+xzasjQy6Nm+u7Tjx84N5utRRgEUPaTIIlxzLgXvIlCafeRl+bcP8TidNgsTxdn31n4jJfPFfaMm1MznyU5iESXhbd42uWWcpTno+TjiOeMdWOu0oThwkwofwuC1bEnYwOtoCl+9po8HCD5f5zcTCu+98qMl1DXQcmKgwqVW1UjRJpU3YK2zgAyXO0e4OCi0boWqy3PEmUdOOxG+d6PhqTOF5efc9Fy9X438idYmIXDDavPE9A17qn8pa869SzQGPyjlZHUUV35WLWn0JKpOcDtaK8s9qGOnFoxLJh5Rd4jsINj+r8Vf9H+ZPOTKQkfChg+KTV31pH7itlf+Mz60uLT0OUJeYELs9xK06bidDZJGWmg5Aq2pRXnB9v9o1wG+EEGUblsoukaXGP3GCfx5+XsDx6+HDtNXT+CxBnjP/ZqvLNAsVHCPMhFcl2yyz9TGsHTZtYlALr1Tkr0exoqPsC5xo8Iffv9I899xv2Zc+3jz5xFex6tiDsvQojmOPcgz85WbjBE77RlCEnOBkDbbcefz7/GVoYxCfY6KZAGg1NoHCRT8ilteZs6ebMye1HikrE5U7+jS6nBMP2ELDyt8+nArqc8qjLf2zbLXYtK+NlVflNosSOiNU2XN+Bsu30yeJ44otvGZ8QcZ555m86bB0HN9iu3ZubG7kdLjm8unmk/fe3HzqgdubHShzXNxRoSeuWCJ2rG7v1rFc8p2Xrp34zbAUSwJ4qpc4XXz9IEortjkc4Fj2t+jLTp2fRgY68bQvU/Fj/bfilsLG1Vkh2vb1A5ILuH7NtQZ+201kX4efiJJg7ejqvxYwaf+0P+WdUQeXcsH2LxYbJpe/yUZ7r698p21aFh716bZK+970JUS2/y2llpNq8dt32s+U/DM7jjPEv44+eXC1+G16tkOxXw2/gJNNHxKrfn3rJ/K9y383/or8cYxoTkHQWZzct3+y+ebXHmw+w1YdwUrn+ihrpJUXJ/zSKILQKjIi+upXaE59yPjHsLpM0fE/eSPtB5H/wvYKjnostME7/Fp4iNCWf5dOAvNMhDg/pk5lrCYd/HMcnv6V+3lOXjv87nFOyzlMP38Ux6Gs0rNdd2mU7Q4oTVw9MS+Or2wjUbTxTflslan8g5GG55/SrUbeFTYyOBkKisKHLkPcebf+hUf+hDrT1VVfhi/XKn/lwyWsDKTLPsQ5gls867QWypFKYd4dwytDneiHHr7367+B1mev68cvI4ZpuvqpdqnJAABAAElEQVQ36FvafArQPEgJtYs0jp8oG+YPIzjhttS0lhOSp+o4NhrTx4kWJ5yq81W26nwsFif0INTP2qpDOSRv5ffpGMrp1159Fd9JB5oTp0+gMFjfbMPJqgovHajaN1b+KFMUaLYRt1fqVHUzTsHNh1sXtdxwe8/xY8eoE2fiP+ruu+9qbr3tVk5b2xNZaZ8TuencBflwCUeDp1F4K09VWm/DT4nKqsuzboVSzqpkUXnh9ih8oEQm4ggWxcYWtnKq4BKmi9GWo3Q4HvSknXOcPmaZjbPtR19WKnmkVfwAhRc1BtWXjVYsOse3/CfJs3DqtCwUO5cuNIcOHqR/e6b58he+2Dz62MPgd4EecPxZPCkbi5RnVNr8cnWBw4f6fo3fLolVI3W3vQ+SdBEGH5Y/LAsevAwelkde462L+efEH8VJn5YioiOlQpZ966jrJ7rm8xBWnvhJ+7pKmmW42jjLvn0E+NPRAdvKI+zBRakraK30kUBBTCNXEFLN/K2Jkx2T1Y4IiQPEAEOUpmKPNKenppsXf/9W8xv2+Z7Eq/4MpuZLDLxGNaNiUuLASriLTmCUdsD2yspv8sg3RoquaPlqNbdq54hUKzt/rViCBF+AGXKkssiqhxAogCLVOEQ01iBeQhWsXn4nUloAQoe9hJoL+zrBANMjtsa0aEAwRTjRWYUnkhDYPgzxKyCiDGrLrUzLCzfRKm8+ePX5DzWhHKAZEJF/BX/hMow8BEAlFWflP+iT1m9VMPK6FcJ8XGRgHd7Fx8l0bdWJ4sSVSBRbCHgHyU56NCv0EpQDkg6/yAK+ghNhbfzJBemqbPWl8cOfvtA8/ZvfcSQmyrRZoOjjZtMEiz4ONAElrQhSBzOltJFP4LO+cHcedpmtOTrwdBLm4E6rh7H0LNAoz8Nv0ilwQ6k3+C45Vjm/WYmcjZgRL18JT/nn1wDqtmng6JD/xCAoyQizLAKGiFWz/MZlhCROzHwLj8K7qgO94MRfyKk6m9iq8zBbde5N95qxnfCoc5ZjrWK0mCL9DRMjZddSYGhWVfjqc5erOCZrozsg0rriHY4V/uWvX2xef/Nwc+YcKxdui8ER78j6SQb3WvQkk9leYUfnQNPVVhxEQZKDFJ1DXoS3KBZZJdjIqo8r3zHTBLdH6ZlPKZSH5iOTaD7aNkIeYV39p9AS0/g1EJV+6+Kw/iWffjaSl2TxnDS85nPCuy8t/owICUDulLKBcBIGHgImbZW39dTFTOJYsVyHZdZDn7qj+fJffoITOlSYFC2WYLXN68df5R9yQStuiZe2qtvywJVSHWpe5LQcnajOMQGWXyqwKl8OWaEY3mnBkdVgBg8QD68FZm64evm33PxW+MXW1j/jdRfxiyRiEr8qObB8JXXltw0jeADe9L4Hb5WtuPR7o6NPlcz6gIqVDO11mfw1ooC8uIvLelkfxdq2P2VCFxHazE9INL1XD7/J8zkPQwR54uf98Lec6oCGpm71OCvqIEbklOLErTqeqkMbgNQOc1injJAQOWfwEP8wy6boX8ZhHwo/yGjhmS4JmayIghd9D7z06hvsQ3+dOso+fpypj+Os2cHnFRQLVoIRBp8qMlWqTG7dzkSWwSkyUvptu54+cOBlT3nhxDKOH/3EJx/kxLJNzbGTxyiA+WbHnj3xQXXi9PnmBE5hPUFlcoLtPecuMiBG0UF5wgBMtDc3O3bi4A//QNPnp2Iivkh91WeZk3BqAzIauhhAu1iyb89eFC1bmrOnz+CP5UR4Yp+qfPefSnFp1JJz505OxNm5nXZwCfPxg1mFdHXSthBfLY4IaMeQwzaebc1tt96Ev5aDzZ34Cbh593hz1607m7tuAcakK60wUflJZK0RLRr/wm/uuYCbfsavjknCd+u/NW/5+McB+3tsxfnZL37Lsa0oTdj2cOToVDO6eRKe64uLnFtJuFx4cSLnMc1XMCnXtF45qQXOZU+6oGychGgK78TVtuIijU6ANyJHLX8nE8oklVPGk7TNTEQ8mUirHJ0qZvsQssN64AlFxhO3Eyn7S9/LNw2tikmNkx1XkoWtkmET45v1rEy7lUHn/G7LiXISZPY7IrU/tWxTt+Wg8tSLm+1fLnkl3EeJ7775JAx+5OiK4EQzetf+Eou4AQHyq/a/RPBfyVNotxzBlW3ROFu/1+OI2aqTU3XIhuMHFSf2VxkXIVdDcPqXFj8wvKRVWNQ4/qofS91IHtrsGZHofGrj1t1va+c/0SvYtN11nfXP/j+JSVf84lV6Ukat0JBW4qnommFV6o03TzT/7/d/ynayM83U9CxdBdtnN7IwQrueow05bndRz7pufYhjd9rXeix4Rea4G6zhWXKngrzlebLZ8rDGJu+X/5TsGvnv8kO4eQyGGj9rtS1zdSKqXPF0LeWGtUj54lhDuad8zNjC8jNJb/yXcrSMKsjglBe3ulbwP7y1/FOZV4y/2/IfABDmGuWf/ngwKWfxEj6Whb5jcEl2Mu94gjbNqTp3qzj5qwebB+7eX4oW6l3T0LYB5N8xHHYfxR/UkaPHmgOvHuC0m5do19X+t+GHRN5YDfS/qPIzp0hScipMzH8qP7y17csbt9ro7+quu+7iZLMbInNv4F3+K2vs16IARpHuaZUqQtwS5PHAKrW9VL5koRkZoiJLhZblcBFZdw6rD8fosR5BAS5etxpnyw38FoctTGWYMIW/GVmmz5UoWOCNfYLFr68l5yQqbi7g5NztWlqX5wQfnd1Sb60PKn5nZs41L/72ueaHP/h+89WvfKV54onHUcCjXAFnhoChvOqCj+Q2X1JHwilLp65l36pSdEHXca/URswTP139WyvxMlxthGXf/pPgR9ZKSXcViXkbPA4eln/mrctDYiyP1gFs79cM7MXtxRs8Dh4Sr//2YfEnx7Zyr/bmo4JQkZarRdR982tNxIhBw8yEhoasfLJzkRmu/Gj1oYg5eux086Ofvtj8GsXJBgZ2OkNaQMCJxYnUYiZ6vCrnQ0S6RZ6IQ82qQaPApYh/gQ8O7jYi8TrJ8a5oMY5JUysFObj4HiTc/S+p3YNxevkfBbadYlZJgW00t96oONHSZM/evemMFcyBahQeQqFpJZ1b5Cz3XEZYid8IXf67eMTqeJ10fjdpCAySIOtKKFCJsxL/ID8dXDs0PoZv0CGXdbTKebQITixOGFj8BRYnnqqzj606Cjzz76CJLqv04iDLoAHeBH+L1z7pqvglPVcoBb+580SIpeb/+cefN8+99CorHlCGlYmdt05iF+hUkJHAJKY4SN8NWFNu8DcTB8wUz8ZsHNM8Oge/JT5ZzeSS92GzLvyBZj6AG1rCfz5E49FR2nGXd/lnUgvUh+W3BKWKDIOqTFP4JBlcAmknfzx2+K9W/gv4brnFU3VwDlvHEcNkcVu5yJ8TZv+nqtMDCL1Ig3Z8/wzoJUR+257MVQoqUCqPpvNy0vHyq28333vq53TIp2mjDOgZ+M+5B5dVXGGIzS0utg356hfLPWayvKtsu4wjri0c+7mDVY+s9EFpJuvWvyQwPf8l/89a/0VOJoOcnNhrkhdzYZ0ufskTONVu2TG2zmFtByNs1WHah+KEU3WwOFFxso6Jzxj50tN9KSS4W09SKCRedi3Hn/x3jGi5a3R54oDVgYdmuCfxLzHjyloosw3zlPYn/b7SMiBU/NIr6tSpq9S/Lv/L8ZNGWO1VuKSXKziClce2nnWoCbP8h+2fd+PbpsAfvoU+Pd3PMmhycOOeZga4wMqkNyvvLR7xcV0Tf4HuyCIu+Fs6RN3HL6xBviS0V/7J/weofx1NjndtAk605DVzb8yMmWxuGsGHxu4oRkIU9UB6fM69zZUDtve/khNwoD1me0zL9QJGh+gag+1fB+o//QVHRL50CGesu5srtFX71BF46qB1E5PmRUyePQJ4FOd7Kh9UsC0KAxNmHawucH/vyDH2hO/C8uQt6pyyABNvVvp27d2RVbyTU5yAMHUBmcyEHZmwwVO0iOM+d61WPOlgzx7iUu5nWfW8MIODPvA4OJYB9s/VRlAGUP76VXEVUbPy4yhonATZr3bXPIN76RSucmSCge4Vtsaewhz9PCuR6zebTyYZJFiXeo6CloHzpvGt7IPfyjhjpHnzwO+be+/a1zx4/y3Nnft3NLfztwOLGJtnrA1VnEhfKgyAqpBCQlfWeamiaEuvq/8Vw+qkMvPwO6ebn3Lc8wVOHnv3PXxCvX4I/u2M9aSnXqmU0LfTzPSZKDO2bptA8TRFP7fYbN2xNYqRS9OXwgMVRbNMNFTIeOymPruceEyyQitPLjLB2UA7Mt4lzN4lfPv2nazmnsskaRv5dxw2g/NFV2nl4UXN4ymrvdRPHUPqvHcL/Acciu7ZTHgsj+40ku3bUb6xnctJUAYxqbMtI3Iz/6na8MX2B1+Ik2/+tO3fR//qSkweufufgA8v/wuuEAMVWJJo/59MgV+ZLk1eZXFyEcXJ+MA5rCTa5+ncmZkW6eynqFGpT6aKJC1ERiYnBqXkI38rD8Eb+WsM+PCh5H/AizRX1a6ivZja4hU+Mbrxt2Hm0Xy0WYX1ftQaRJq97OPowyD+PIqS51882Pxf3/qflC3WWKNbGCuo2NTqWx9ZygZxKLtsXVyCCXCQiD3wySl1V1pcOLV9W1W8HPcrHwf88zGXEVoYPKb8hXCV9teNMbQssB7aJ4papaGKeLf5q1i0rmrZ4FZ6+Z8JsXiI2/k1+TD4hZUr+fepffAGPf3xz7D/u3r5V28hHPnbKk6AuZhxOdAXh4qThq06d9+xGx8nOIe980ZKD4WBdVSLXeTuNArkN958vTmG022VyUe4Hzp8CF5sj8x1/qXVxxWUp9KpQkO5G18iEOs2yLB9ge0zyH+3+bz5xpvNvffe23z9619vHvjY/VGKyj/lTSl06YcoC0+5uXj5EgqO2YJNGUxs2Qb/UboC2359DCWb428Vucp2ZY7WLNu2TbKdZ1sU1ovQZv9v9sUje40b9w0qduPTZDIKFcuWjBPF8iUu8WeRW9PA1aLSBWzlo2lU8F7hBEQngyqLpqfPNc8+/avm+099r/n6k483X/vak6GjlKpV/yzS7irFSVvWfhw8Dh4Stf+WJgGofOsHdEAH92sGDmL1kPYel6ftv/3vxI+cEf3wWkYYn9tmNIwweKqY/fiDoDUeVsbO+xqJ+5/6z6tBVui14wxT9WObZd/bfq4y6QeuxDPTPKTD4J7OMQFUYe+GJWbd1yFEjSsnFbw2oxNYmfzqN683v+ZIxI0Tu9kNqSAHFtpHdfc2B2PaHkwlAMCkcxCfzyGMIBulaROvvhIv0Aqn9EaokaI+G3lwKbKq+dkEC5bRBvnvYmarjS+E8l+B4Iqp2lu1mLvY7+eARhx9/KCW/NCSpMlTB9RcXCf+wCW++SkSSOlVv/no2wr8id/iX51/uYcIRgCr7bZDyhGS8h6z7qXF1uKkdQ6bfdcIKxUnrk2ykKi8Im/8gCPwoaHKv3hpds2/NK91mX8dCRpsR/4v332++cOrOqBCO87AfgkBiKSMHwTN+7QycQBgHQgrrRsBLIylOCh8E+eFG7FSUWMdxIlQAjaxrIxtmuK/5A/LH/mawUYspAp4wJiPPv+jJEsl7eCFDaEjT3zueC7K7rlFzq1w5gSFHn6hrap/fMxxxO2pOl/gVJ31jnK9JMwxTctH71FmidNwgGF/YAT+/CLnKsc8sKe23lJWxhFebiPNawePNj/40TM5FrzB0gTX+5QFZSZvA446lPiVl8oj2OgAtcpYZAVhmhMxHNRvZwBvB01I8REaHJh2+C0LQfq3Vv5F0+e/Oan8eDcVb9z67e/a9d/Uw/anbwS/aOk26mAvmal6pcJOsSMNKk5UDI0suFXnCj5O7ohz2J1by8eJKgB5XKVQaVry+L786uNfnn/wygSuKE4cgCgfaQMn8SlxkYnOKO1wFMUiMfhXtIsnxSF+npU/CRKUf8U0Huq6Ov6C08Xr4K7kf0ezwAM6+AtRh196kp76mpVegDiAmmHytokB3EZWx7v2V2a/KqA+CH7xtVm7Bv7V+a86a7l2+IXk3wfB7+TVzGfiwOM6BKOWFRuwONm/b1cm5ak8QBa2cXP3wVfT5ku9G1btKMGDn0pTipMAgXDTWk/1YSQIxp7NUz/6FVtg2VLDaTkLSyhKIq+tBixRsE1iEU3IHP2Wp9bs23djc4r96PoHWcKa5dbbb2xuYFvOJZQRmJU0P/7xb+jjRpqt23ECiOJEwXvm7JnmNOEq6zZh2q3VGcVJP1g+O9xqc9P+m6B7EaUGSr6L5zMBt92MqbihLttXuoK4Zctks5tFh02sJF5kYn8ak3Anu/J/zsF8KhyrpAzKPTp+P3C1kpiawoLl1Kn0wzLLwbb72ek1mCiwYo7D1S0T2+iXGV8giw4deoNTVC42t9+6o7llz0Tzqftvbh5gu85OLE5UVrinX/7JY/8G5W+R+KG9hRzjWWbcV45/jOqk9QyWOk8/92YUV4dRnDz74mtMUiaaS9SLS2yz03pDoxOPDadZM3kYZ4vUOfK+GAuZBaxNLqIcdeV8E3F1iig/tuE8UQsR247+aKQ9ihMUIq7cqjgRhhY5U6zoOonYirm81gUXmVguV5zMNvvgvX5rVELJV9vfLM+b6HfH8Q2gw17N5LdyVLK+kpJpMm9Ntu1bqwYM4qnYJReUP8TwA3/KAvkVWcCnuv749tdh7CAO8UOjZIRSqOXZ0U4UP3wtxclMcx/OYb+BxclnGOOocLPsRil8t06kn7JwmJiKJ9QWgpLN5slqk8yZPwMLY8VvudOmSeCKH3l4zfFnL63ZuZ76ZzxptbgcARXjXXE3xEve0FIg/hLW3i+8/F7zrX/4R9rmdjKP8swt0YTPU2BaoAUeP2Yv7ZEHvyW3EgQix/l+zDdS5WrfVbxUSQuggrrf686/CDOORzlqmQBTyyg/q4zV+ttJs1ZUkzmRhWNrCRedSvn010SOpWtVjJDwQfBX/kkGHPEmK+Y/L/XRbyXl+zziG8HW/37+GUHwz5mO8k7faPJNjmPNQdwsYsLXOIftjiP2VB0UJypa1mVMgMxlaKf1xsxFjudVeY0l6qFDhzlo40hz55130f63IZ8X4zxbvrldWseqymLLs3yOWD+Q54y5ddT6+oFXm5/87KcoVvc2Tz75JCfwcGRx+IjiA4WrVnFar0Wxgdx2EVnLOOXVKLJI/2rikgtVvy0vrOj0aYK8sf1p5TEOHcbX15nhWp5qjTdG3XSMM431SLaREkfY0i5Uy1armFiewFyPG9ZCRmuVnMCDUsgxp/6w6mRNTuYhf+uxeD43daZ58flnm1/88ufNN1EKffmvv0RfgVI41PaKU9HGtdzipL6tKMr6OPit0GvHGUSu6tP++jXp1kjc/9R/HkLqnir02nG6uC2+jxD/QHFydQK67qMXY/A4eBhSmKerfTewC+vulXT5W33rQkog9WIMHgcP/QQ8X+270QzjN4Kl6xCt+oXF8BIMiWZM/tkd2eAJMZCGW7IEoZXIwixRoo2CMY+fmW5++fQbUZy4CsNJ4fyRXMEYaAo8nqIlNzVhSC3BiS0dsm8tncGvMPcT3wu/kzOFQpsPEwcCFCSCHyq/9b3orMbj8/L8r0dYd9Ykhjr4c6Dm6pGWgDt37EJ4SzcUho7CnxWUSM2iLaebfAj8lftr8D9S3Swtz38f/+r8ywP5qeBuy4vVQyfZep5uFi5gcXJPtuvs28VKBHEUVpr/1+oenCAvYa388mFQ/r38L+PzCv7LK2i2k7+McP/Od3/DnvCjULCJkTgWDuICNFPwdJROR+V/pzhZItBv+U9vM8N++0OHj8S03EEi5HJZlygZyjB5yKcQK6T2T6gVL4oc4oupW8kxlpd3Y9bFEyP2rv6lyiR7plVFCGXQ57aVJMxPP/8Stxr/2u0PHsxNN7dicaJpcZ2qQ5dkeUuMbYee2dUHa25RSYjE+p22VFfR7BKQQVabwXdozisBdtdS99qb7zVP/eQZnDROs0LNyiWDqjnqC2gKLtiS/5Q7H8XPzQGlK9xXmACcY2K2iU7MSZJ+CGxHiSBdpJbHRahvxfeV7a9yaXjF5JangfwxIwkkJnf5b5lfu/7LAa+C6sDGMstKeIaccqDqHv05sKTXfBHHEXYsTuaahz91e/NFturs4DhifbtocZKcwYTgTyohiYd0IbBwdjQP8h+aW76UVDQhcEwmEU1z8tTZmK868HVoHEjtQNhw49leKv9i9SN/q/CbP1O3tBDv6vwnWi9mvUlnW/4D/gPNPAzwAzvEV2q3bqlt9WjZ6Rkcirryhc+HbDNjwqdVQkyrgVs8qfoghV4dpfX2wfEv5/9Hk/+hyTc8h81jWHcsMIDfxPGeN+zZHeWyReAlX9JW24bnjWlAe3Vl1N3r87I3BtjD47KrnthPCseu7grbG3/yi+eatw6eiJ+oqQtoNNZ5PDDKDdIuItcX0XLs27uv2b5lFwwda06xJ/30lCfsXOZ44duaBx+8r9mCefNlzP6eeoqjKzndYffeO7g76DyHQ3eO+2Wg61HD/GQbpAobzaa3b93V7N61kwEtjlEv4KcERYiDasvdE3Cc4DjYVkptYl/5nr03wKdSGpw65dHIVzJAVvHqEaEO8HOUcSxY6lQe96+fPHmyFCWuLKPwUfmvD6tFLH30m6Kp+K6dezFDX8QPy0kmFqdQBiw2u7ZjdTMy23z+0Qeaz33mgWY3zmGpdiggbasqJ73L8ZX1r2reVeufNdMyQP5pGXkCPxEvYK2zyCr+25xS8vSzL9GasTBhomo/twnFvj5wpj2u0334KHAunD+barKVrUVXyMcMYZ4C4VYbFUfyUZN7F2u0GHHiYfubZdKg9comFEsXsVKxlm3jGOdpJiizrDJvZSLpeMWV6Dh9RCFyiXGL45fdbLvSEkVHnx4Zat41jdd55Pg4231Q0sj/7du2l+KETPalsEtcV+v/3l/+rtH+Wvm3mv/VFq7KfxuARQbIvvyx/7dcvZSLKsD5YYJFW8WHW52q43HE99B2aU/A0T8eD/w5JmK7gEkCPGDqBzj5Byyg8Q3kPH+Y8U91wtY5OHvV/FfOr5p/qPB/fAxxt737IdnNsxNjrzCI/gtlAwLjEnXq5QMnm2/9j39hV/Q2hjMTLGR6GhRKV2WFTvZJ0+VfGELursh/+BTkBMq6jv/pGeBnt5AR/rT4K89Fy/Xlv4VLgZZvNLjuZBv462k/KkamsWy1X/FoXCfI1nnLs7bNg8tcUEyhn+95yq3L1R/P/w9S/lKUcRC06BvNwy1cGNTSx/q0hNJbGteNYH6dU3XcqqOPEyxOrJvynQMT7DMRl8gwoiN/3mO7ztO/fgY/V+80jzzymebWW26nLJeiSNVBriecjdG+PYjBq7YNyissMlgYO4Fs1b/Vv/3bv8aXyd/87d9gcfKxxFUxtUHliGM7LRfxUXKZI389iXIcf1H25cJTsZGDAfhufVB26az1ErLKrZmenGM/obyx7GyjNp9Sgun/xPEBijCE6Wbk0LiKGuBaUlWPVH5qhUtdZavQFBaQ2W6IY3RP2tmIBZ6X7V8ealGowm0Bf1pnzpxq3jl4MM5hn/zqV5q77rwTRbuygaht+7Nep/GA0bofvIJadrX9b2pUG0OW5nHwsCxF1b61oQ3Dlqdd/tYH958TP3KU2tZn2SAHg4d+LpY9L4/RZbCNsiyQFxsx/yOgl/F0WcQhX/sFtQzr8GV5yuvHXxBILR0S5EOefffZMP4ieCrINMb06u4+VRdrzUX0pkKqOBnB4mQGi5M3mmeefbnZvIUOWUewNLAS+tDqQFCBnf3cLXrwxZQe3JKT7QGhwTcwcENmJCzvfvMtwV3+fZF2bvnujTATO/GE1Hw2f7mGEe2To3ls82/c0rZy/jl5y54+7pkUEneYUkCFP9jCOz59QPxCWYv/+d7+9HH2nzv8ioBQNggsgVBmlXyUD6wyataZrTqt4sStOh5HbHNw64WC0TQR7+EVQr7L0Bq0BD+8qRUZ8RCpl387P+sHDrOb7/z7L3EOe4R6gtLEFUwnhA5mQ7ppLSTuKkz8yGOBso4x6GOgePDwUUyX2ZPNwL4GwUZScPqOkO4161XlDzB1/hGgDMQrX6T3Cn+4pwyhuaRyghJsUPtmCunrxK5wFMAh2MA24lr4paAidBEL/+IVtupocYIzuy8+ssI5bAYE0lvkmT5Fw7ujmUVWpAfvzrDgR5e33KFHXmVgyb3aK1t1OJLw+z96mu11nqJTihPN+nMM+AAgOFJA5lHMXtXdzDHwnmJbwAYG/9snd8AzFQuUHSsBtlfzqkx4v/ZnzOI/OWz5n2+9n45bicpP0VFvxX+/EKsXsc9/TV9sEcqDKivLweikti5AsHVsPRNjj0BusDjxVJ1s1Vl2qo6KE2F0FJD2GvV/mfxJEuM7mOgGpNR/6hpDxLDgzBkVJ7V1QeeqdvIqupRQuSDN2KDkMrPVZv7Y+lf8B1yP/8Mcimt49VjMR2msGqGDOl8c9GiCu5HtdJ4CY/WxaZtPT3qo0uugeOcyUoJ4T32r1wQlwvCnS1lfhvhJmP9d0fTLP6LFBI7+E6GD4p3rKvgVAyEHAJKHeMwAftPG1sdJ2iblEdQFq8g3dqYeQXc9/T/qD5JQS/lLfOkCsPCsmyocXvjDIZw4z7Fi+Dwn1TEZ97QWFBkq9q1XOsu75eab8TtyBWetZ9k+wmk8nEQ2j8+eexiUf/yBW5kzsg0NGfv7P7yBwpQB7rotDKincTx7HvmPU1cGzyrP3VOuCbbWolqc7Od4SsOOHTuOtQkWFC4mQJem3Q5cL7IqetmJOFuE9u/fn20jp1EEHufoY4/TdkXRyU13Up2TXWXNzTffwoB4Ow5jT6PEfS8TX5VuymCVJpqXa3a+EZm/jXg7d+4mD/hoOnKUQfU5tgJhGTN7odk0xmLH6SPNk//l0ebJr3wOi6Ad1Dlaq/II3rlCaZlU8VtWMLaKqQrZQvSvCjBBFoExvaw5Wt8cxffLC79/lcow2RxCcfIrxjoLi5s4UUfHziwLMHnRJ8sMZuP6aJjEZP0cVjTC0aGik4tZJg5uV3Jfv1uYnOxvJZ5baXzXMa4rtypR5IWrwFqcKMe379iBIuZ8xigqTlwMuEQa8+fWOJUhWpLsRbF34QKOFGmPE+NYnHBXoeKiw2YUMaU4mYUmrBGUMZRH8UOmLL+kvfvaf5Y7/hvKn5aniU+IvOzGXwFw/e2vT0EfZ/csL9LniwL6IxMJzCIXxxF7qs43n0Bx8vC9KTzbVJxIOgZI3XU2SmJp9I//KX9uwWFdaHM3zH0bRkhd0PAB5X/wSIwP4gwan3m4Rv0zmnTKQS97hLwxpgupAuK/2/c8gGEW5eTvDpxovvU/v0Mb2M44YXO24lYJIItHeSJttLL9/CMbhNy/xN196T8bJ/nvMtILXCV/A4AI3lfk3/rDyCTt3GfrvuWrpYPPF3CA7cTe7WxuhdMixUvFSax6oT9+ed5n/BcGrYG/yl/a+JMXK65etlJcwxhrl3/6aDKa8QZy2UUXLXfSl5N4CV8z0jK6jrnQHKfq3L6nefyv9XHCccSOB5C/Ood1oSfHUJPGvujNt95mG8oPstXmL/7i87E60dn0RvpZLfu0xLavcHyho3utNmzabvO8PKMT1nPNSy+/1Hz/e9+L3P3GN/+2ufeee9M/azHsiWfKH+MpQ1Q261NJue/cQCVN1Z/ycaICZQbLkamzrXNX5E+2/iFj5rFus89yfKiyVt8kMzj7voAy177Dk3Y2o8SPQgX5Z70Yoy+z/JX9WtS5bVmFuw5gt1P2WspptTjPN+NkIYbquqgyBmX09AX6JaxyLs3ONJ//7GfCkzEdVnf1G55X8VaJdhYnw0Ltl/SKStC+Lo8BPCge1Idlgby0eK3uNWbrYC6LaFWodjF86CKuug+iJuTPi39gcdKjeAUZKwjqk7+c8kHIys/de937v4MkPHSx6tsQ6/CpHzvPy5MMgld+7t7rLjywUZAKtOHVxfKLITZ5BSeD+36HYGNPMkJ5cJBWcAoWOki6o3XNKY4yfOb5g81vWIUZG8c7P6umbteJUAsMGggrU9nPLR0CZQItxtQcBYaTP/GDIdRCh5+kXWxW0qr8eap3v7Vh3PJkjguSnUzxs2Aas7v4TqNVQGTVwgQAv8y+vmkGKq5SKwhM7X+voWwd4leY9JU7xcvrw185CmRQ9Pjf5j+IC5WR1sQvaf4ZrS4neMaFMv6QodCnxQlfFzlamRWZ8nFyN4oTLU4Q6g4yVWQQJTYJ8GIkGnIB8SdArrXyb5B/HX7LrUqQwSb8nAP2t1GcvPYGA2M6jQWsG+bZzw/jMzDvVmNSSuCMJQcw/GfnrnLACeVbB99F8z0eB1hVH+Gx9RQ4DvY7GqVTeqo7tkYHctFIBhwj1tWn2jTWEAW3205a/KZalX8/yKuKI5RK23HgaviNb+zuKvxanOjjxFN1vvSoihO5J68hlP+xOOGDvO/MhgMhHxhkDLCbMWlo18p4FJLlKTydwhpXoK+ycv0DzPWPncanxnoGzwysXLnO0d/G77V/HgVUF6sI6dhwYqjiJHuPMffWFDwe752gEF8zStOZbMj/tfNfgKXto6r/lU+5If5agLFclTVmRh7ICfDJX17FLOnRaTDw1sfJwxxt+sW/fKAsTuCNXtjlv7wpPltLOmj1hVe+Vf3v8JezY9KRllII3koFTbyx0YFEI7EQmI0TICmDmPzBvRDZQg5+n82DA6wqUd+E7rUKP3GltOq0MbtreaqPgv9u1VFxkq0DDJqqRJn4IwP0MyEVw6uP3xxJYXGtaG252zGZhFb5sJGYpuiCKq1vxYeu3nX4jVkwrx+/CrUgNBckU3GyhMXABnyc3IgPCcSOJORHfFYLX+W/Q3v1RPXe/yVgcNV3X8vqz9wrV8gNwAIexA6GZy4tNL9/6SBmzhtZnHiF40VPs/+cukSkdVg4bGVCfftttzVnTp9tjrz9HhN0dCpY9s2jGFxAUX7jvi3Nrl2bm3OnjjS37L+ZgTFUjm5tDr93Ls4j7Sc2T2ygDYAf6zF5dYmj4idYbbz7biafkHr06FGOEj4PN8gbTvtyigWD4MusDGodMsnAe+fOHVgxbOUUtVPxt6LMsf8RHlUg8M2vixO33HIzg+EtKBmms/feSb+nusiLy0yU3KLCLn3qzlyzB4XJXpzMXrl4BcfW7zEA95hLVkE5sWIRxdDoEsqfU4ebr/31Y5ym8mUmBvtoPQyykUsLDKY9UvmPGf9Iu3k5Rf6fef4A8m1zc/jo2ea3L7xG/7QFh+fz9FHQjFJCJ9luY/SYzC0oRC7Ex0mDaf1OVlwvw9eLTHZ04riBNOXo1UmhW2m0GJGP0q1CRGueWJyQxnqxEz8w55gkqPhQcaJC0u09Ti488lmrEn2c7NmLxYlbdWiPW/EJoOPPSyiw9FMTB7Pw37jbkN1IEUqkWuba7Y/M899rdfsjwDoKbZR0ReLX6Mr9P6b92Y4Cqb0JvMOvzHbcYlAaHg9pd+RTH1X3oDj5r1hwPsKpOtaDwdGuyHHrevk4MXHl3afIENqdVNuuxJWra9i8dPjFZoqONKOGltyvLv+vLv8K74D/4BwsSAHY/EbwcVcsdSXGmj8B1b+HFve2YU1zAYvQl1891vzf//1fGS/h8B1F35xOSanDKjxH9Y2Wkh9S3ZVWwZObgQg8sIUX0lRkfKD8F2fDrbXy7/ivxp/IWBbBtF61vekYtBTx+OrBqsE2oqNj/Z6Iv5Qlbd4VxmESQYRZlvIo/QBEW1J/TPsHQK7ryb/tSXzijxIDGSUD3V4ZygbOYXHeOn++uecOjiOOxckNyDFSWkcZI6sgVqHhCZLSfvDtw82Pnvphc+C112JxcjdKDxUPKpM9Ctz8Gz+DGBUn/Nn+Z2nnU6fOU4ycgHb6VPPLX/6yuemmm5ovfOHzbNnZB49LEaUi5CxWihdQQOgMehJLYpWsKijmMD20jY9xgEQ3Hr2svFK5gaxWobUDiyCVvAvEvUTYGIqtcbZxK//Pc6LbFH8eU6wvpq0oh+0/LHfHDF46t7Z+euT0OfoEj6BX4Z6tP/hl8rqCvKwtmHUQgXM1HaJLC3qXZjN0zFy8wDbrTyb/Vt1Ujbb8LQcec9nXDl54qlpSdd7ya6NV5O7XerBGwMrP3Xvd+78dIO9drPr2nx1/FCe2sZKHPRb18tHPap57Yclw3uHgsu8tMyj8tZhb7GmZRLo/J3473f4VIRLy7Xz45zNSpmgyWzb8iJ38Jp8lCQCjltvOp/K/gLBWWBw/Pd384tevNb9mYLdhYle+OdgIUCSO3YmXgnuAH7qczBArC4I1lSj8cUZFJ9CmivY1ACQmtBb+JM6PeeSvwxmiBykqWr0O8C9iauaKmfv6oCCkajKbwQjgdcZWWm2CGYQWyPbe5l90NZn64PirszL3XCRfyX+5I1QF9tXwr86/nafCFnVWW076OEl33h1H/GmOI37k3mYPg2kFbg1CsNxIqYkQdNycvMn/D4bfxHZUVc6cZtn883d/3Lz22hFOgED7jGnzHAqUEfdLgkTep3qa/8pMcMaPDrhVnMxy2sihd96L4sTVtbIuMTb11ME+ac1vLgskdbVe/Q0P+ZXTNdnkGWZ3/A/fEeKp08aGfy3n2/wLUyTVGds/qSysuiYGIVdNzbcWf8Hr8BuvLunt8C/hHDZbdXAO+yV8nNh5Jtyo8DFQEx/6UtmKjqB3ilaFQ2RoMIhR31DnxQc/AsXBiLBUpRx46wQnHT3fvHcKJcF6lJyjm7MqMLJEZ65lknxN7JYLJiT/tlVPNnI/7Nmz5+KocRune2iW6UAmzhjFISmhixff+ZND3TVo/+DoytxIH039B5OAgpVHiPFftlfJNL7L3/ANh8kSa/b01eBEZAmLE3b5No/g4+QLf/lJJoLsTUaZFhlhxrySnudBxjp8fDAOvBBJ97VLZH1zAGLbiOd6JpxjKDWl9xQr9CpO5IhHcUspVKWeDYrYUXOQQjGP1uAgyY/EtPjb/HexhdS/VvI/8QTFQ0EZlr+8uVb7JzcZAFm74uOECZ+nibiNq8Ovsk2T3O6S/139r9KRXdX+hFf1T0qqbFLvryJ/h/mXUi75L2KuDn+9DX+vB38aDAxRKauV2eh68umWEQw9btzjqTqUUMuclDRM8p+zLidew9wO8XZPhV9em1IqUdbi/yk8EQRf4l8AOaQa173sT/3kN82RY5dRmpxtjrLF7vylOp1l+96dzQ7Mo12JPczg+sqMp6VM4FSP1UjKoRmdZXB7tpmZeq+5dP5U89BDD6OU2IMM3tC8jXNoxr3EQ6lFXXT1bgk47j/XNHrfvhtyms7x906y7ecM3+uUl7LSdDsJE3/8Zu3avSPKCjtxLU30aWIenfxk3zr5UbGqwkSl2o037c9E6PTpkwzm2faDQLWORV7CDrkyy8qhx07uRgmwC4WBg+7j73EqG5YUKoXGsNqYmT7bmMUtmxnU42jxS3/x8eavPvcw8HcDD+5S78Qvzv61qv6L0yYV3IW/k38WsXmZJ+z4qXOcRvZ7eLaBE+I4xeF3B5m0TKLEUoHR1nvousDgXUXsJKcJncdCZB4A21ScoNFSoaEsUXHiHn63B27VxwmKjUs4h93CpEJCrsRCZGNWbVWiuGC1c8dOVoSnIn+1FjGtDmVVZI3jADjOYYGz94a9OX7ald4JfKZo6eNpPK7iqzhRuaIMd/uQ/UEynvwrUapWKvkth/xGpsEE/ufHe5hFDNql/BxciVvRum+J3r0YxId++08TaPtf8Yl3iJ+4K/Db/3ekuBXchS8bo5NAfd7ce8uWbH3tH0dcx5Yq7y1Rlbht3pS0ku9n88KtsuB4wvEPMf34vvInKdvEViRzwY2/ldcHrX8BUiB5pO8AqOzPJJt6UVgIod9gnb+ZRk79/pUjKE6+3YxgWTaC4kQn8AlHMbFAm7WL0tI4ckpCV/A/3AGJvf5AW2NmjAuDDAlffM9H6eD6EOWfPIHHybSywLHK+jwvxuJEC8yt2ycz3lCZAmrkL2VD/Rd1bddu+RD8fLSStVdH9uBd+o1SpTsgv775vfo/8560/lxn+Wv1A3HB77YbbGiARr/vAglXjiOGtjEsThbnzzb34hw2xxHfcSNl0dZPnMOGCRSS2ZGIIxwf/NMf/5ST1V5ii82nmvvYZqND7fX2s+BzXqU1R+Z6jGvcEqk8uYgydgRLol1sc3SLzC9+8YvMaz75yU+kn5aHLkBOo6zQ0sRFMK07Jtni59afWpSEX6wERDlK9qZRgpxm4cx2qOxSyQLBpYSBYHOr7Fehq7XcFDLQ/ikn7SCPcnwx+XKLp1DLot9F61kWCS5ifeeJYihetExBuW5/dGWOfgyFseP/OBiHNx5tP40/KZ3DTm7B2TbWfAcOvNJ85rFHmpv33xgYFrTF55V7qoYK/U7CyOA2cNmjsSvtsJ63ERMkoETxY3vxwQJr4XRfV95L/pncmrYMaaL2MBeKZXjEQbSusi7DZdifBn/P4mRFdpYRZ1j/Q5vF/qdB8l72B+GDhxZM732QbsXDqij9D38kfipyifEeTvkLCrFY0hloWBjtl37+jRNhQonrsM6mkfIiwCeGcHjCn2Y17K3mmedexts9q9BM1kuU0jBIU3hEZUr+gFViiRAKe10EN0HEjAiEuAUqdyjiR/FDm0UoFa15JqlVWxp8t0ImQbDxxQwqW7mvlf85toB0jTaJoS2N11Uc8OzAGZv4panyz1MQ56OfE7ZAumX40zLyJXHWxN+CqFvxo3gjrsqXOPv4zU4xoU3M++r8g5fviYrgDgS3s/gvipPycVKn6mj+hpiD96N0uk6Z7XqlV8hDZFfP/2r80mZ8hp6AsG/7zr//uHnlNVZCUZwssT98Xu07NFExKq6NvR3QmNptIw7CHSx6GtAlFCcH3zqKRhlhjnY9nQPxzJvlRL9a9bcjO/wvzkFIlT/fYjYJHrGaPbNpkryI3/pmMn7M10r+23cugE9xrw2C6a6W/4QJrMNPzA6bYd1rWZyMN//1iQcbncOOSbt/RgCXVPgvbaQvFFOnu05ZRQ+tCYarxIvSIhkhB+Zbyx3SOhS0hF9+61jz1I+fYwLGPipMeZdGdOTJpBDHlxh3Esd8wY/cTVT0oEJA0eIEyK06UxnIbJ3ABJ0O2ZUF+VbOTym7tmwl/c9b/1t6KUtzOwJPYCN1nMEF/+oqnno6iIM05cd6ItkOysfJZSxOVJx8otmpjxMmlVqGOIHWk/3gMm/guD75IzWt4oS7ihMHOuuQk66InmYCqWWVsIYTEVtjamvubmsMf9vBsrjXxG95pfyhNI/mV7p5yVV5GH6BM9T/8q/SxfHe5bXSZ+xY5BQYYQPTdJ5y4MqUjt82sVKkH4tcxIkSN5u1O9gfLf5V/P8I8p+6IA/hi3nM8cNMSNZzkstNHkecSmUOHSa2dSwsC0fCjy7/ETQDXubrsh/5M7Jgb9jx2+KzPIDFp8so+L/1P/6/5oUXDzcX55kg4yC24XSybZxas43TWuwnj71ztLnEvsjNOAJ0JXKaurRp0r3pbA+ZPcURxMfY1jLd3HfP/ShVWJm7MsIWE+q9JtP4CVmYu2An24x6bKnKjRs5RphB7tFsi+HYXQbRHGaLTGLCgmhQmWNzmRjfyKk8uxhob8Ia4kxz/PgxBsnyyxVjLT4wHUcB4okI7onXJ9KtWMfoCPbMqVM5hcFatECY8DQ79xjj86x8juMj5KYb9wFvHqXJiWbmPFYYTBSEfRklxDrk2rrFGQbJWNxgVfOJe/Y3n+B0nd07cXgt46irTrr0xdLnbZhPcF/+X2v8Q0mwYr+OycL55unfvko+NmJxMoXFyRsod8bhI8p9LTqgW/PwC9NTuW9RccIx49aQyW2chsRk5jJKEC1OPHlqFmWLip2tnEChYkPHsa6w+u0ikwCdzTopuEibkje7UZxMwRcVUK7AO3G5pKNFJilbWMnVkaKWK3vxcaJZvJMnj/l0Ff8SipJYgrU+TpwkbdtWlsHMrG2kMMTGbW4tkWrz3gnknzKCx7R/+wYuk/DRviX9TyVsYbV1uY2zSv4YbPxcvLT9rx8LTIffCMvx2/9nDEBEm6HyR7KcGI5QH+7dP9H83dceLh8nkGpc9ZxZjbc9m1/bmnXEu5dIiRPsyah5NJHhhf9q8m+V/OmACQcwa/Z/gDVYtOK4Vv0TuwsiAQdJRRUc1TohozaegacDdMfiM1ic/O6Vo83fozhZt5ExLA7g5+njVRilkRMrikHLPBeJU/5dXqUInAL1SX4zbghb0tFDg3wkuVLrw+afpLlE7Tg8W0KiONG6QSUKCkmtD6jHk7QRnRtrzeqlpbiT/tRPFQfSKDxJDmC4FIZx4171z7D2qqwOX0yc/BpXMJUm4LpnA2R+d5mEsH7+3Q7tVyM6StQhsXJKn49eKk6sFLVV51xrcfJQ8/E7byC+PKUOZ6tO5cKqKshpttu8/IcDzR9efhlrvTua/WzLnPAo9ECjZCmXUkZyrDlKhrIGQRHCls0dLG5tQ8miH6mnnnqq2YEceejhh6KY0Jm3/bZ+q1R0aOWhcjXbG2lXyQ2NrJRuyEEsAHUEq5LWxZGdnnbEfR6Z5LjQbbumkWidxqro1opbue2JOG4hciHFOpW5V8svFcUXSe/8Szq2IKf08eSCi+MkTx9zMjFCX2LZd0cfO25Rfmpx8tbrrzc/+/mPm//yV1/CouYLKJPhtQxqy9/iK+IsHWVA76pCv8qHti6simP0Xj0ZhA8e2kLsvfcwLHtcFaX/4X8/ftpXmlcyLBeXk9fytf8xuet9yGO9++uVipKnYbzuqbsneNlPhfTDB8+Dhy5B70Me691fr/fDP5AoChI6i8o1hUHDSAUGhn2I8MrHBE++U6kVSOlK04hIQ6WPg9cWitNQhaoWJ0/j4+TpZ19hr+9OjN0zBQ+MrJ4lvumVPAr99Cg88pxHsIDTzkl8oaedfNW7uSwiFTCjRLBNeHVzGfOlgDRmwswuz77nt81/8BuCVFI0qLQRn6fKXGHA6UBFIb5z166swKXKyAugpPYAx7fgKOCFoW2gfjIs8dv7Wvg94UWA6ZCJXPknFyIxccv/7j3wAhj8hAX/AH6HryaImrR1PHZy5rW0yNaM+Znmsw/d1fzlI3djcYLgtSOiTDWl17pmDCLSqQLf+pEOKbxq89/h7/Lf4Ye4iks8nu3enABqcfIv/6aJ4TGm5GrHPaqayShlq8JpSWHY8t/Jttu1FhwQINeWUJo4oVdxcggT9M0b61SdlDKwM2CKdp8aYkED2Y5LzgC18sF3t5ihu+MPRRmrbzGJJE8D/ksrNAsitKRuVhmYJ/ljpoTjGsKSRyK75Qh6M7kJf6xJ0NT+LlIp6126ZI70ioS/xBI/lF4539y2h606X3uo+fxn7m82yF9HycaVB+D0nzDSjcHY5J/gOuK7rfHGl4/E8r9gpKZw2YL8Zlte37yic1gUJ0dOsnULk/115MUwO/iyeBG/AMhB6oePlA3vdlpOAM6fvZDVDvfkD/YdW/9J5ypknaoj9gKV3xXtb8B/IoX/pI1yQp4P8AMz73KTeAHYa38D+CYJZxKv8m+C4lucb8oj04c75NgJlbzmv5M7V1vWzbO/d+lS89CDdzVf4DjiXRxHzGZkJsLtKiXKC5gCjhXtL59b/DwHf4EOPWKNNU++KTUxgQV3fPRQpidRnMww+VJBkwEQsqnkb8loJzVhH+mtzKAg3JfKt7/D/FvOLb8SXrEKQMUz06m7gFgpf2AK4IBWWX1f/ruCLs1Ofs8zedvExHADDuUsLCm0z8hpWG35S+dHib/gfbT5dzKoknzEtk8+9NGyhEWGp+rcuG9n2kHqYvJCdohb+TVR/ntLqPfVlzxoKXelEADKS8siiVxsECZxZuHr33/ru82Pf/EHFBh7moWxiWb7rhuarZwuc4XtH55Ec/GCp0FRL+lfrcucAk8dmGOiwdaN3VuaPfgDmbs0xWB7e7ZNsuulGUHpqTPYBUzGR3Aiq/+TSUz7PZXFwesJHLvqoyNtH18nqRBg0XJCB9ETbBfxRJzNm9fjKwW/Izjosw9Zj2LD01osd9uBfFEZcgPKmC3g1E/KCfyfzNnetGIh3Ji28xH6A32nbMUR6m4sJ87jy0QFyzyDe9vMoibs1ifkvmbsF2dONZ/+2O1YiN3Z3HHjZHPTvslm67jKSK2DGKQLs5sor6p/1bZSdBSSrL/a+EfFyUlODnzut2zVacabw0emml8/9yoyHCevsyin4IkDdU3O3W8/sp5Ti+CPJurmayuKkzm2Dc0wkdjEKF8fJzMoM/Tf4B7+WSYZWrpu24IiDMVJti5her4ZEycnPNLmqvE5Tj6axUS9HLuWc1j9fmkCrxJFBUlO1WG70DwrtxOsHgtPKxO3/mzGGuny5YtYtFxuJplMLaKw81/6YniQ7diWmrLSd8sR+n3nFToMU/6VjOMWsWJnZW02xHspU4jD4M64sl5loGM964NxFqLNUD4X743phEYL2A6/6SzDAX4eQ4PfhQtdNf5DHmuNwFYdfZx848mHOFXn3gAWHE2Ci/CAhiLSO/4LaIPyqagv+UeUQZ57+IO04vkoHPPnT3/800YrGWp4RRAoTyKHV8h0eXs99U8BIx7zIjOLZypOpIWP0u/YibHeRcr9Rbbq/Lf//h0cUmhR6mIV7dc49nfwwWWyQf/LF2EaUv0vbRHawnPo441nyxf2tf1ftdmAzPeAME+8JR4/lp0wfc9vPvgWaRf8CaM9e7pLtourkLV9Y1lgf6yFlg5CtY7KVh0n0MCx/SsHvMpqWczXjz911Dyuwf+q7wC2/AkPZ/Kc6MFvLgt7chY6UqdTTsgpMqbSxPGr42nHXeHxIuXAtzGcwy4ssFXn9l3NE27VuZtTdRhrjVI+qLPTfrQcXkLpYD8s3w8ferd5/bW3ON3sZqxGUIjBA7fjeDlfGaXP0LpMPx8qYR2vuX1mC+NmFRZHjh5pvv3P38aScF/z13/15WYH/YdOpC9ycpE4NmNlMq4jWAuFtH7TMsP2p1xXMTmF/LHP8bQux34qY213KsnNvwpy8V5EAaLTcenZggxSgatzX+vvEnyJ5VDGDdRZZOA0liZu+7H+ae2is1s62cy/lF/2R5bzFeTsRRzXKs/0g7UNGjyJTKflz3Ec8Q9//MPmGxxF/PiTjyNL2eIDLXWZn6qN0llSx6dh+Q2eBw9t0n6shFUEf706DORs8NY9dfdEXPZTIf3wwfPgoUvQ+5DHevfX68+Bn3aQZlHIevSEAipHx+hB0OAhMYY/K7+v+d59bO/Lb30+F9w/Ef50ROYY/MoKZWfEW/ANs1S9SBVDS2rSdAP0Iti0bQdIpczkDzF8HMdpTz+H4uQ3Lzcbx3fHqkDhkQG/+/xsMTaa0FHCtSb2HVQwdvknXX/iVQPISptKDyzbgA28hDoPIViEPnjz2cZsF+PVxjH/0CEF+j+wMQ6SEOAAxj2VCvGd7B3vzAELRu9X+AGJWJWQjvYBMBuqiAgr4uomOvD2+S/F3Z8hmbwOUAFfEMByUB2zVWLnuGA+gMEkGZCUyar5BbcB8N1veHAwFubgmO2xH/zRT99TipOdCCdhUSHkg6us2XKR7Ni5QaWEiX/lBVFmjdSEJAE3I/rOL+Huz79CJfg2ipNXXjtOR4cz2vUeFVmTwxApM0jfpuKxupk4xvIr/10dO3zoCANhNeEbELokwYJAKxkrgeWZLRCWA47PnLibjqlpOiKVJWMKX7TaozhjHTOfK/IWHkNLBkvw2jajl3ORuY6T+sY3US/gpXyRlcZ5Oo15VvsW0olDC51XEMMHOzU7PoYA7bYhkgAAFAxJREFUxR7yVdtobD1pBTWInL8QHyfffOLh5vNu1Wnpkp7wN+XIM/mRu95lGf/NIHdhVTklD0RImLEzSjUiVPNRyxxmSs3Lbx7Hx8nzzdET0zE1b3A06WBlxJVikXAZNY8gc3DavuVu3KkzFzIx2cIKqRLVeln4wZXZn1Fb4jtowQ9s7tLcPhCxvRJQBAzxE1afulh1B1n4M4AtUCOa0hvP4NdXTL5qceKzbUY8XFnV5y6VmgYvOijDD5DOYT/NROxLOof1OOIMNC1bIfnHdZ34K37lxjqlTFD+OmF0IOLKjvX5NP4pLqIgtHysO6pWgilJW5yFWeTBX6Kr5fEgng9GIc0a/B/Kf+Pw1wf9IfgfxStA/JeBLpM3J3I5VYd6p/xJWyJP4vuo8ZcEJBMfcf4ZB1Z1UrgA3KbtVpaN+BS5ieOIHRzWVS3QOMoOL0mxtqSxhsEto5ffWpqJR7szKD/CoNzMV1ZKkZWXkV//x//5j81Pn8bx+uT+ZseeW7HG2BXFw9mp0+mvRpkwacGwnvqzGeuPmUvTQCifGR+7/47m9v0MkqeOQ+No88tnDjRT08g1BtTNBuoapxKgwW22T2zGn4inIixyssIMR1ji+I89KqMMhh3E5jQS8jPP4NXjgW9AseFRuqdOHm9OYGnitQGlR8qcSj6HwnGOgb+n8ezevavZxSl1+g97951j8WOiRUXi0g6U+ip+HIRPcIzvbXffzfG/082pE6eZ6CO3FUwURUzL6QNU4M5euoAyYXPzyCfva+64aWtz203bmr07ODmGk49szxo5pa1DK0wNf4f1z3e4zm1wXaP+z9GPncBXwPMvvAIZW3AOe6559nmsT2jHl/A34+rqJrbLbICfF/BD4kqpTl/loyumrvhqJeOkRjP2KESYtIhf6xFN5XWKOMmzW9500KsPKc3SPYnCazcTHZ03utq7Hf8k1hmVKk5Wxj19x1N14JX+YNz7r7+AOvXMU49YDSbeZspY2C4QeZpPZngQoSWltU5lspewFeNDS2C/2SgM4YJXVlXrSvyheVoIvEhNpv6W038nfYSTzjHNOsp2lD5VKyATL6i4RqMRS8kaLNJwhON4ID1ECElQYeWX+P7r2if0jKwjDfAzeXMyejMLEY8/1Dz60N1pp+nbCDdrKf9AofO0aZsJL8IF2uYuMnrQ77VR2htpjF/9T8aHfjA9sDpwhsOEFj70GmAU0XKPdE+aYK8f3wM8Ueup+yQ8n714zC3y3UwQQNrF9Btsu2C16nevnmz+nlN1BltxkSOCz5iUMnDLcuB1sGRG+BHIBA35n64kGIdh1pvKkkRJG/dk3neuJOoAdnH47piF1y7/brVy4cX2rEy1D3E84UTZ7x5HrFW0k3S3mtk2LJfuUAC3e+SY3A+IP7lLmlBbPwMYPda0pPditXHXzr85rkyKQUtV679LstZR2aIfJ2QdY/M4h72zfJw84Kk65CU1nzhaoVKjgUVJIMu08DuCf6dDKE9uvPEmLNYmw4crKJXW4yDWYnTu4ql2+lJya6QnnHl87zjhto3nnn2u+ff/+C7+rvY0T3z1qzj93o98Opfy8CAMYWgJm3LkRaWJNLsIq9+Ss1gZjyLfVdJOTuCnhGeVIPFTQjyV45anR6RPY/24yAKo+CewHnE7YfIPJ6xbWkVavlrI6RfNNppxA4oet/TUViz4pkIGmJa9l3LNI9mFuwUfXG551DfW2TOcGvSbXzc/+OH3mm/8zdebr37lKyiXJsgP43ARy1luXftn9hB6kulA5ucDlX9AdinXgGV44R0Etq/d15asHgwJtBQGKXoPw2jLI7TfB0D7793H9r78thr2NfIfxUkhBkpLZIuqPvMr6R3Kflj3PAwbPnVh3T0hq4J7H3pEdmm8dzG6ez+sex6GDZ+6sO6ekC4YXD4WdBqDby3+Lq9drpOEjx15alwjDOwAEYYFiU6wAJawZkJykonUr559DcXJK83Gib0MmjgtAu3HEo1eJ26u2gp03WI1ADtT+W/dUkkS/LxE1LadS1c86m1JTHTvxnBa4V1q/KaQkrb24lPoD7AKqxBjtJLbCOTH7QcBYXoEhD5O3BuowkVnd66aFQ5/K/9CEesAPw27QvIxgYKXIA1sKm7FD34DQqCEEodbTPzIHyQ06w2jd+28aru/WqudUTsrwgjNnxYHimX56MCklCryF+7kHWDQxjCSBORjAcUJzi8fe+QerE7ubPayEikd6aiQKJkjQC9Jkh9uKedr5T9tCD6uzL+DA5Un0v7tf/tJ8/JrrDAuoXDAmZ6rUcavFShoZDBf+YAW4tu5QxZxzJdmeVicHD4Us+URVj7nPf0kzj6ZDDNochUyNceBGI7PRvhL54QGewzTwjHuG7CSWM9K23r2leuA1forfPPoZRlkosLdrWX0XPCNVU5PdnCgQQeh13NCmjlXVDF/naNDuoLD2stMWhxALzqT5dKsf4mOy0mKgxn08MCvLlQlWxldMTgg7gjOYW++YZwVskeazz12P3Mmy1E+0G56dcd24p98kbdetW2D8uXdPxVVRgoP5bEPBNhxW8aua7sH+lW26vzgR89ySgSKE/KxRH7iQI8chycksx7a3PgPvgxz+Oh/OA2sc0yqnPRP4gDMy32oDnDEMoJG0n+ht4WRSHwjJn8CklZp5FF8wWyqwrsMfxsmbQW56AqcUNiFFBY/BbzAeFEZpRmzseoCCggoHjJX8kffLXbk+jgZY/X9059EcfK5jzc7UZzwERimd9A+xJonmSSvQ1kLvo+fZ7FWOgYCocncojDhWSdq8v7U6amsXNsOsvImvSZs4XZlUXBM3/IxKCtnAe37VfAXFQHKI7F7/A8YfoSRLPXKv4M7zEeVVmgDTrb6UVlLcUIbY0ClmW2qI/kLxmQm0AvVR4S/gH30+acqZJClxUnko+MsLU5QnNyI4iQT+eRBVoKfzCobZH6anU+97CaouNs+EhiekN5FQ8vMP/6NIH+sn8ph5dwsg+Z/+Ken8FvwHopnVl63YL3AdweyM0yqPQEnLRQAkY0ojxeWZlmJ49QrlP/333ErDm0n2cpznAHnOCdqYW2Gn5T1bJuhRw5Vk5hnu8VlbGyRQehp2vcM/fIo9RP/R8oxKq6DW+nUh80OTLQ1p3aR4SRHA8+xAugEXpLHqAv6JJlnIrABGnZw2o7OS2fpV89iqTaD2TmRIxNL3gEW2JsZK0yiNJnczhG6yPl3j5yM4kbFA8uSgS0BShllz2YmBru2jTd33LyT44hnOJlif3PP7Tcgk9hupHde0sXyj6Obqw77u0L+8KXCpLwnf9riiXwCJ0XAcclYnLz4Mv3YluYdtuo8i8WJdF7CkeHsFbbg5BQJnHNi7aGSyZOAzrcTP30BaKmnX5iNOJB1O1QcKTMhUFniKUHZysQE0bsnXGxGEbOJdnQOGMoXt+A4aTDdDrbZWC9niOdq8gRx9bPiIsPefVjq4FfgCluRJ3A2S8ExrkFxwoqvFieesDOLkmUbxxGPIXtszXOwRV5RrcIiRF76lFoBT7FTP+n/YIgLaaaSX/Y7MIG6Lo/bsR31N0p4FCVLrKy71TH9MGU4hpJshL8l+og5Jj2LTqCQdVoiZzuAihPqj8oUupS0oWoetnGwGiYNlEe1FcqN/ld64uME5/f33ba5+duvPNg89ql7AwNGZeJIlaFcLH8Tk4e0P8udv+Sjyp8PsqzwB3LVDdtYoonSBykQTv51aQ0kHuGpael/gSeIpOaraYxAvIJsWD0J1nbfBlcS2nXeCVRGJGoIKBiCtVVAMm2VLXj4N/rtK5yq8w/fRZO5nfHIeNvvmKmK23XwhVU0lJl4+Be+9Phvn2Veuri2v+o8/chX0lVK3r38ZBT+pNd0dfFhRfsTbsqNcZIy1W1qUqNiRKX8RZSLptqCVYFWJlqfiFPFgHLBMccoY8APi1/ahF+lWGWYLHXlT2D65F4M41eeeFqRf+Naccx/LE604gUCSyTBssQYmBzzDX9yKvla57AP3HUTbakWmNchrxYc05oWcI4FPSL4tVdfR3HyTnPHnXeXU1jaAoWLLPTkmtpC40k36j4mcKi6me2yjkXXI7+VKT//+c+a//iP/8jWlq8+/nhzF8ppHVKPY+GxFVng+Ef2u3hqnxIHrnxQaaLC26PPPSlMS7qciEP7jlIS+rIADS/cRqifEi1DNqs0AbYKEccGlpftL/7OYJAWJp6aFN9XyDn7Ey1kCq9tK9wMj60j2c6IfNOKRRr065S2QgWd5WS31189wFadnzRfe+Lx5rHHHm0mgRdfZOG8ZWx5WeOUW9wtuxVXV7bdfUVwXodhw6eV8RKyKrj3IQ3k/z/4kVdSvOLq5SchK98H0VcHrP4yiDx8IFJ1NqsZtSa+qwJdHbD6yxBt96RFgZMsK0uJBkIkxcQ+pJX73n7keysPElYVzvCq+Dnxw54VIVgmnh5HzFad51Cc/BrFyfheppz4s1CSW7N5swMdpXO0B/FLJz87/Ab5PKCP95ADjNxDJy/tZMeYQnIyKLz0IyHfrzwkbwSQOB0CEZxsKPRFJczsvUfwdPmXXvfYXcTc3Ma7HcXJoiMm4QYmz2mpQcSXwp8sGqM+GzmXJIQMfqr8l/O/g2KVVMGUYqCsxphRCUvFiN/0K+MHzZjVygcv+XESPme+iGsUzdxq9UdYcpL4xFnHvnW71VGUAEucsqDi5DOP3tXs3emgCtAINQWM5oCuyEuXf/6Erz7nS0tHCsSc1SA2cbsoybD5BTQ/UZz868+al14/jvDEbHgU5YkT8SjiqE+VwdSH6nSkuvKZjzhNdIXs0OHDODFkpVHtOUAdWKkQcgIKQLBTrphHU2A8Y4fuSgQD1VG07xsY2K7Hv8sm3secGIeP5q14GUVFKgVwoMsBNzWbP2A52MNpqp/XoYSx9nik8hUsZ65s2NJcwqfAFbyNEysDTwu8U5wsylMVKpieLNDJjaA0tJ572oWMzWoeyqzbbhxv/vbrjzZ/+ejHiEO5RctOvpyzO4IDqLk1TJ5YJimCNustOIIMt4ZbcLYQLgvHuoDyUr6rrjnw1nGcTeLj5MQF+InyjDYg71VqGt2E1qngEL0v4HV1oRvUTk2dzcrpBBYnCSV/g21fIUhA7RV65Zz4Q1VLl+Ei4lvXKA3n/3L8XTwLoSPMXF69/rWYq4zJiHWRIQagi59pjzAyllHgX4/Zfx1HTNtne8PDD97JVp3W4oQBjWmLMPlvjs3HNfAPo1gkkb+Wm+alXir8HKRQRaxu8Z+gj5Nqfah62mzGtNq2YoevgCL/hV8EAS0H6+pwdu/dXZZa396X/22CVfw3r338HWDzAw+VHyl/TKuRnQ6UPMbQK4PwNvmfAr+l+afIvyQ7tsqMmQLMYI6B3sZNo81+TtVpx1+UlzG5IKKqRbU7B7yDiyjX6v9dU1CMBZKCJq2RO5XACSW7YppnfvtW8zZWDsfPXWlOnr3SnMWSYYZVPZ0pqqjQ3FkHpFOc/HIFmbWHLSvbt9O2SbuP1cTbbmECxUB9Ew73fvCTF5pDODcdZaK/xCh7E/Lshn17sNIYxVkrvkSA6753ldSjkddLWIjMM/iezUk7e7Ae2YRvE1eDdQRrxpVPV+gr9dHkLn6PpFzPFp7dN+1DkbEljvymcByr0tXJr9s0dAKoP5T15NPtXds5Onf37u3xY/D2u++iHNJ6BRrTbzGJQK7II/flTzJwvuvO2zi95wTjiplm9tx7zRc/+wkcxD7IFlRkkos1yMuRrIqivCatW5H65dUyHPqBall37ZrXrv0ZxzapcvP4yanmeY4jXlzkOGJO1Xn2udfJB75KtFJFcaJCSWsSLU7GmPxoSTIVixO26rhVCh85l7AEcguRvgDknyuxnjIxdX6KbTsL4ZVboVRubMSKRIsTV1mtYHtwSnwOhYjls13/JNB2gfamEmacSYLm+Spm9uJQ1zRamWzBXN1V21n6wU34qfGUHpUr/un7QN86wllAX+FYLuMw8y9G71bDyEx4x9bZbB8wjMCq4lZ8+g+3/DDZElaWCqgvo2xNymINDlvH6OfG56gfTOAcbywib2dRnMyzWqwlJhIEoBCBIkVJZQF0xWGPFqWK5YSFSsn9oEqfq9LeyxNX3Ir8sTsmm7/76sNYDd5rd0UbUp2vRZNwgaUcNXMS6zdfBZEGX5+VW1EQ8FoZNbfkmYhpon7nImnqis+hN/yq56Dwh2/Lxt98sn6JPlwkX4V/8LFwtq+BLfDUU3PPYwEPfklX6lBFyae+i+abFw6cav7+H/6dMdF2dBybicd4Rnz2JSIewO7gAsTxXxEV+ohVbOJeCcwcMaBFXBKtIs3HXAOYFVtQucRr+8v4o/3Yxk1ZMhHWEbT1xy0clpfKd8vTrW0SsYUtbJalihVljfLYeMZRiZKrh78+tL/XwB9K36f9D4FX/qugzXfVkX7+HWeUMJHsGkdqWa9Vsvlf1DKL0hj6ONndPP5ljiO+R4sTysY5FhI0PGZQr2UQzRcl8jFOxPlVc/DgwebRhx9r7sZXldtkaiFxPlZoWm/YF6tw1Sok253kJ3JnnjHs8795vvmnf/4nZMxk88STT3Ac8X0oUzdlQVKybUeWQYYZjDOs/zpk1eJN+e8pazk5DJmlIkOliZZU8t/2otXdBWSaYwG3LKr0lUbhqkSPvym0l4ar4JlGKaYlTOQXlnBaxMWnieNf6VHTSd7d4ngBmeeCtnC3uu0H+VlHVtuHWGf003Ky+f3vXmy+/MUvNLfeuh/lOemBY36sf85bq91YQm2dsXCtkhZbd618776virg66SBq/8H6l3bSR9KLsBLfyvdB1NUBq78MIg8fiPTH4v9fAAAA//9YjmfGAABAAElEQVTUvYd3Zcd1p1u49yIDDXQOJMUoimIQs6hgypRN2bJsj71mvX/urfXWeuutmXljj2dsjRVGVhhJI48VKFKULTGKOXTuBhoZuLh43/erc4CL7qZEBfrNnG7cc07FvXft2rVr1646Iztcpb142hkphf/1Mmpk960N3bubs4muj0MBpmpfc/fHoJEy0oS30Ymokf8q9ZcBlQFEcK1gCVWRFMP4j+S9E6h3YTQf4SPeB97N2SFsm7v065Ztws5eXC4/fOql8sMfP1/GJo+V/sgYaYzvkG6b+jdDn+7A8iXj/voDItQakD6wJlUS5sm6qN6MlEP9uYe6CRuhAP86tt+AP4nObYe0Pua5CeYtAf2yVUoPbAjPNRiU9c2tsrq6UjqdTpk/eBCcB3vNuls/sFOo9LJgq/Sxw1/7HBwIHwDoSEP45paqgj8lJ6zWDvxSoVN625YrufgHOXaCHAHBr5a53d0p21S4A/AjItDfKV3w7vIv+JN4kHyUNxhNaYPBWun0l8qjD3+4fOLhj5SjhyYr4NTb7fTKzja4dqGXuAmAvz6IlzfBSgT488/f4NGkSTqea7vbliNlCwT/65e/V557+VzZ3JkE3snQJFyw06fAAWUAM1fFt0M+eYt/4ZVuWVtbL6+/8RZ8NVk6o2PhA0AlF7h1umUEekm7LiV0B1tlsL1O2Gbpbm2U7ka/jNOmYztbpTuyVXrc5YcecI3AUKIkLOLSgafT2sA1In+Hx/sps0vazjbx4LTdmyybYxNlc3SybIxPlY3eWNkKjRteoj12bAbu26UHYvzB19v0FSvv027iN5DeO+vl1lMHyp/98aPlk49+lLaWmPIBKQBse2cTHjSb4ea3P1G+QFNOl/C0i1mkZafmNbm8Ef4zaiTczkOvvPDq2fKt7z5b3jl3pYyMT1MsbU82+SeVhiokbdubu7UO7IDAZ9rLi5dLr9ctB2YPEGawwFo3MIiDRBYG6SwwvDc30hgl3E1YQkziu4hZcXMjq5e34JkH8lo+AbWuGpcqia/pav00AmRSXhEh7RLZlE//EGLfRuH7nf42uMEzZb08+LHbyuOfvKccnOuRpw/V4K6UIQDUzz9/g0dgqgWJ+nD9w/jbH5UnXp2R0UJ1BfYNPS9evFLW1zbgr9rGSSSKIkU9dvFarq/iL10r/oJFkvxdv37izM9fxd9XafKb0n9/WdswRAfays8ra6vQslfGJ8ZSZzgH2JRRH0T9tv8HgX/GGsiWxoFOsHrZ3twGr045dfxIuklLy9RvUvEEHvN2HQwMk0N8lP71ZnC9CEhexaBtSIr0a+Sf9Ya3aKfN7ZHy0isXyuvvLpcXXjtbXnnjfLm0tFr6MJAwdOXd7X4Z642Xzc1+mZufL0dOzMNfG+Xy+SvlyOx8ufO2I5S/GNHysxdPl4tXtsry5nqZmpstRxjnxkfHy9LCcrl0eZlyaU8YcxwZtw2/Kn/FZ3Jqohw5drBMTI6WhStXyoWL58sWNBmdnCrbyOCNjUGZGO2VzsZ6mR3vlUPQaXxmuqwtr5eLFy6Ujc3N9DPQKf3BNrKRMQwZO9rtlsNz8+Xo/EHk6KCcuXCuLC0zVvWsH/ygRI/6+8hx+92Bubly8MjBMjXRA7/TZfHSu+Xyu6+ULzz5yfIXX/hMOQmMg8FmeK7bHYVGyBQaYbctJHUaZD//1wYckj+7yZDjyMYzF66Up3/6Au07S1tcLj9++kUaeqKsMs6sbqyVyfHxMjY2Wq5Amx71Hpg7gJxcDQvNzc+U9f5W2UC3GB8fg7bQe2mpdKHX3IEDZWFhoWxvbZWZA3Nla2sTmq+ViYkJ+I10V5bCHkeOHC2LVxbLyspqmZtDN6HtV1fWyih1TjI2rvG8Ae2PHj0K/a6E3tPTMxnXQ/vxUWg2UdbX18vG2maZQ3Yrm/3nUKtk2lYgcUWWwp/KOOtR/vTlUYczScRzeM/+R7o+POxY5Bjcgy/HNjdKj7+RjZXS3Vwt4/31MgGPjkMDMwx6o2UF3DaREzvoio5TFjyAh3lIv0hQyreyOpapr8GM4R3hddzZgX/Ikv7QGSyXj9x2sPzZkw+XB+7/SOLT/xwXFa2kE9MBMOzKv+CJpKLCvfGn1iGe5Ko3n9ALfSd3fiXagELTVymnymkywa/JRZl2Z8APTrV+y+QivTimirwLyO5bkiSS4KQ1fUP7Wqd1GEA24pQ7xi+v98uzL5wv/+E/fZX2OkhfRs8D0OiRwGsNFd5K57bGyLOrxt+AQ7mmsZ19UDKpew74i75NmHG7uPncvAyPf8HVMvizT4f+wDtAr1Lf7lJZHyR2kH29Xg+Zsl1W6C+mn52ZSVv24SHbvAMPDJAhzkm6COe2/tDZOqw/cIFpU/Fu/UGjqZ80/M8l/nvt3wZyr0Wljpq4tr8RaeMh/B2XnXNYv9rfiGMjf4MRdCwrGjAuQsEe+thga6Hccdvh8rknHih3334S/cuBgKzoBgPkzSbye3RUPEfKq6++Xv7ha98oL770UvnsE0+WBx95tIyNjiIrBsie9choUZ6gP01NT5XeGPVR76a6L7r95uZa+dlzPy//5W/+phw6dKh87sk/KrfeeluZnJ5GFo0hE9B7gL2HTtuBibeQRaurq5FD0nwM2Xbw4CFwQGbTv9WTk552UpZvbmySfq1sIrumGA8mJ8ebdkEmyFPgbh5p3EdYXLmyHDk0hgyYR/6NAYNx8pbt6jjaRRZZ99IyaZGHjknT4Gb5yqte7WzwCTq64zNy7fU3XmV+83CZmZkgvbxk4/AH3FFvQ2G6p8IAeiWqCUsDy0DvdZFePLzq41DAXmATaZyP8h8PglFvBteLgP9d6qdt7B3AfRV99gcNvQ09tvj+xve2rPY+VND+oKG3oceh5L/WoyiLLreKt60FU4UOTUltVF5t6bZFFWS+JyxsICsQn2ELZqRjw2znMJz84McvlR889XwZnzqGyBhPh089mQij9DhR4F8E2VD91p3SLFYAElcHc+f+w21lPOI6gcFJaEQMeMOgxvBsigZrnhq0zZxMCDfqGGA4cZ6SK+Eqfxso/xhO6Fjzh1FOEEzD9Tep97LwtNfXKkRG+qQQaDCqT0P1i2Pao0mbKH5E3TG+DuLkJlEdLAk0znIpdhu4FR4qoQoZjQcdJ4mmkBYkjqAivsPEHWqCyxqZVsrHH76zPPbIHeXYoSneaRSU1U4HBYZ7ZFzoTzki1kobHq1/+BoOqs+13jYZ4q84FHzxK98rz7+E4WQbowlGh8GAYbPjoFIHFodR88hyEmqbOCe7YjsComvrm+W1NxvDCQI2AyYZ0OFQ1sAYQd5BePYcpPoozFso+n0Ggi0EKoraOHGj4NmRmXrSUSMLPCLNKCd8Y15olLIZ1DrShXfLdpDrSCPSqOANRjGajGs8mS7rDDSbKMmYZEIvy7UJsMpgy0KpC2Fksq6lpS01NFpxH2PXCBP1mzCc/MkfPVo+8eg90IUwjGLMGpJmB3wsQ2WhGpgqXQS6lt32SZuH9qOvxfhC+tpPGgThoy0eNXS+wCTsm997urx7brmMjMEDwO8EpScfAan93T4EpUSDMq2D8IZhvavojzJozszOEgOo1helh1pRgpK++TW+5m8feAuPGVMvQEsa39r6g3eAqGmu/q11DIfuhdQn8BF9S7azyFMG+J9OFmMwz1bRy+STp20G6JH18tB9t5ff+9Q95dCsfQc+UlCE4JZcr73aGvxSE+1CdFt/KmveB/KPtVHOCMYF9ELnAHY/JqEYTuDzUFrlPMhzg6bCLMvsXb40bdUEtrDsv/smdk3aXcAI/yX0rzQizV72ppb9t0TDCwMQ6AKzk+wVlC35YgJFS1yqAk1fA4Gwj5nE7XdUvyh5pdhr7oEwoYn/NfBXWU0/xogAwFHMBluMauMaTg7D48Sn8qa9eVXeW485uj5cfe3WT8RQvPbjXdZKh4MxIFaMQtsdR6nyLy+eKS+9eqH87KV3yutvXyxrzD+78JBgKOu3kXcaLw4dPFpuuOmGGE0uXrxQLp2/XA5MTpejGABXl0+XsYnJcuEyXDg2iZwt5fDxo0ymxzFsXCoLl1bgN8YBFNRxDQ5ooxtM0rcwsByYmy6nTh3HeDKG0eRyOXvubFldZ9KOHFR+aGwZYC3uADcjf7nxyKEyd3C+XMY4cPrdsxlHNRbUPrBTNjFOygM9lP9ZjCvzGE003L7z7ulyYXGBRYuj4AiS0MbJwwiTCsenGSZQx0+cALYRyn2LycdK2Vy9VFYuvlW+8LlPlT/7o8cxnMyha2xhZB9NE9p5JHc1fPrA26/Bf8rTPrCdu7BUnn5Ww8lMeePdBQwnL5VBd7ysYSBYY2KioWMUQ8jS0pUyyng6ixFkgQmCk/R5DFQuyqyvYziBDmMTo2V5aTkTxANMHC5rOGFCMYs83UTmr6+ul/HJCQxRGE6Wl8JJGkQWFzWcrFDeIfBgkkz7jILn1NRkJi4aRY4dO0rZS2V9A0PMjOMt9GZiY7pJ0mmgXcfQMzc7tysf1SOkk2NGRA/4RotSpiN/1JcUn/5JO5IyotWrg8F0oLGNMbPL3yjtNg78Y8AyurlcRjGaTMGfE4zHjtHKhc2xblnB2LMO721DQySsAKRtnJCHq6nL7mC/2l0EUE7D9JZiOYGZfuClHCrbq+W2mw6Uzz/xUHngY3fHqDji4liUKtLbYbyornnKa9OVebZM3kS07ePSxbBc5qp1C8WWY3jkc6VbLV65Z2KN7sQDVoqEsOpXZKfs5s7NS5mRLP4kMQ8+70aQTbAIq3/BPjRobELRBS16icWif3n+Qvl3f/XfysjoAYwMdYHMibRl7riqcxX/k63Wz313/KV0y06EsDQkUN93/GxhISaXHNPONYKeCZpMouRf6o9s5cUwkmzDMxqkMoa4OMW/UWUK4atLGE5IMzM7Q3N0MunOpBrZpDxwzOk6a841VB/vLYWMSkxTfwC/Cv82yrS7+JPJtkvmvaJNkuvaoKoz2bjGASBPSm/5RUOGMGE44aHbZazfWiwfvvVweRLDyUc1nDT68EgHwwnCeaOPfAU3dYVXXn2tfO2/fb289NLL5Y//+PPl4Uc+EfxPn7sYmkgD5c8E/WkUo4l19LMw1y8HpmcxsC6V1157vfz9l/6+3HDqxvL7n/1MOXzoCPIHOU8jbKHnCvMoi4Aj9DGNsJeRw1voJJPTk+UAxnjLX8dIYz8bj5GYBUNk1iry6AryZpv6rH+Kfq0OYP32DeuwbZVD5l+h7E3kQQ+8DiDvpjDeKCMHlJU+TJ4O+pZ8sUz7rzHOiN9BYNDAouFlC51+AgO04RqcWfVCZq6UN994tTzyyMMYY+AXOqP81batqlTtd+ojcuje1aapIUNvQ497qX/Dp7as9j5UzP6gobehx6Hkv9ljW1Z7Hyplf9DQW/NYDSdDGerjUMI27jpBiUr4UOTu4+7DbkPthew1Xlv8/vtwyibmOkG/af2ZfMNAEVxN8XaStop2GPC9vUybCVgkNykSgChiQIkyDKMrURQNAzrFeRT/f/rxy3idVI+T7RENJ6RwJGHiucNqf4QwExjrHq4/Q6nCmPCU2kS6Ym9Im5YX5IFCG0XRHkCEYFiik96akjAEs1krhObimXdRSFlksjYn7rGGJie4EK7hREur4XNYZnfovMmf31rQCNJ0ty6CBihx++q3nveovy1LOJTJtftWSMVFGHtoMd4tpLaNuBgg/MAJIbdBUMXXMkS2q5ACSQ0nWb2hAAcgJyx6oWS1AcPJyPYKFtk7ymMP3YnhZDrCWhx7TNRdObZOBWkHGjrJ9grN8sQPdQzjH7DMQ13DNDG5oljqaTj5OR4nfQ0nDOLb8JD0MlZjhMqaJVTlDZygr4ajaDekXWN17LW33o3S38nqAkY4mEs/mh4DbQ/4u6xmjiBAuyid/vUQrB2MJmPEjUHYbqO9OOD7rycPAFxPslJWB6kaHJzNauELNgzKPDNc80cZ5BH2PgroBp4aGxhONlgV3WKwEVdLkGL2D6lvKYxCFKVhCBypS7S2+BO7vgizqnDDyanyuc8+Uj7+0F0oOZTEwNdBYc6sGiNKx/LoY1HmKAtyUzp/wOYq/yheOGmTLG1JP6kJPLaVMFC79ZI0barh5FvfeaacPoeBEMOJK3a7g0rSA1/YLaVYEv9qOTGkgdnlhUUGxC6KOYYT6usBnsaX8DA4C6N5av3W25a5P9x4S681pZL8VFrV2PrbxInTcP9LJe/d/+Qj/w3kOf5qbWaizkZ5cuV7zAFYAmE4YbpYHrwPj5NP3YvHiZNB2z8tSzbacah+sgb/q/m/hdm7l/jDtMBhy9llVQzhv4ZuFy8uxUAoJeSZFn/5SPrXBiEu+EsxceGqBE+a5E0QTzRAQEuiWn9t0xoQcHisd3PW9E3y3frJuZsmqa+mP31HudlVHtH+Gk7ki4nxidSfFUkLJf6DqP+Dwt9mcoKmYdnJT4d+OMCIPs6E74YTB6PoB0HiHCflCVtKGtnCLozvo39i+LnOtZPJF/mSndyUFe9DxiD7ME4d5Xs/fKH87MV3y6tvLeAtsl42Y3zuxtjp6uvm2nI5dPhwOXrkFAbSrXL27GlWDfFg2gQKDD5XLr9Vzr79Qrn9w3fg+TBT5o6cLEduPMnkHKPJ+bPl8oUF+LGHLJlBGUW5Rw46Fio3Z1FsD87PMnmZLJcvX8Yr5Vwj25RB8A2rqUp0hFYm/8fxYJmbmSrLKLKXLi8yUd9kdRQZhWFnCzndw/iE1Rp8+6xizpXDeI/0QfKtt05jJMBggJeEsn+ZVdBpJvoQHpG8SdqDGAVYlMEo8u6Z8yj3izTPKt4qE2W6s1o+9dh95fGP31cOH3RBgD6LEcNxokPfHm4Lm0BSt2G/Wv4wsWCwjeHkJy+SbwqPk8XyFIaTHQwkGnjWNzGIxHCCx8mu4eQAHiIr6R8HoJ+rv+ssyozF42SsLLOS6gRj9sBsWUSeOgnRcOJq7zqrt+NMQPRO0QgirDGcLFxBP1nC4+RwWH95eTVeLtN4A61grNBwcvTYkRhl1jGWaJRyMrIO3cc1nOiZQpp1vIjm8dyRDpUSjnAyrSHIOJk44798DYXUuYhSrtt2Pig+R+gfo4xDGks68NsI9XSZHE2uaijBmMT4Nk7bTcCTE/Bhl461DV+v65nDhGwdo94W/OaKvPLDcVhcvRzH6mXfqK2VERZRTLKG9wBH6zO5XIhQft98arL8we89UO65984YxHujyFOTUFxnjAfSd6A7M6vKB4bb15LIhCSVrxHO0XVrquQHEGK3dfDlop8Ct2CaTpL5UiEVD8ccy6q0tcwmSe7tTzXKNPK3JkmTNMlTb8Q8GWodTV28O74JsTUof9Q5VjCc/HMMJ1/FpjkPfSaIo3WhvbAFz8Cfogl7P/KfjF7U5+JMvx0Og2wdfy1NOL2EqYW5abrKU4bz16LJIzoVZjOMtRpP9DJxvqLXYh++XVlhcYc00xhMXZDpw0N6EnSIl69N32Vivk//bsofrkdatuPPcLhl//r4h5rkrDja5m39toHtHvzoF2gVPMvzGtFMj2Ga91HmQ4P+YrkTw8kfPnF/uef2U8Q71shTpLGzwaKOK1Lr7XdOl//5Tz8sb7z2Znn88d8rt9x2Zzzbzl24jKFkvBxA3o5jFHe+pXyUhsI1phEFvWZtfRXD+OXy7W9/u9x8yy3lk489htGafkecOsmo+ibXNoytDFlD/iiPNARrrB5HDkvvDWSK9J9C71PHWEE2LWAA7mNc1yg/h/yynbzM7/ivsUOKbZF3GSOL8kcvF72IphlbpL9xfdpSTyONIXoXLqFLrGM06fE+i+FsinFhwBiygfF5G/w0Ass3W8ib5SWM+affLS+8+Hz5w88+UW65+SY8I/F6oX7bO31pqOGrPhcwjeUvLdYGXDcokUk6lH73cfdht7S9kPcurlY4nLIB4TpB/3/Wjw6EmpfO/54g/1LE3wufBl1uwymGn9sUsMm/dv0wmZDUq9avTmhYlD47GVdlnar4ZfJHsGyXf/R6FUhFYpRhBkxjW48Tt+r8CMPJ95/6OZ3sOFNNVhxN72CEgkQ3iuBw4HEAgZ/pTEP18279ivAwOfVVkUFgwmt8BFPTGZKBuAgk7/xzmE/hZnLUsjDh4C/iTMGkZ0YyoVTR8Vwly8VdwaDCp/uzq2AqXl6maOHzbohhFp/FbJ4tfx/UqaOmNke9rk//QMBPDAbQSFVPsOqwWOuKCLC+1IkiA2l9FrAOglbviwqBYlw6KLTB1ZUPh1Qs2CPby+WRBz9cPvnInWzVmSYcejj4QIe6IqEC0dSa8qzgKvwNammW2Poj/Vv8W9q4VUfDyXMvncUNGxc7BvFtlP4YMjLMy1dV0NZ2Z3WPwUL+lLAjGApWUbxfRanW4h2h7HYbFLIebTMOH/YwkIwgPHu4qfdQ0noMBipobsvRQOKfjreuZG27UoLWJX27Tlj412UGa3qNOATzH7o5YMkr/OkrYqhp+vDzFjhsMjHUcOJ2nepxIs38Iz2wo1pARwoDVzkjjJJaLYMQaNwHv20mBScPjaLo3R8vh8JarG6WXQYW117GpyfKCBNRV3U1puwwOYl7MvBEqZNOwUJmATfavCqAYkZ7QFRbxfaIsgf8L/ziYvn2d58p75xdLvjSox+jhECjGBJILbbp/zw0LZP8KcdGIu3CwqVMkF0x8IoLM3St5TBYJ1BiBsD6nt/r83+SJ36oflFqcgq/z7XFDDRk7/Kt5f7dXNDfvh4K5C420MhybRv4IfKEgGzVieGEbV6shj90/23l9z6Bx0mzVceBtlKS8gLI/vqFpMaHer5WGob+FVpr09hlm3RpRw0ndLvY6bJVJx4ncBBtZB3C2eakcP837SAeRhrSXlKmqs9teJXhQmXqNv37oT9lkSl0aorfy23plljxVMarULm6Y/dZXnUyOJoVqdA4rULygNpAAv6/fPx5//U34HH73eIvvsFZ+Q+OaSeU9jEm/TecOFLlkwkayrb4SWchqR4nLdVM116V/uEhglKEw6OkyR/0pB9HXtt3kI0bKIt/9V++VZ76ySvl8hq8wSryCMbaTeDZwrihy7b98OTxE2zP2C6nMZosry4wScZwgJV2BL7ewNvk3beeLx+6+eYyd/iOcuqWD5dNir/E1o9LZzFqO6llAuskS+R0jd5gIn4MY8yHbrop+F+8eK6cP38OmFW0XRG07fUIGUdGM0HGE+/UyZNsPZlBPlzG0MBWHoSdkyA0XXqQfZHtJRgPeqM75cjxQ9lys4kiffr02bK0iPFhCnmEUVoXfScEmyrOyOyjRw6XE0ePYMDYxmhytixgwBljEry1vlTu+vAN5Wa2Jn3srg+Vu24/jtHGVV3wGUXG0slUxkPnNMFV/A9EwuQl+f1tODuhbZiGk/N4nPwYj5OdMlveeOcShpOXMTpXj5NVPE7cBuMk3S04PQwqB/AyWWCrjjJxbm6mbOBxsY6Bawy8XC3VBb1L2znZyFadGE4OkE7PFLa3MMbouWV5QngMj5OFRQ0neJwcmI8ccdtO9SRhNRhjhRMkjUvm2cATZsbJBo2kXuM2okm8WNZoqw1oPofhZE//kw6V3+Rf8c42Z2rWsJwhOuOLsY6LTIocR2mPGEbAf5QxeMAEqgdPTlLfJNulxhnfWCMu44zFY/CA22P7GGTW2QZ2mcne6uQM46cTK+QrcdVwArYOmY3stFVitKHmwIZ8rHoZcNKRXFSw1+1Q/g4LQzcdnyyf+cS95Z67b6MQ+iKdUd3GzWHaghyr5F9X1hHE9FV0JHjUcXAEXhnQ2R1j3cKqZ1PGW/pEW1cdG4FEIuUHeLmHZo4rwkmbC631s6wQXAwnIPHJ5ptB3vnLIz/RrLlDjt0wistz9L2kN4dkQpa26XhH62DrxnZ5FsPJf/jrr2AgYqsOfTq6OFQOnTKW1Tr32r+pi/LC/+DR1mk95qsXgAkbfTIANaHGOkI6CRdjknARKv4tfCnCcvfkf7xXNHLSJtXjhHaiDCfsbkXUm8FLw4nxmYw3aeVr/5xoW3SVLrX3JpMVtwROgD/763//+Fd0K/9ZTJAZKrXWb1+REl5In+iUUhQplLCdndG0xWinGk7uYAvl537//rpVJ4YT4GuOObAP6CVtncvI4tdeeb28+dY75fbb7wDn0fLOO+9iTDgQr18NEDEioBdreJVXx9map0ff6bffpRwWNaDlP/7j98qHPnRL+fijj0S+ujWqbrlBvyTPyso63q+XgBWDOVv5NK7K2xvNFru0DeN7j/ZxjFjAA07jlVsK9WxR/rWk2R3/zY9sWNEYw1YeKTXLlpvJqSl0Wry1kUWbGlxpV2WebbKCh8wiWxQN09NkgrSOdTHmUL9Mn3anbnnkwkW2sL7wIjL5B+Xf/sVflE9/+pN44WGsl4hctkpaLHzt+FzbyLDKq6aqVxvW3g1tn9t7k/Q6t+EUw89t0sp/oJjr+ilaXtuDrU3X3tvSrr0Ppxh+blP+ZvUPeZw0heY2XMFwuJUNx7WVv//7bm4e7MN7jTRcz26qvfp2g3Yf3n+lQymTG06Wma0/Rg/fW7xqAnK0wNX66K9h4CrGjbaMiA1SktYCudyTrzDPVh1WYH7wI7bqTHPGyaAKa1ftRjScMHB6dXQTbOuXewJXrbMShzoIVwmvMDWCsCZNejtyu4/QNMHG+AhK3lIFAcJsKcJNMYREuPucgiI06Fjma8qxg7tXTyVvnj19AxSE4Gv5JuMvYKcxLb/WY4mpoBYlYVJuXpt870V/4da7Rm+IVAACGkJEhN9qEPE1ACBIERp9gBiwCpr6+e2S3gFeWJkjMqAIKHQnHeuS4I9IdxuCW3UevIPtOm7VwYofA5iTfc0DlUgKmz2PE2sQF+sXyTzu1tu+DUU0aR0uNJeV8l81nLxYzzgpPYwnGf016jDIqnmrpPiWmxMH24tabUjC1/BLf+3t0xGq2utHcP0dYXVLpcxtOD3c/kZQ1PQw0ZjSJX/db+0klUGYAUmGigswNLNO3Ytx7qh33wEjW3mESaKHLlaPAi48hqGYbSPI11FENvDUyB+Gky0GblcUqKQaaWgfzwByyFPj9F9wppCsTEtL/g22yQPsJw6Pls986r7ysXtuJV91S3SivkkJnXHULhQ4Fewu+7/d999DmUYLZgJFe6v8WY8luhLoBT3TP/Ji7cJTDScdPIte/MX58vXG42Skx+ouhhlz1smerSYDkUGkNb41vC5PSQbr0rVczwJXSH1X8VFp0tuLEY0g84olt5rl15I/oF5z0ln38V8DlspX22+tP5cVtRU2/c9Ko9BpZYQu1WAK7fkn2yln5DddYlXCdPWW8g/e23qckMdJGP0jdVZoqGa4suvUL3ykMiYtxEPOHDLExsgKqwO4pN4p8TiBh01vXOQvdWgIs6TAaqRZaZeKfxLzI7FIVaus9wZ/s7T1CxJiJmXstowJwH83fyCA4hQZHN+T/jWLqAyi+EI/iFkncpxxwiQ65QKr9Vv+B1F/4P4A8I/IkP+V/+BIF6xnnDBZv+HEYdrDuCDGjQdfQ9uKp5LLyyQG+2B8nn1vYwxHrtjX6uTUngjl5UX7MW+b7LH7P//vvy3/4wfPwZkz5cDBk/T9qbKGItnHWDw7O10+dMNNTJS3yrkzF/AoUYaQv8MEeXWznDx8qJw6OsPqIIr21HQZmyL/+MHy8ptvMHlnYo9xITBkkuj2A1YsEdwzKMJHjx0u05OzeEcslTNnz8RN2kmNCmxfzz6U2KwoYniZxuhx4403lSWMA+fY+uO5GiNMQu1jW2wbmcSwMEX65aWFMjU7Xk6cPC7zlNNnOKcEg8A43gcah9cx/nQwLOq1tLx8ucyz2njqxLGQzLTLnvmBPLRttjaWy8fuvqXcc+epcuctx8uHTs5hiEBOym/2cvsueCnOwn9pAVtMytIauTXPhDQBaXMVePuWHOwZJ2fP79+qE48TxoI1jAXZqsMKp0ZDzxehB8RwchmPE8ftObxKPONkU4MIBgzPJVlexoAEY8XAgseJ+/ldWd3CwKKr/EQ8TuqWHkHNGSdMVJwozLFVx76n27vtka06TGT0JDnG9qulxeUYTqbxOJHGbgMcZSKlB49pVkk7z7kye/jTS8FXXldiyv+R/44MdmDkQM61Y1xUZ9AY14X+o8A8yRkmUxjqPNeEWRTjKgYT+MHtsV10kFHHUGDQQ9RGwI+prDKGLTDZW51g8QEj4ABvJ6k9CrM01ZHUXrQn/9OApgJG29MrOo9AU67bNpTfepz8/sfvLR+582Y5ANnuKjy8AlIMUzECuiVXbyllaSd8r5EEWU+CTusdII8Bgt6BnRhafEE3g96OmZVYjvv2VcrSghEmk/cCnZnBB3ryKz+1+nfyNrhkGxRx9VXOrO2QLb0t/oZRZOXbShXfDBV9JYUiX5xWkAk/fe5i+fd//SX6yTyTTz2wyEkbpv6UIoSVlruwWIHlGA5Ktf0r/UVLfvMyWbzBrThvNUIU7OtBRWAoy0uyWJ43J+kWVPGodHKRKQuY0N92sp54QNOeTqDF0a06trXbNMShepzQhuDlImeutsrc64t1+mTltWl4G6o/QLV4tAnNcRX+Bg3jn7GxZq6FW4fZrCSDte3i3AHep7BttssnCrlg5+pw5uNO/0r58C1s1fmshpNTtKG6Fn946++4uqZ6iCHEbSvKUY3Lr77yBucgHcBzr9C/1/EydPukWwTpQxBfw4Qt2HOBDbK47e/8u2diXD+LZ+F3vv3tcuL48fL4Zx7HwHoi7SX/q08o990aEyMrffIgHveeVaI88gynbLlBT7IJlR96j0h/DbJ6j3heiXwmHDKj9avb9BkHNJqsIx/0NJl2Kw953GrqeSnbbsOkfRxTLdvzVtxO6MskZ7ZowJG2esEoSy3DfrjJlnxlWZ9y9cJ77uf/XP77t75Z/o9/+5flTz7/+WxTbJol5cuSLf/bK1NoGGSXSwhrnneDdh8s6te+dnPzUNu/LaKJyW03FZHD4aYdjmvzvv/7bm4efpv6hwwn+yvfraANvqqiNCyE96pph3K0j7krEuzqw1ebYDhs//M1KQgYRvS3qT+ikMKsI8zSPOVNQIlzkjcMs7JE4ZgYKq/wtWlUSzSG1NVRReu5yysYTV7knJPnUMyOIwLo/AofFUysqyo5wSHGgApJ/Q0UApFO45vXXv3UTUI7tp1KSFxRt9g6ASexBdEPqscEk1sTmtZC8sggn1HFcukyFujFTYFtZju4AQoNV7t6CKu5w6zqIAhEJKsvEbCmcgB0Um7nM55KfIywlZJVbTNKAJyQx10zdewmTF4pbLI+k96dHn+GSiOtCM0A7nw4Ww6o30meB3OBENXVvBoZsppHOkGKESbAkAJa6IbqGR07GhzwOHn0IbbqcECsZ5xkyoziXhVcELBa81BmVl6AR6Ba+vtS8Zcm4N3E1wYhIfAZ7yV8LqZ+8av/ozz/4nncy92qg+HEtkysA4VEc0WGfJKGrNvQK4eeOmAQuLqK4eTNM6zodVm9Ig5BOcbe6TEE/bhbc3QTRlnroKzpRugAmwFapU+aO4Ch0Kn8qRhlixY4x9NElBtFEC4IGuIgZXUWidVcGOUbwvrkX0VZ3uRQ2PVR3IwxnGxgjBhoLLBMlURwICV/lAP80hNi5t2Y3T3crAZ7BsuJIxMYTu4p991zC+lQKx3IwGEdJYKFBehFLsqRV0egiZOLHocvdvBMGWEAGpvB5Vo2Bk+QAWhf/JPO1A0+PgMet1E8Ts6Xb373J5xxgmcVOAyoxNVj4bYJvEI2XqQJIOc9D8JP5JXFS8DBSjcDpsXqDu8hb+Kc81koa/eSntDPdMJRy8lL80McZUomr5THT5U/PPhfnkl85a+sxBC2l4FnCdXwXxOTepPPQ4GCHWkoMKXAK+Ipn3rQWFYxm606D7FV59OfvrccOiAPunIErRsA5Y73W3/ga/C3x7ZyxgPgdHBR/4b9WM3mjJMYToA8fF/xF7pKfPFvYa/yb3//oyaSXo3/cP3mvvqqdKI3DtHfNFZ1Lf3b+sVf2tV20923g1u7E3DPh1LBcqU85Ugz0gJx2jWBQz+/bf0pSrBk3N8h/nZ5t2BqcJav4lGD0j6OIfMkE9MuBtjMVOQk67fPhv8EA36qiAW8MHD4vnm9+ka3dXhKEmSUckrDSd2eh+F4c1D+r//n78tTz/6CA7ZZpZvk8FMQ9nyFWTwKThw/iYI5VU7jzn0Bg8Uoh6b2xraYZKwikwbltptvLB+97RSHx66hTM7njKPnfvFOWWB1cRQ5tsUKtWPcOPm2kaN6NLh6+KHbbisTeEecPXOuXDi/AFzINiaNormD0Vqi2+cde2688VSZY/XzEp4mZ/Q0IZHGXrto+hpoaTyxnx2aP1BuvOEUMmSRcs8xAUA+Kz8g5CY4dVCMc1gthoUbMJgc5pDVrG5evoTRgL301O+BkHoDbK0tchjoDRzmfGu548Yj5aYTs3jfTDBZRE5JSsdK+E85aDPxwl99at8Mc9QebrLwP7xrLlP3+3U78tPPPkf9B8pbpy/FwzYeJxiw1jz8FLg9X8D9/p5x4tklC27V4Z8HIK7rScLhya4Ea2xqt+oc4CwU6daecdJu1RnDyBHPFAxF9rVjbMHxjBPPNTnEKqz6TwyV9LcpzjJZ5UBdzy45iseJ23vUZWbcqsNkbHN9K5MVt+rEFZ8Jx9w8HicSIVjSmrWx8i7ZooIAu96hkc+0ecezYxiXxjH4e/7XGO+TGIMmVy6XSWdz8i68NArx3doqz3ieWRYmrIdxbIOxbY0zdhYxnKzh3r/ZmWChrYthBbd+IFJS6D3rduS0lUBmHK2v8p+9rI4JthGdn6TbbKUb2Vkpt90wXZ7AW/AjH74Z/lM2EUkm6RAZnr6r/taMOdDRlXdFVQz/eaBI+LfK/Kqt5Fwja9a4otzmeZx+lwkfixg8oLIBPRPCbB9CL8jCRuQ54Zbr1dzCl02Y/T+8xrscpwbp3Nk3+1m7/Vz9z3+VYxV6YE8SJ4ROWLeJX17fKf/MQtW//6svM0YfAgaM2ALLj9Qa5n9Dm5oT45uw1PZv0lJ+pbWxtXYbVChzBUTS0jEr9GmRlGc/Cj4Nb7X1m8XLYL0YpGH1ONGg5dZZjLOEr7JVhwLY0jEbA4nGWhDCQKkxzD4On9kY8oqMEfrUWmoN/BL2K/UPKhFbL2Fr8Q/9Cbga//ca/6Ud3NvgDt9rDMlcAVlo2TuOi+qIGKfZqvNhD4f9/QfxlDsBz1d9uN2qg3NK5HHl71555+13ys9+9hwyewJD7Hw5jtzfhoB+2EH8c3g2/K5RQj13DTl+Gbni2XVjeFI9/8IL5Utf+lK2/H3hC18od975kSCr8UNPvkUWwzbZ5qV3zwHGlQ7lbKOf6vkjw7vtx63ybuHU43gLfXsKTxDHoB5pW/rnbBOw9FKWrSIflP3iP0Ofd+uP/SoeQ9QbvRb5b10ein0ZGdejL81j8Ff+qV9Yl15i9kHPY7HJlX0rbnfkRe/C5/75p+VLX/5i+bM//dPyl//mzxkHMBhyOQ+0/dMFqFekNWYOX8LWtn/CbXMC2rCIsualph3K0T7mXvtHm6/W0SYYrnH/8zUpCPhfqf5dw8k16A1TZj9Oebsm/b40LdrtvYlsX9t7WKeS9JryPsD6kT8wqjXuApLniJe2hTNwICZkLGAx2NT+5p8DEG8yYNjJQklVt+rocbKK0eQFDod9AXfbejgscgBjgOUxSOWrOpSoUcCbP+RPaRbc1E/PYJBgAKBDeemG6qWlOZNrc2ANTTqKyMqc8DqwpmuAQx1PgBsYnQ1atoOG6ZBsjpvWv61HA/gGEu4K7lWsncu4L+sKd+ToQTqzsZVyymYFaL2SK3H13XDDvMArMFmP9RvjjwMXwBG31ymIDy3QnkfZxkFKV7fq0IkA5F8POKFKkvWhpW6STr4V0qEgRXec5Vun9fHf+mJIwhrvl4xCmxhOVsojD9/OGSd3MGFHgNlIDj4YaaSv9LL29jeFgZEUiLWWJC2WQTU/phdCAOGpvTSQbIHuF79Uv6qzhRdSDCfRyBgkSJ/tMQyEmXRA4Hrwm3H8syH5W9Xj5M23OOQQbwus7F22A0xgbZ5AeRvHIq+7MNDTzNDQsgBAFpAePowwoUtdUDd0lLcQvniL58+8I7xnYpq2qBjEwJF3sQdWaLtJW6xzSv0aK75brJZtoQBuMshDQapDoFOmxqgR3FmEI5QM/YWIdrO8tA93iONBtieOjJYnPv0xvBxuh+/ZTiWfMJD0ubNwwAX04KK3lka1bMVhYOEx7dXD7VlFbYQDW9zmo2GlQx8ZwQV6x/4jX8D0kmOHCfuLHDL5ze89yxknfJkDBV8DmXtqR+1D8o8JufnoQ4KCiQfeWs5O9uT3WH0/MMNqAKh5jo5fyCBneD55rc9OE7ZITMoz0XD/Dz3oo9K3lT/WPHy1/Cdo114GVvoGcN4CusEqs/YneNF/uSgk5+iQxfLsR6PWDQw7A/iLLx1d/VUd+2EOFKOca0G4lv/b+lv85S2kWtoxMNDfnPjQzHY/JpCLKCS474ZwVfESZ//V9MP9z9IbXBL5y+onAeVbbC0rD6H/Lj3kyaSR/qTntd5aTMV5uP5UWmuWlhgXVN5cOXJlqT0c1nKq/KV17NsJ+N3WXyH53ePv8BYezQPwO2ggR9wecu3hsNQPnzlBlg9btFt0d+97DzwJs1SmHoc6GTHyz9zID+slPhNNJgj/+e+/y6GwV3K2xpVVFwfcCjJXjh87nn5z8dxCjBBdto4ws6W/s23Cr4sxpt1+403lvo/eiDLs4bbT5WvfeLq8eW4xW2JU8lVM4+GH8NpGyR9HyT7OquQhDne9yOGyZ9ka4/YRJ/tdVjbto3pHeLnFzdXGYwePonSvlnffPc0Yuh6jqu7cekU6dpve/n0IL86TlL3CyubliwvxnlAm5qwIytIxTAoI08EDB/OlHQbhcukSZ5qgJIcmkG2LeN3CN1cXy4duPFhu4kDYR++/vdx1x40YKTAsiwtjd5f2gpyUL8UtmcxhwYafCckF/ZV9tooQ1N8milufNrhwabk88+yLyLl6OKweJwPkp6uxa3iqjnPgq9tw2sNh/eLYZT16yOtXdTYwLFTDiQYWviqjxwm0jsfJJQ6HZVyawTNlC8+h3a/quKWHMwSE/TgGkQU8U9rDYWWZ1sNLTyINXq7WtofDbuBlMn0Awwl8u85kx4mH6XKOG1ugDuBxIp4OWsr/9sBExUGllhM/aAEvjuCCpIenixR6lozi8ZkFDHCa2MLjhLYf8wtwDeU8OyzbcpjTxVORiqJ3EY+fUlnDsLrMhGuDsavPWLXTx6uS2vTfcGGrtoVQ1DZJyzXto4HAMo0znTyu/rGFEWeE8fP2ExPlDzB6333nrXW8RcBm8iojkBLAuNv+9TV6jx6+je6QbkwUkOTXPBn/1U2lDUkdMyzKPBAHGNAiHPyFJX9GIhMwqIz18KRSDzUcgZ8zd4TbMdsBwHykJYgyrBN4+adpInUQpy6nYcswYak6LqkEBqCkHaNLxvKl9e3y0+dP81Wdr7Cw4jl9LGICo8Zuz/nKeEgtFi6U/jj22T93jeXCZ3gTb70ZnxNQwzNWJHuFn0KISIlAQ7G+SuSm/4V4tcZgkYoTrWcGba/OQ1up/7df1dHTTZp7DpDxej+kzQEu8wFxh47ST7qlzPaXfO34axtVqCt8UtG0eeOnxf+31T9oJcqFxpbugpyyGHw8w8f6q+FEXbJ6nNyJ4eRJzji5+9aTSuKkQRMDHkqANXai03YwrC6WZ5/5aXn+uRfKnXd9tNx88+3Zur6FbJEeeqWED203QNAI69e9lvDEOI5sthGe+vFT5Stf+WoMJ3/6Z39a7rrrbtJVj48zp89lsWMKA5Xb+3LoKnI38y5bB9mhp4fGXr+oJu3dOu9BsHqc2H59+p80HSUdCTDkruB9WD0aNShrBJnAUKqc6etlIs/B7zG6IM+Va5f9IhkycR5PZs+CyrgK/fSGsf0jo8DRg2s9/8QFmhkMzJ7T+PqrL5cfPfXD8lm8aR55+KF4q5CBOKCi/eVhHoBRKG2j+pwAmayJy/tVP420uSq0fa28ZPta8u7Vvrb3ofhryvtfvH7GACAcwk0kd/Hy2eiGgMPhu8RoHq6JuyrgqtfkSth1IoaDPoj6LdMGTWM1lck0vnvt1g9d9gSItDBV7cwKXnMoGJMXbkysCiP0OneBzxE/81L5vlt18DhxZ6sGCwcLt4zsrsDXZQxkwvXrtwplnKcmK4C0ciqAPWHbwYOuhjEGActkzwG2i+BQNuvUkk9AktcOkLMEhD1tbSAX+bMgT5iw98FF7GQJO6UrWH7ebwVh7eFhRw/Xg+iEx/FAtrATWqYlRlT7gMxKVQakKusX0nqZX6tqHfwII78CIPQnDhWPd0YXtA2t754w79YPT+DWMqq3jGOr6XRH3eHLMH7S1isDJ4B10Djl2zQ14R64JXYO4k64qRLlhEl5f6U89nE+R/zQh8sRPkccBZ120mqvpVgtPgORg5GDLLVaqLUF/7zW93h1GGjhktosAcuaTQONiapnnLBVB4+TfI4YrMSeVk39DvspIvkJl2CWizFIAecXBl7FcDKH4WSUFe0urtCe0j+Fd0ZPjxNoJzlcbZErY8SgfBU1yw2fQFfdJj1IVkXc1TLdh8dIYC2eX2JureAKc99U4oUl/Yd0grQFTdZHNJxMl34MJxhQ4BXPK3EbTr6+IxzwfAwx0pA/uaVSjZIBTLhcUhpB6Tx2aLz8weP350DSbT/fCCwDeFzDky73Ufrtd5Szg+dEcLQA6lEVFDCVGq3uHQ6v7KCMegaCHinMHKyaGwYUDDzs/Skvvna+fOef/oXPES9ju6mGE3lELyMqJTmtzqNQWjf/U4bw51wD7u5tHWV1LW6UdMDKKeANA7h6VFtf4NqrCbMsyrb/VXpYT4LyE/ljnPWLs5W3eQwb6n+81jjJexX/tfXbP1KXhkWfZNAgxKurZrxKwzH43bbfoT1GUe0feMAzTvA4GToctvav1BqQrFbQKPY96ze6XvIlMow6TN8eDms3s9td4nDtNVaIjdZVvMVf4ReYSSf/VfylnTTnXQB8uC7+RlZ4eahkbIIMDez8pp1NkPKH2r+hf5qAuL36a95UC7/ArYAgfw/qVz5YAY/SY8kKBYsmc6r+Hdf/QeHPsNLQpcoWF/X8XPUYfewkX9XJKnMICP1aYoKqPGuwY5LBw1dQz89wKE2nvZfEiokYXxFm0tIAy9tETvzo2VfK22dXOYD9+XL+MgfxTc9x7sfRePdcwMNjdZHP3sIcHfr3llslpv2SCt4ZeHPc99G7OD/pNlbkOUiQ7X5f/fqPOGCWUQaF08OuVdL7rghSvx4KRw4dBM9eWcFbYoVFBF2h61CeEUXIguAkK5/urddVfIDXysqVBbxI+AIMSrwGXJu+zzjkvx5eFrqYT+OVohv+hTN8GQi5bvvJPeoXfeS4clcDhAr20YPHywBPiVW29rjyvNXAodeL2zgGTOZ7rNwePTyF58NKeeJTD+BN+REMLhwoS3nS30OL2+6+S3/gGm4bcfPddtyv/+ylY8GTrxHxOeLmqzqv81Wdp5uv6qy6VYcJwsQUW3UwLOUzw1i83Z6zcKU944StOhgd8jli3M+dhHjGiZOEWejiaq/GJSctTnp0hfewx9bjRAA1iLSGE7fqyGgxnCDbp0jr1yf8dLRnnHhArQYY21ODjFt1NDRN8ylPzz7RG2WWM1jktPR/kbePKv+lAy3hcODB2G696TIpcpFiFAPMhEYSyhjnbxTDyThnmUxgSOkxSXQFvCoC0s42hWcIyqWMIn4b2mgwWWY8X2dytI3HyYhjPWmrPkdO2qh6pgoL73YQy4Jh5BHHIS/zeC4JgcTRSOg3t3HGyWcxnNx31201LEJToz4lJRuTWd+EKwJEHuFPnYxyNFAY55dN2vHPd1VZjd1ZtKKEeN05YWU8t5+qhOQeeK2owj2i26haBvplu7oeQ4rC30IZRzW4kTx5alxdIAnapAmYFBljLkjUDxKYwT/wslMKHzhcwePkJy+8U/7j33wVPfYQow6GKRCTVj31b4jrP+noVSG1bP6oSP1H+otL6GQa4oQlcCSPrdJcPti25KlFGtDSuoa1JbV9sX23XBcDXbR08aevYYQsXXUqJ9MxnPAJcvqIW4M1nOQaAiafI9aAEJiNrVBbzt7VhDXItvUb3+Jvxfv7f03VljFU5V4eI62owT/8T1vLa27xjccJkW5UMyyfI6ax8jni/kK+qvOHfFXnHr6qY1o6K7TgS1MQxoWo+ln7kfLaG2+zzeY7fFXnF+Uzn3mi3H3v/ZTlQhYwQju3Enfho01kjAYFDbk2lsbSw5zVqMfe93/w/fKVL325nMLb7y85B+T2O+6IUdbzqdY4iNXPp89MHciZVM5D7GPKgsxdoG0f48USxg3PI/HzxLN4m7jdBsDRSyv96zyH7TbUr5FFjxMXv6dJ6+HZ6onZyiMjQVBpqgeJXnB+ylhDjl7MblOsfQnC2tcBxMUXD9i2/iuMB8LmweV+onibvAsY1197/VUOCX8sW0zVT20a29dGkt9992fY48T4hHP/IObfVumVeoYrq8E1vH2GLvZTr+skbVJdJ+6qxFe97pV3nYjhoPfCHx207frXA21fESRoydnAOxzdBLUo1Ch+I2F2I+tDwyRXhfJ6dYHD78PPTc7rBLVl1Ch+r1N/lHW5jGgxyiDkS0ihcNgVj8SSIolqiXmU5ZR4MHAGKhNkMEL48ug5J2f5KsQPn2arDqf/T8wcY4LpVh07NfGedOF5G7nIA4yeG3F1/UIhSytgBigRDmTuAc7KhTMKmQrhus1A5SpVtnEjaHoMcHodjDoDSWePiAopnBzVyVytPZMQqhEPBxMnoVZrp1QZdsVGK7eC3D3FduSmNNJJDS5vgspDDhu1rLwBH2kiwBWk/BMb6/cAsrQDCU3r4GSTAAAB1gC+wLoNvpt4DGg4GbDXcYQtAhp7FJKyrivlO27TYZAWXoVB1mkyGPKukCGNwslSNW5km4d14d3gV3UeeewjfMHl9nJsHjdTytToAEEQjA6uwIYuoMhp/0knk+SH+LYLVetvsEldmeTutrlNoamilL/9Mlt1XtZ9e+irOsCYPaDUz3Ce/NUbAxxAWPbwjJMR9j67+vY6+/EPoFyMa1hYuVJmUFYndR0eYAwyN7TwpG2NJhAXIOExlIQc+grtbAeG4gxmA9pURVDDyTgI620Aw4GXbc0jZYU3bCDiLDcXN91010fYqtNjZdAv6zB56KMYUVrwyYF5to+GMGAAkvSDEFVcwbsv43J5Lg2+5uUEhpPPPq7Hya18QIJzaJxgyy9gRm7+SG9eJ//ula2lgiew2dbO8nh2dSPoM3EdMDnxs3aG65kyyuSm47kTfA76JQwn3/3hz8qZS+wZ5R2LC3XQ4vGuAuJUC03BvXKwz0AhGDSM3KDhxEFzhoMgk9d6NEoBinpsJZn5K/+3gwGJCSH8Ov0/ma28qd8H/xkuFeoP9xqUuHqQprSyVFIBaPpfMgC9d4NjPJSra1m6cNsWDs7m7KHUuu2gsErfHfGrOreWxz+p4YTVFCZoGWglgrzSXMImNcQ/9YtTw/+pXliMkRjSx37kn3Slf+tq7MSOoHyOeA3Zk7a1A4aAbV0kCL2ad2FI2eQFBwliffk1XPYPAEP1N3BbQhO1W0eTqokhRRLVEo0zR6gm/l41iBs0g1ezoky/sP09c0FjdyYCxIMesEBh++Re1mvqSPkNFrX891f/B4V/+pTtJOzAla/qsGWmbtUZ/hyxvQHcbZ+0P3jW/8E3PzZGQ/+9QJ8qjoglm3Ov7KxMFAAAQABJREFUDNsUWZS+B99iO8BL7Fx54/RK+Z8/+nlZWAUOFNxxvN2uLC2W8+fORdkds99jWOXIWM4LcSsPW3Fgho+gIN+KV8b5079gz/pUefpf3iyLa8gh25G+71q/n6tVWT3GAbOTk2N83vdSucghs5hdUfJpZQzUWZFWRoKLho35uYP5VOU2dLl08WI+Xazx1XEhk1Dw69N3BlinZ/HenDt0mC/QbLOd6HJZXwEpxu2Ohl3q32DyvbnD9g9W6Oc5e+swaQd45C3i8eK5IJqI/Rz0JnqBBx6OIsv8pOWJY7N8sreUy+deK5/FcPLEpx/MV3Ucc/V8EsWM+WHEynb2GGnvJMmQ/JP+/EsKM9lmiTEcMY1ocDvd0//8PONa+znil5nAjrKVCuMO53y4H9/PeS4zPnX1CGILziKfI3YVfZ7tSTn0FcO/Bym6Sut2Gre1zZFOl3onJhqiXV11IiONPedFDxZh0FC2iCGr9TiRaZaYQPjFIr8wYR7PBjh+4iirzUtMMDgzhQNYXTV2G2A9C6UaTvRAOYBhR+72n+X747tCM+eZwM/ZPsCY5Dli4xhlJsBzwq05jL9T4N2j3cbgCT1HXfzx62XqaPkqCB1B/chlM2kZWUwYLcjYCf7QYZPPWe+g55jPSwntfyEJLLyYO392EtrR9oueY0cDRr02jFd+d9iedvvJCYxoHyv33nU7Gf3EKY1nOnLaDZVTkdu2MWVmVGie03mTqMYZa+27vGE9FVTK4x8e1faHwOd4YtclPjlcTOFJOe8WWyuP/qwsFEdhoa/YX9zqGBjhWT1SqjcKPdNnZSt4V68saUBSxivLrGSr8keDi58EX+VLWs88j+HkP/8D3XseXdzxXfzRifg3yIIUkIFzHeOFMhgkVTxndnPUOElvktwF3fbmjvZBqdRfc9LG1EJESossqG2lrppAi5EOvli/T5RlvDq4X7Yz1IVMn/XKMl7jYl3cq/OI6BMUYf09+qD6bi1zr36BNazq/9ev30aoqQKxmOW/sLVxAu5bfqRDc5nvavzD/7YXafQ0YTmSfPQLdTEC81UdeLGnx8mWhhMOh33iQT5HfALaMQeg3hEXs1RFNFpqXON66RevlX/42tfLL37xSnnyyT8qDz308cC94SIgup4841YXZcMyXib2j2mMsH4NbQz6rOCl/dNnn8Xj5CscDvuh8vk/+Xy8FS/nq2d8YYsvlimTengsxqOLvqQMl+aZF2EYXUKW6XmisVfD7xiyTraWVPksdNqQz95T1xLtZvt5npNnW2loyWGu0CRpMYI5nmgI86wnDS05M4+v9eiNl7mSbUHZJAt+yusrzM30pPGT7n41Z6yBx3PWtvjazlnOwPr4Iw8lrkc7mHdP/6plyXP0OKiaVgp968/w+/Bzk+Q6QW0ZNYpfOwj/910SSECuua4ucPh9+LnJeJ2gf6364ctc4NEiIjRezXuAqxBeFdOQucbVPMPPNSS/ycjPbh1DcTzakP+q9VOfzBIic9u1qiYMSR/UjUAUmNYRDolcxYV4iJBvDqTiomRswug5nini4bA//DEeJ27VmTyKuNDKTZkKeN4cPpOHzij+cfsbqj8Gh7Z+6nPgVRk1XRQfqrPGHit+McTgFtpFIeigtHXY29pjxWyMO1oKAkuBQ92AqZElsIujMOdOFPfaOaug1lKqu5uWzxUEj67nHo6k0pFmtH6LlD4pxIIQAkTuDaTWZMEmti46KK/qwVJhO9byGueXBqp1VmWDchhcXc3YRiBtaUHl8FG/QON0306u8UTlUQNUPCsYsD3PQyRV3IXddM59qwFABZfkAND1JG//YTgpg+Xy4P13lE/4VZ3DfJ0BxdbBpauwblmBcrPfmXCxFH/ht7b2l0cqIoMhJiKy3vgNjcCHtlY5/+KX/aoOhpNtViaZqLsVKYYz6KfxJNKX3A59lr/tJ3kbDwG9TvwCwBtvvFYOoGhNcPCghpNpFLlJD85CcXPypAICdaANhgXySJoegCkee9C2B6wOYohq6OAKmu94nFCfY5Msrnt42lT2JmdWlkQaGqZJgyB7tAsHwnq+CW20ifFkC+W3H21JfOwj8LuGQtMDV3WNVYlimAUw1uXCF07GB1vLGE7Gyh9+5oHywN238zlHFAW9VVyhsB/w5wR/xP4ATBqT5AkHtbgf0zga94KovEH5Gqw0lmnkiWIFHCMMMB0GGw+VfPnV8+X7GDnPX+EALs5oYaQiAWWS2H7Y9n/tOilYIgKRvG6QPObXHXoQbhqFJlwgk4RITrQkmsxkxfzndbdfJBmYEFe5y3S1fJAlLTjbeHSqTEat2fiUUbm94Uqy2XeT2QLytFuPGcjov7g4qywSZKIoamazneEL2JGJJ20BYiMo3p0OW3Xuu7383ifv4XPE9j/kiXSnPEEbrt8CATvw1Ru/0qGp3/pa/G15FS3lr/0NUcXd9ir1cFi2pJltR4YEKMu136b9rScgG2KDiJkh9ZcHnkz/3vWnvzbRlf7mogwzJt8w/Wv9iQkCw/UnAzmAhewqQlml5L7sVh1o6WdUQ3/6TqW7xK5tbZ7fVf3CLRUCJz+/K/wHCK70HWB2ZTmfW2b2jK7GZJ2v6ijY087cKjkCg/UbHPkWxk3wvp+rx3/06tBKPMLTGkAbpFTE2XZenv7pa+Wtc2tsWbzA4Y8Yk+FVJxW6Q3tIoDSXzk6KdIP2s+YzbNM4eOAQcnOyrCy8W95582d8xeYIq3UovMitBSbaGngnUWon8ACYnz1SZjGsrPBpx5VLF/myKxN75GwHL6wRjNSeK5AvsCHvZpnsT5FW+DxQcAlPE23FYxhC/FrZDoq3aO1gNBljr/kYnxzWw+Uy20nOsq2IAY7y3GLIPnqQ3VRRZ1vNzMxEmTk4gbI9w6GxF/NZXvUH97wrNB1DXY31izvjnHl140nd0FfLmTdeKp/hc8Sf/cwj5QjbYrJ9TAWc/uYe+PCF8PCnnK/SmBfiverEboj/wcueViPNo46zVJ75yXPIv9ZwwqeJGc/WGE/W+arMJOOTn0Zdwrjh4YxupVqANvZ3V3M3mHy4VccJSr6WQ9v1mDAbpyeJXzZytXUTD481vpCjgWWctnEhR1CO4HGyhGeKZ5zMx+OE8yw8HJbydJn38ESNr8f4+pDGlk1oNcPBwS4UrHk4LJOVKT8BTBoNZbMYc4Jj+Bd8ZXiY14UZLz04/DJdVwMcY/A4Y+80utEkK9pT6BJ6fY6Tvsqw2gkc80M2OwXPfRQXS4OyjNXWJodj1APmZYz5ep7sZOLLQlE6EmkiZE3nRU6LJs7oKn8Z2yjL4IyDykt5n3G+sDB0C1t1PvvJ+8o9H8VwIhcil1KzuoZJ7R/kTuvaWSkouioy3kW6ur2DQHVI//Nn4tTPo+NFZJsQQrN66LflVVnuRF2dLpWR1g3Xrl6ol9W6lPvmM0+FLcmNhMc9kyUH1Co7nbhGQYEeTFhz/pnh6ME57ww9MjoGWT1fxW3sOHKVZ198q/z1334TB1M8k9ByVG2yWKHBnsUojTXqW9ZvtUFSBIU9HjTGh0KJ96nSP5RMW2pUSytQrUktJu/SNAHCzx+vtbVqIttuePyXz1r9e4C+LT17yBE9sPRagOKcVzGXyfMA44Bl2e4aJHzOhJw6r1e/gVfX/+vNfyppRGcYf7GNrE4Fe/iDSU0YyBw7CKFxWdZNIr3INdyNdsDTM05uPcThsA/Uw2EdR7WY2B8wKiubYZMA8BqfIf76N79VXn3pF+X3n/jDcv8DD1MzGoWTD3mJvB4arSFVA6xfz/Kg6Qn6mOcFeojqu++8RRlfLydPnOKrOo9izFAWsFAHf3leiv1nwBxK+rt1P4vUyOVN5JZyRWPrNPIj222o0/rjcSIf2+b86T3iOUx+ZEMPE42znpMlj2rYsV+YVKO2esMSsq0aTfBK4XBwv0xm/9/C6Gs/Uq/Q4KvMXmLscLumcm3eL+1gMbdt5TH79CYGm7fffQevwwcxGE/GEBPvLUELtCYWTOAGv3oRlkvu4fLVQpPegGTJfS+0SVsTJ27fT4rkJ4JjX0wth/r/Vef/V4Fwtf5Ro385/ujlQB0KDJEhiJK9pcdVFV3vdV+WfS/Xpr4m2oChxsmj2T6g+mVAL/tgxIvvzeDkewwTVt4CkkeBkVTm5TewOTG0w9hRVIu4RwB3olR8/5kXyw9/hOHEr+rE44QkMjUTOrqCNXCvJsr2BPK2/uoqST1UZzrrIGd9t8kAwAGFLsREmG+7M+EcYTDvMNnocHhoj5WWLoJWpU2RT0+iXpQ3CtPQwIiYUlOyoxQD+q5Vkw6cLS10Uk+lXkG5yeGwHJ4Ww4lwSD3LAQ7hc1CRMvaN3QGfAUc8vHZJGXxc0SY/M6UMGuDjyprudY71O4Rv8sc0tvSnZksfhWbAiiJuAsCOIqpCk/qhIzgpWLZVbqB9PaW8Dp5OuIQnE05wFGMDpFoeWTUcoFg87Fd1HvVzxNRBOreXxHCyC7TkEbBKs2DkQHUV/pbpAChN+AmIlSi88erAjKpS/q75HPE2nyPe9nBYNI/s/bR8EtatLAxD8iRFKbg1AqVcJgIbCP3X33y9zLGiN2bbrCxyEB2uzBhNuvxZt8rOFn8qMioKGpr06JAXeuCiMUPeQfXOF3liXKN+lRqNKvKaFLNV5UnL0wMmChUlClFgw+OjzwrZuiuefI64r+EEJVk8NeBYlweHxW2YksQhvBPyOBhSj/0BAMVxm0nJyfnRajjB42SHFT0VAj9Xq9JXV6SAI4YEieMqKsYh0RQk2yQNTbk995hzt8+RIIqw75S3BZ30OBnpTpdXXj9Xvv/M8+XCAkoIBiA6FuEoU56HYntaNLykoUBvMpUsFYIdjaA808Wy59bVB/cea6iyn9tu+dpCZEuKyQ/Jc9WSbSOJwV9oQ5m+S1xxaS65Vly8zJd2aPL4bkt5+Ww5/IQmKaK+GkORvlAW9NO0mFy0t/QX1Wpkar+qg0xrD4f9mIaTu6vhBN6JV1fKVf5dv37bIxUkHc/N1WIFhar3kjwPvbflT/IoSutXdfQ4ITM0F38LE0/zQbBdmhmXJifUVC3+LU2brMbkqvXLhdKUHNelP0lTWM3T1m/etIQMdx36W5YKl4YE1cP2Kx+uCHmpIKQMyv4g6v+g8M8ec5hJw5pcowKnF+S4Z5zwpRmbQ8QqbYDCtAkjhFd6xO5V6W/a5gpB8hPaMDQmMvnpQ45V9gnTu5VmgxXkr3376XL6IofyrbCddKNTzrJtROOlEz1XZ/kObFYD9eTYQomcwVhxnENsp/nc+AIHxr7+0jPl/JmXy9HjpzB23limDx8tFxYvx7g67/YcvEdm8W7YQjFdOPNu6aL4TnbZZLNwIcbccdq+HpLslw4mUIo1gDNxZzvP2grGFdpZbwvPSRolPAKRSV+H7Ssjc3w5hU9DMmKXpQ3GOYWl4xoHyMd8TQfwkOuZgxw0iNHDT7KvcNDpmfOXMBzgDg5dnQw4BGloQdCULkYFzyDrMR5OjrEss3KhPHz/h8tjj9zLHn4WBKBsJrBMRDRmeIWnKSTjvgsqjfwhYcZl27BJyKOp9ySQ56Fd5Es1P0bH2d6ZLq++fZFPX77ERJXtJizarCPHx92qQ7/OV3XicVIPh9Vj0c8Ru7rqlh5XYcdQ+jU4CVv9HLFf1dHjBE9GcHV76gQ4ulKrZ4rccNTDYRf4HDETpAO0l914FSOKdJ9igrTCJMixMl/VicdJ3arTZ/KjsWScCcikZ5ygN61jbNPTBcqH1xSGynt5txrrFROMmfDSWGs4Wb1SJjicdxYcJtn2OwHP4S+SiZk+ec7hwrXQ2MUC31w4Uf9SOoxSuIY269zA42QFY49fpvP8B3uZulrkOuVEJwBBW4RX7gbWcOWIW7cFVr6rkyC9mxilMHzfcsLPEd+Hx8lttLWTa0olTvOZOLo3LvMmYBEy+5q0DIOFBayBf2ydDi2IUkOo8o848KnyN7kFLle+ishz6GiBIabZNGqqi4gP5SBTIBp1ymHCY5/nzv/cc/PFh/rnlgwPS9XYkjFDlgaJnn1MelNSd5xnvNCWOXT+5y+dKf/pi9+EPY/SLhixlWOk83mg12p0BNvbRhMBn8WXJ4UXD7ZF3gkTzuiV3G0P+4yiKs+EmQJyeksek7SXpaeGppzUNzT+07nBCdjRL9xWRkn0I3Sq1nAC7LMchupWFI2Lqc2ygNtnFzmV09afLpwEbe21bt8Mzq/4DtV/jf5BQttLOJocPCUwyFlO+LSWljBrsf7oKaGpZag1OU9Ss1RvM55xUV7s6nFypdyhx0k+R3yKtPAleOhxEo8tZRREdiy6wHlQP37m6fLSC78o99//QLn51juCf5e+71lOypUlDAd6XU0j++fwXHNMdlselOV5J0bob/33b+fzvnfceXvKHcVweRAD7wSHHHuwtN3KM0o0dkj/eLBgjNmE7hPT48j+qRg90v/Ax7FH+muo1gNwmTOdlF96qhxAlvnJYamo0USeCR2hv940fhY9Z6JRhp89dntv2p94jVsZe4kDDdKts+1xARqyWEeZU3jSWZqsGl5BHp0+/Vb5+c9+Vh7/vU+VW2++GVjr4c61r6bl4U0Ko3z7giHCFp7hngby/j6uyhlNln0v12a+Jnq30v996kfPaKdCLYLBonkZQrEJHo41UX1/j8ihIqvAtat4taVUxmlD98e16bzbuPtuBub6TeqvYr4OEA4Aqb9hoHCN0nG3dB+G6ufRzhyhqcAhTpaT+ZSQrmrnjBM8Tr6PIvFDzjiZmDxWNlEcVHDC9Z4g7So6eVRAMka09TeIyrwRVbkLa61DRo8ct06uHh1kcuV8GV+5xAQaaymdDy2gjLGC0qWzeSiZp7lvu2pH2Xaa+s15uqx48O4AJhAeZigy2VIknvy5eu/haeLmHmBPW9dibD4v7w27ty8JJ6tZArcDivSqbstmoIM7GiHU+B98VMQtJ94m0Al9sqwhgPpY1jfxdNnEcNJHqdhikj4aJcRqgANLdT472wxuORTKkZRBdLsZfJQJrMtRvnAgNDkQNC9Yjj189JOPfRR3tnvi0pyzVSJZWPkDT9M52XGVykv6Vy+bvBoZuE2YAVupxlXr4kHS8pcv/zCSOCfwqzp6nPTxOOnzaTyVIJWPulXHeqphp27VkW2Msx5aikF+g5W0N958vRxAsI7ijtfjQMBJVrxU3PSYSaXQQD+NfI1ooFbB4GsZ4IR5Cp7QSEIKBPxIhDjhCG/PDnR8ilB31KDuGANBSCFf8a8tLpZ9DBebCPNNjFp+WWfLPwZ5V0EjyK0P/uvTTtHN0SYrxbhTTwwnDtoSDIVpu79UTsyPccaJW3Vuy9Ydw1UkzGB/jcIXXHhqyiNBLcNVMS7Za9CFFrxqkFExUmHzXQ8t94/q0rkzMhXDiQc5X1iAzxkgpUwOhXPFKu3JYEX9fh0As33KM772danJ4bCuMDNoz2SfPEYJ6S//ALsQXSP/5BvooqGCR+Dih3epHCD5rc/ejWtupBvmP7PEWDSUxhxeuyXx0NZf+Yh3aBHPkyY/YFAO/UmYKMtDLutWHT9FzueIORPicb50NM9WnS48ptlFuOSJyg17MFo3b9TpD8/8aQS0Ph5rgClQmDxc12Ruo4osgElczb7oGSdMbhKrVm9GEkaESlczcVX8faKtmjS+GZ26fOZBPqsSxlgCfiX929ympzRfvf0K+svzrnbZ/q4greDF4BcRRjFyWn/amTQapn95+/9m9QfxCmpADvQU9dviryioJHZiCC3tXyiPY3hQnDp+iP4B/ZWb0DmkAtWIaZmAABcG5JR6JQWPpt0LbePy6Umz2XAOHmlAuBLaO91bXx9Ehr51ZqVcuDLgwFE+G8kBkC1b5LPA1DfFfu8FxmE/O3vjLaeYoPfKIl4Sa0yiN5bPlKXLb/KlOCZSI7i9T82XVcaDaYwa84w507TZzhU+IXzmTBnBo6+HfB31qzwYqZWzzMkwinJHButZsYkxZ30Fzwgm4fZ53bbz2XSU+B6LAJ47MIo3Rh9jzBrnLi3Sv5ag2QaGBr3c/HJEn0FPo0IPGaTBRFjoNBhjrsQwy0fm8V7FkIK3iyuYKtIaUqbwopjCpdsvNLz99qvlxhPz5UMn5vgyxaly+y0n6bN4WXDQuv0v2xykuvwHwdVVpKvvuaS16RJeg8A0bZib4cTrUXge2j71Uw6E5ctGr7/DRObpl/E4GWerDp/D5NwPXcfzOWLg93D5ebfg+FUd4J5jMcS9+atMcDy3pDWIeO6Iq+kLnnHCuBTDiQYWxrwJVo3HMFKtUJ4cdPQoZ5ywmut5L/Nzh9P+9jfPEBj+HPFx0i1yDpirtjnjBD3HM07GaJMJDCeb6Ddu3fEMFnmtLvxIBuRV+FD05VTHMgxTTIhGhRujyQS8MbOO5wlh8gXrP3YT14QwfDhSVfb17lvf8qUxZbuokLMbSOcWV7fqbGCAG7D1VRkcf7zwv2VESwtM9j+19irL1Qq4KMv+F4/aWEGon06rd9SHjk2Uxz9xN4YTJ5d2ZP/ER+9eELQOxzmFaS5KzLN3A5r3MeqNHLeqJq1wmD9QEMaYqr4FeoGVW+AcoIh6YH8mgNDaRJKhHf+qh5NlVkojQHnOq2zXwFHjHP9Yf+eMOOgnHsDk0C6F5VtzqgZ4sOwI/WtVw8kb58rfffE7+VCDHj3ST5S38eTYYezxK4COffa3OiPyTkGkSV8BXg0Bjvu2hBh7NUnABb1AgSfi/E8c95RBouq10oZTNumi/5PYLJVgtglxMI/ytG7VAUPyK0tsT8/UcMtLtuoAs59BN48LONLWuluPkzStAPJnu8kvuSS89VKwfJZQgbAi49p0wc4cxjU30u3TPxLcRtY0/OaymNBS3pVvgDtmQoy76ljBH8NJ+pV6vFt1bsPjhK/q3O0ZJ+ijMZy4dQ2dQLb2jCibmp5Y3nrrnZxxctONNyFPjmSuooeXh0frmaOh2PM+DuLB1mUM2HYLIdvqRqGtB7n60YsvffmrkQn33HtvzlPyM89+dt4FYuc9apseIK1uv4yckf6eLdXDmNK2gfqX23CkZM6Xod4+Z5l4EOw6csutgxpvLFceM7+6Qbwi6S8uTluuBly3ImnAjYEFObmFjLRZNLxQPF7qGH2RcY4zlmW6WXD0DBzh1RNUQ/TK8mL58VM/4nPE3yh//OST5Q+efKIcYftRXQBs20tZVNu/6nM2WxNXWzC/+9qfEMDYvWrqtryrIttURLf6Zw1q64DXQ7U2ofc2rn32PsR/9c3AXDV1k2c4a5vAO+G/6/rh3drbUud7VPwewcOg/dLnffl98dqlfo3N776ENZm/7xG8l+BXPO3Lz4uMGAEhDPWl1kF45EfKk6GqgDOJwjLtx4v/zG/HN0MEklobCey4reHkB2zV+eFTz+Mo0XxVhyQK9uolweBBdhnHfDwGJF9DGqROE5p0+doK9emhoWBUsNq5e2usepx7s8xeOZs9tx2UjC4CYhQhi6mGiXAzilFoDCaInECqgBQHOrDrueJk2Y44ouWVVVNe7NAG6ubq5CYAefMSWIvyxj3KbhsIDhlkuKs0mEblLQBALx+TCaGTujFOBA5Sek7GgNWCDZSoFQ7nW3dlyMk5e9Zz4j116cmgP7SGk+qxQ3kqKxQ8UPHl2X2CWU0I3go3B1coA+47KL6D7dXy2KMfLY8+em85cng2gr0OoB4USxrpQZs5yPrTkIYXLuJVIOxCRtcB1qe9yzfp7RVliAL+9iv/yBknZ5szTlCSMuqr2GrcMbVDg7wUJKNA1RAHxy6WaT1OOONEhZOBoLuyUKY4i8JT/jscEOuqK42XcqNwkSfw8ePgBeCg0mfoYhqCFb0DHzHqZGUUJyZYAEjTmLSLygI5RDy8q/aQhqsC14nMmoYTlP1Nt+pggffLNHWrDpOA5CZzc+9QYD1fQwylJxRMHdKR+vrLnHqu4eT+8hCGEz8ZvcPKY1aRaQcHJHN5Gr5/lXaWZDuJt6DyHIuhKzeGG1uNH9IhsIB/zjJh4H7ltbP5+tVFDSd4oeQAP3lI/sJ4oqdKBi54yhmjBhP5bKByR9gW1S4yGfOrPbO4TAIdz8JGfuiYLT21Ytku8OUhsEgbMeK3PibUMFMqX2wKoEiQz9e7ktq4Jlvuzavt11RflTtLAz97Z9O6tgI8QRs0IQ7qbi/wqzodvqrzEJ83/cyn7mXyg2Ic5UWeqrRtQEtesu+7rHe4/mH89UjKfnuhw5jp5JzuilJRPU5cFVaeSs+KP20SAtDCZiGm6XkVlveB/3D9FdCaKb/8WK7PXsKeyRN1tvUP03+v/iYtGYUoWx9sf/h5FSVHpab9HLHTVWEGreZKzalT2H7b+ivQDfDN7b3oXwF4f/VDeJgQ2MFJusAeKLoYWsc75SQeJ/azIMGPJZqoluz0jHY1k1ciubfvTaqkrRk8hqHGW4YTGTgoshHi6Iq9trFT/t+/+Vp5gS12C9gyltaBC5nudhVX3DfdHw48q2zvmOXw1VN8fnhqehTX7DfKxTOXynG259x8A96Fg4Xk+fkLZ8sCZ5xMojQfv+Fktumsnudz8W++WiYWznN2BWeK8HlZPfrmmBkPGHMRRBhYZuL+rQeD+939jLCHoY7KyxqjYWQP23Y7zyTeDV3k4zITgLOku4LnxM78oTLA2OHETknZX2eigGzx0NgDeD86Xl/GC2YZTxo9KTuTBzivAfmHXIGl4BUc3SnTs1LGWSm9yLkgr73yfLnrzpvLw/feXG654VA5zKeIJ/FuwWZEE/oVISbmGHUc12Q2FW2NwtYrb9ZPvxMGLdv+7Xho86b9DSX9Fu1wBuPm0z9/Fdhn+RzxIh4nL0d+7hpO9BDBiLHINpn2c8SLi+z1p53zVR1otM6qsNtvXNVdYcKjG7oTkmzV0XDCswcgbmio0HACXB4263WULTiLGFHconWQrToObTGc4LXiFpxVJi3qLm7p0UvFcjw80UM2dYevZ5ygYzCR10jrJ5JbuewUL+MIdFI21u2B0AGYJmnPcc5mmcRoMgn8E5wzNs04JX+M0iaSygUm9UBZ2vm0i1i2FxIv44vPLiTYrRyJ6+GwbEuifbY5DFOtjFjGGP5s75RFeRbIZR0Wnq05cU0Bef4PG05yxgn6zY3HxsoTj/FVHc84QcjWHkqvBAb/5AsncnWSKswpnd+GBiSD6+nEgciac1Xs2rSAQ7qUzZ0MApf+WPUnuYl3EXbAt+7UE6BFJUFV54TiMInlq8NZq7wZ/jOhb1arrk1R6tNGO9wakVfepbUePSv9bnnhtQvl7778HbZscMAyh+8iUEIrx5dtF1CE17Itj1suy7A8wpzQVj3ScttL+KwY3cItJxDIRabkJ9hLValN3xqv1QdE1rjQi3fzSBLAyrgbjxM6ne3iOOLnc2M4oY/YHz0c1nmAXhBmVu5Fh6bM1nBi/bymfu9e1pH29cHAMNRe/QmrKY0kDXGWn+cmeeKbn5Sxr6jh2NTVtB51u2DnaGDfoF3Ju9Ms6nVYTC5b9XPET3I47N0YfV3ICwzOZGQT+G8ZY6Vniowh706fPlvcsnPs2AmMG7PVmEC5C5xT4sKYB3UfYu7QRfj1MUA4X3JByIXhUbZ2vfzyy+Wv/uo/ss3lYHnyc58rd9x2B+odC2HqPeDsTxdDsH3Mc6U8vFoZMsEWS43y+cJpQ1jpbx7bU4Owcsi7eqtnZWnUVsb2kTUblKE3kV9lM4dyyi/zeO7jYeCdwBiiwURdVwNN5C/pzK+e6Zkmzv88tHyyOTxbI7pbH/Xw85DyxSsXy/f/8XvlG9/4h/LnfDHo3/z5n+Z8J/tg2/7yX6WvoYwB4tzgY0x7vUdwG/0r7/vy++K1W0+Nze++hDWZv+8RvJfgVzzty++L129ZP/OjsMgvh84kMIYYSOzdOisIv/R3D2hZygZ6j2sv4bUJfsf1tyi3FQlTqNAKzzaCe40bxl9AffemEln/pQDCFBLDhpPvP/UcHeEEE0mUKdNKR7cmKCgQBgraWkdb2d5dcMLMbRNhiZb+7odTsG5hYSyXz5fJM6+Vw1fOldl8irYq+T2ELXKmMh0CeJMJkNtZ2omHw5GWY1cv/AytVajQuE80gl7cGtiyusWz++2SJimELgjs4wkHByeemXgHMylip0QYEa7hxLrdeuEnEiOSEWi67Wa/JoBIE112O6ymbfB51xU+A7mC4FkprNaR38m323X0qciaAyOlUKuMOAJukx/MaAfSQCdXf11p0etDWJykeaU+DlO9/55b+GrIRxDAGk4og/QjuHSr+PqddrXOEYwUOwjBeC7YhuCmsEp7pjQFUlrYqEr45iaHGCdc0q/9qs7Wtkqsp2VDy5SF4YR2U1mwiHYBA8hrW4C3eLk/8/U3+RwxMPlVnc7qQs43meQwQU/8t3Kp4/5+y46nBeXHCMQ9h81S5jhK1DjGqh7K4A5tQOp4KHVQ2OuqlqOVihVFijO0D79ypxXCm27aWMW4sO5qGUqf55z0GWy2MXAJh5PsUNv6QVMPGI051YBGmbSJk6PwpWRg9e7YYc44wXDywD23MZgyM9KVWKMFcNSvI1GuxiAZVRyFJeAJb8DkBzz16nILCvHmhRGom7zgVBflgJFzc157/RyfDn+xXFrCuo9StU37Q7iUwVPFnzAdTbMylT4CfzmbgE+pJXv3O0yaDmDVz0pX049sQzQZSyHcfsFje/lMvEmsrj7w3ISZrMYRGQSH4xJDQJOozcxr8hvuM3+2lvQI7wKQ/+KpkwaBFqSS/vlEMyRUD89XdeB/P2ep4eRBPm36+CfwOJlH9sRwonTxEg4v7ul/9TEV87hbv/iDwzD+9gknBbUEzXiQihe6QONxIi8LHa3Q4K98MaT2O0pPBZZgS/nC5WvzWEcd2/P9078BaK8oG+d90t+xRZziBg4iy+xD1vvAVaDQvyk88Af2BtzmueKTqn+j+vcy1TJ+V/jvrsLTZlLaCfYAmT2OZcCv6sRwYpXgL//l0dcGX8848bmJqgmGf1v8zdMYTsLHdHJlfEoiwD64hRfHv/vrfyg//fmbZWVrrKwPcLfHcKJiqau+n/3dZrvIFbaS3HbHneXw4UN4R7xZ3nznVYbeHT7LeidfGLmJQ1QxMpPnG996plxeKWX++DGMJ6w28kWcZVYzR999rZxkQWJ6bYGv5KzihTJS5qcxCuPlp7z0gFNN3Yso1OsYQ+Rvt2TJw9tsE2FWg+I8UmawGk2iqPuZ2UWG7HPgt4Ki3zl8vGzgQaIxRKOhSu8USu+EbteIDPfnL16mbug87VZVDipcjlGIryegXGtI8DDVEYwT65yYe/bcu5SxU265+TifXD5UbrnxMEYtzl4Zx3ztp5gZy2yBDWhQ+wtvyCSV/oz1yif+SXHbSR5t20s65Y00jk9O9M9zntqPfwZNOXDzzPnV8pN/eQUYMVjkDAAMHYwJtsUyk70eAneWrb4aTpQ1c3jmtYfD6nGiwckVXSdFTgqrx0k/7u1OQFxh9WDYMcrUCGKvieEEj5N6OCweJ/xbwTW+PePECYn5/Gyx56xs0GYaTpwYaZRVj5pitVaPE93k9YJp+VeZEyrA7NKjjl3wPmPRlHzvgbCeb8KfhpNJDCd+fthFBw3QOR9BPoDU0lHDX8Y5SnPMUw7Wc7rUUzijB6+jFdp4jfGz/dKeNbuFN/xPGWaNDE2rKNOkAhcDtDLTQm0jPSmNE0/POLn9OIfDPnZ3ufvuO0jrpND+ZAYSWQwdLge+U5iiLh3cOw+OnU7IwwsszGjstv+HNvyCSgOEkKD7qUvwmLE5IUb7TxqGgzC2skU2yRNKXFNvU5TtmEItJwl9tVbScxOm0KLB2c+Np0CBt3Ivi4Ao6i4bnB/kYdJ/+9XvsqXuOEWpn6hTEUn/NEuoYRVcZvXKnbjgo+5HWMZzq6gIJFE+XMAW3x29V+i4pg8duTuWBnvHH4yuUjRlWL7pmqLa9KgrNBH6LP3NybHPGk48AFrvb2WChXvQssYcjQFe9l/HUwuMzhqlrSmfOq87/iWn8KXI3XYPTPviCBFfbqFF4pKrSeWNd+nZpuHupXyMxg/+MkUOVwbOOjuxXuQ2b/pi7wwwnNzi54g1nOhxgrQ3HxqkBjBZYJN+pk4/QI98/fU3MJy8Xm699fYyPTMfo8oacnATmngwuIYT5YVcppeH/Kes2cFTcpOteT/4/g/K/0fde37pdRx5mlneW3iABAl6L5ISZVpqGWqkGUmzO2b/vt0Pe3bPfJiz3a2Wozgy3fIi1SJFiqLoAQKErSoAZVC+ap/nl/e+9RYIcmbV5BztBeq99+ZNExkZGRkZGRn5d3//d/g4OVr+83/+38qDDz1EPOYi0HDmDbS5sqqLH57qJf4HSK9VijS8jdLK8T48NPIpygt4yTJK3g34i+2nXxWV2vEJBhz6pEoa8Ok/FbjmL/7G4Pvjo8w7gFOFr/FUPHtXgbZOnuatQmUcC0OVQubh/Ez+rAJU+daxeO7KpfLSC/9SfvnrX5Rv/btvlKe//Lc4GYdmQie2TJqqNhvP+gokq1tf5Pn+9r911JtDJYOarb1aaD/g2ov4/gh/heXfYqtOA3eA9fnWVf2wetp5Otctkzepue11xE6K+vAxll8ZDIB1yhdIXrjZ970k1lr1WoEmRv1ox3eQIo6MMBHTwWWKEDBhOof9ze/e4O8VzAOPwkCqs1E7BU4lKA2GR3rjy/RSvrmnfAgsg5XvEhtER/5ZFeBtRMeWpNlAKN+9fKlMXTlTDrFVZ8KTUcjK43pVVCigyKIy8Uagwr9SUycYWOJIynROG4E78yTyRWvOm8xahmACTckMVFDLIOr3fKnZBWTexZlZxYcK736tygyHOusAbAknb8a6eKgWXwwOG+wnTDnUO0wLOBQIN8enyjIM4gYWJytMyBF3YL6s3kXLz2AUKVtBgD/yd6BVWaAlguxYAb+2kcadVpB8yVtAd2GQ2yhOHnzgtvIJtiIcPuwqpPnA6jHxdCURYODbQC/DVfCEmcY7Nc8R0hG+6sSYZE6QmwHGaiqQSlCVYdjWaL4J//vv/5w9tziH5VSdbQRN4XVFq24TErPkY9ulvWwDKarWy28qTnIcMYPCIEJ8LxYnCm6aCmtBUmkHmM0gICgw076UId3uaKHD+wD1G+JvwMEXXETos3Q7gZNm8cVv+896eMXSI08o5KARFSeraMbdSrWN8mQLpdgmW3iCRxratncyYV4qTKQXFR8KXcKjQOnlquoOzmCP4Bz2K194rDyuxQkm8nUlrAouWpLY7yDPKNDSjkGUeCZQIuMjtaUQ2492V1KhXPuD/Uh89imhgMtdFSdnrpTnX3iN1WsVNLSJ/RE69HQhIfPPiaNCsDRFrlEOan0CoeT9OpODPpRgE7PTtGOlE/G1rcUKg7iVbbmFT+JC2YZsxQo/vAimn9qHBmc1hFTNu1AJfV792F5J2754J4b5cjdNPrdxVDyRYS3fKOKq4tWUsbCxL6iMLVqccBwxp+rMsFWnN9vBpKnat9osTZerCfC75bf0b7AwxzLIsuVp1CT0Rd8Qv1om2LZzc9dzHLE4o1FqOtMmN3FRc01++crTTYDU8qWFPOVzjd9AIj3Qxobl4iHo4uXD+L/x21xrfjW5BZhO3qnQpCXaMg4k3a6gxYmlZmVKiKSxj6H8BpLcPsr6221CfzwAeVWy4Tgvp+rcbHHSoDetDYLS00RU52rxX/Pc98k4Do3+S0HEhW/JA8zPiYn+r5792Ys4iD1dzs+xEujuGBQnsmt9OMiXd+FfJ0/exWR4hv79NhPlqwjEKDOI9MA9pziS+E7o+BoremPlRz95Aaet02Xy4EG2xCyW5WtLpRcriWnG1SML75ajbFXDsIYDvzhFgTH2MFtHNd/Wkd8Cq3/6WhjC0s5Fh038Zu2iCBmALw7w3sdpESMol1Q2Lm+gkGEBZQlLytURtgYNoTzg2RXvLQT2GRYIZlG86pviyvxcucoJQSJ9hFPldA4pJ42Cg8m7x2q6b99tPdfxDXIVqxT7k8dDT41h4YRD54fvub089vh9nISF3wz4qEewqyRw2wyUKJJBKbDx5+V4L/1nHJcRiX9wrqxRxzLiyRMZ55bJ4zxbpH7/53fQPR/gFMH18uLLb5cenNjmVB0UTXEOy7YkTed1DqsDXS1z3AI7pXPYKERQsNA3cqqOihP46RSKk6tXOVWHsSyn6jBR8PhQnSvGOSyKE6nhsFtwOoqTWdq+y+Ikp+pwLDJyhccWq2xxpbezVWdV57CszjaKExUpreLEWsubrWsciltt6RGc9ENfgypdsEIax9pkFJoYXucP2vJEHflolPOk8F94HfloealC1UULlQ/ZZgO8Ljzot209ihO2JUOPGOCDI8YWxmkXMexBtpAgCYn/5D8+e7Op5Dk+ZxKXJTPCFOhwfn83zmGfxjnsg/g4UU5ScWJqE6RPEyY7cpBu+V8dmwzkj8hAA36rtZOw7ON/NbPEC/Pmu0AJp/gwL/O1JsqDKhPSv/mQepjeShhqJO7SG//JQRhqemO0YabX755xk8S4RrDU5MWTk1nktLWtASbYV8rffeenOahhl4WzOJBPe5hTcgguTEpWhPFPOCkwYwR9QzGi1iFYTFmW6LId56HTtvjrkPeQR41nGp+og/ApPzrBBVCxYynb9jkz8cfKIcu47cP6u63CZ+lOWtVSSuWiBWiVpjWDW3UqXoAziK5jt2nMVHxbljJY5xJRVg7Q9hRxhAlDGw3cyH9bOPc+1byMZti+qw1MHn6tOJN6av9ppKeUb+YqwvWqZ3tr/X293I/i5GsoTh68i1N1ADB9ifbSSoSNvZW+wc2V+avlxRdeKqdRnDz2iSfwd3QMZcEcylhP7JrOFj8xq6JFDMhj4twVXiL+rmOd96tf/qr85J9+Wtzq8++xyrjr7rtZjMaiBAsRcem2m3WsAKPEJV22xaBY19JEnm8+WsxaXa18PGpYSxPlc/HvGDGjMpbO5bjkfMI2CxwoYfTNdJXxw36qA2uVuJtYpERRRnt3t7+KlKvwOnHq1qNRnGZbvkoU6cbyxbjwaF1zlaOIz5w5Xd5+5/XydbbqPPLQgyh7yJO4e/IXWYQ0bVMeui/yTob1p/tLntumft8HA9L+zZf3EUkbgQ/EkwxvHeWvu3zosultXfVpqpzbfgQ1b21ge+9O0JWq/dzeO9H2BXS9dD22cfcHNW9tYHtvI3fu9UP7ub23n8OcpZMQB19tPa99ESvT6cJOItgZHQCdZDgiRBNv0ydTGQWdhNdL8yvl16xiP8epOntbdSDaECqdrdcODHOSicBYMjQGDojYSDBQVRiW7wk8205ssBjJ6j3MktMVkRKYVM5fLJNXzpcDNxY4jhbGQtkyS/m1XdqaOTiFgdppfCbPxBNmXrL6T7wtZi45yg68iJqsnBBFhYadeZRVLusvUwioTT4k7bpMKLO3rAwZqYPghuVSpuWpFXXVzq00q/hmERebCDY6kssJPlMH6eiD+DaZKEtMRlWcLCNgbqKh1sogk1M7eypjSTLZ2iwyBwUPa6/WekflSoNrV8vYbR7AdhVyESweeuAkHufvgCFhAeKkULQgzMQsHWsTFS6iiuwQABmMyV/mFwUPAqyDoVcPq2vWS6ddPQqn3HsZvKNwySQbZQyw/+MPf1X+dAYmv8XKh6sfwBplhYoc/UcgZNbBOrUKdThwBOkojdwX/u57jcXJBhN2cKZH/2GYpttu6hYmfW/IDhULxATIEUF5r7TdR7sOYKEygIC6hRJOLXkPSiJPNlCBNsYqoI7mXD1z0Hb10NMmttHeX8dcfRV4NxGMdxhoehGKe2mrdeBbY1BgqAIH+HWgSEsNhnjQyin1AE9BKoRap2IIQs7T2apz9MAAxxE/UT7JSS47rAxk0iTeyUm6jaAjAdomyZlnw1M/m0chlVgoKGv9/UibgGcViiFIJkNJBP7fPn2xPPcvf2bSBD0wEal9sa5ee+JRLm4Vj/ZV87YfEAgdWD8VJ70MdmNYSNnX0qcRRHeYNHsyVFbboL1MTujZHp8YpQB3J3zCK152FF6FGpqy/7kHPm1nKfCAtOgH9T+jUHL6OumEQrw3pOuXyhOE2w+WKazecg9max5st3J1dcf9/KgrH0dx4nHEB1CcaBIQy62ks8T914eVn8Ka8hXgVZ1Y7x6FWR4Z/+E1HkfMRIdJXuofAdkyKFBs0Bkj6Af+1DjhxvDaK18Mg1PD6hd+re/N9W/qULNPTFOYz/v5P22Tcs0qCGji11t4I/VyEq1ptb4Z9rbqmKfpuNPOe/j/6MoXio+j/raJtCKZCLdzkh0m+0PD79+qYyQ5jknEP80ahRi3eu3Dc9dL8xhduCiBRuUO8aFgm5sa/rIOP/rFc6/hzLWU5158rbx3iRNT6Ns6BZX3D+Ko6eSJ21nNw+rk2mpZYRK/g9XUOlYCnjbw2P33lcceOYkiBN9gnFzwyqsL+CAbKZcXrrPggfNX4BhHYXho8Ww5cOWdMrQ4hyXedpli4WBqeCK8OiuL8EwFaPu20KksUx6I8hoYh+EHE6RZ21gqcwuXyzLDUM8Ep3qMHiwrKBh0eL7GOLPGisYM223GEIZ1ZLuI0mYRAXyXxQ4nf+JSwunFUmQX3jrFKqZKFvlyFAeePuXECn4Mi2ZRgdOFrl3kpLiHOVXn0yhj2J4CDlwt3XWVFEsRcZu29Mf+0F4+2g6U2dK/rLq95E2q3W8wnp1jn9TLb5zDke2BOOr945vvxVJzBYuTG8AS57CM79dRWgwwpk3i0HIRX1AuvkzNTGTVdxXFvyuymq1rmdJHBVScuFVnGyXVGCurTg5c8VURM4xiXlN1QT6MoitbdZiAeKqO3XEF57Ae/+0x0zpa3GDC0W7VWYceXNV1G8gaY9ggZY14kg3+Ttz+qkNG87X64lwHkvZ/r4xRtLXWuX3Q0SD4HMe32DhKk1GU/cNY5qlk1geDY0xOgWHkHoX/b1CuTsLl/fpa01JgACc5ylU6i93BunUJ+l0fRs7B4rF3AGskFPisGlA+YzLjmWDI0cV/y3/qnYYE6IwfxpFODCJMxVMfcN2Bc9inP8tWnQfv4UO0krS/OUIH9lX7Vu7JICVZGoGpvyTCGzKlfN+8m4Bu/mcELz8lL6Ft+W8oONAbJrw5qYhHxzRlXYPJ2MS5R0HoOM74a05pmNrLEjfbizJpVM4jWefPlqr4sl/2e8oTizivv3aufPcHP6e/z2RRp/Jf45m/CpgKQzvcBw7qIe+y3Rzv02GU74Ez4pjgNqDbJ7TGTTpyjeVJcq28wXL6kKcgjJRpkZZerZAq8uIomH5in3f8sDyP7tb6xEZVSSIt2dbjk+NkRRhtbJkudGg1ZN9S1NQnX/AvrslevmgdDdOKomlIkhJuBL8GzeLDIO4o7uRvQkzOhAJL8CTdGKlinty4ah5UKChwMbC2OlYxKTtgkkWV/20bIgRnynf0Bsj9ernvLp3DYnFyj1t1wJCsSvxzzwEC5OWx4mfOni0/+6efl9def6M88cQnUXrcRzy3OU5HCRv/dSlAyCnT+oAvj+mdvzyPJdpyeeP118t/+9GzWOidKt/85jfLydtPUmdkcXipWxoX2SZ59SqOqKmnDsA9mUdeZQWVXOS3Q8i+XjnBB8tDLZnCf1CkO39x0cR2c3HY8t0eaJ1WsBRXOexcYRqYVcbYnsoMdeFFNwPgEnhWyHeRP9E4Sr4TjBO95LPtdh7yVWmiskdn7dfxDXWDLYTGJbsyd+l8+fRnP1OmcHguHhukd9o/NE37xzdkbc7Up/3ZH9S8tYHtvY3cudcP7ef2ftPn5rXra9djG3d/UPPWBrb3NnLnXj+0n9v7TZ+b166vXY9t3P1BzRu3PYuTNkZ7b1Pe4l6jfHDE93+5OWTvvfPUPrT3W5TbBtUoHxzx/V/2hzhQUPVkVxkn340iB9j7VHkDv5XcjOBlRyQVTN0rSg4fzBNqlW27Yn1lYQlrk9fLb3AOOzTqccQcKxUpXKbVWJyQTyZfMgZLgVnqv8Pjsnq5y3oyQaObbrJy3sOonUFZJ3aU338DB5eXL5fxubNlmu0aQ5E4m6rU6lXeRj5tfQW1xmhqRVkOdEaJFQ2lhqHDLESHVr1bCBkqi/RA314yU+vryknw0X4gPCsqTZ7e2vIs0fcMHeBITeuQzuRQBPhtE0Hp6vw8jL+vHJo6xESdUw2YiK4iJN5wqw4DitM4rRWc0PcwmY4mGo4feMhb5p+BgYLScg4smNBW9mm7Kco7CScFgs4ue4A/8eid5dGH7kQ7jEUNzLBa1cCY4NbuMYwAo7BpG1NfknJRTxi19dGSx6BQBA+e5qNwlFmg5WVcZmKIMmW7f7L84J+fK2+cZ6WyB4F2F8YLHrW2iBWEjNORj8sB28G4n7aQdmJJQ6t4WsDZc/g4wfneEMJ7DwLcGEJYVZw41IMHync1QwLoQZvfuVKFStP9rCD1YTrYjxn3GtrsVYTBdiIwCAwTmJGPMVFwz77MX6YOmFSU1VMUJ1fZRrMJc9e/yRDO+Ubds+8JO4DQi8ASk1gcgDjwOHCpCIi8Ju3kctoMfsyTOFFRIYgePajFyRPlyUfvAfUIjgiiOnCreIfabNtItKRF6LJJ/KYQUv+Zp/Fo+5TlR5UmxrenEU7dLZh1ARQn56LgXFjWnBPLMGJI/QxdPJPOXChTqAO5AfWJb04kdODIKjV0O8aeU0HzxCtXK3VQrL8d2zRCSoQiMuAuYbjXVyWco1yczZK2KmUogmgKZP6TJDLupWQiNfzG5bvab4UpSYCGFBKmV27Cbp14Bc+1EawEf7xWXga+bJwqKUKL9C/65zaKtUEmnSpOvsCq5ewUgFB2BCjbFORXzNTszO9W5VfIiJOLSIGH/KXPpKBvECrli7/5ucU4bJSf2mdN4VXbV3oihPJztLVVsS7Nldzb1yRMCF+5p//uvVdrL8P5LMOzE5uWd+kydeHXz+3lc7DZKX/vq/iVb9j/FYCcyHns6RCmvTU325pcqZPvH3X5whj4PuL6QwoNXsLlWDmDo2L5MYyi9WicwzYFUn/7Txw2BphKHzntzPdcFYM+dp7aB+5091qW+A+uoE0C7QeaaHNoTfnzW+eZrG+Wn/3ypXJxgZNsMn9QCJ0s04wXI6yuXVURcvEaNKbSF5pxqyr87jDHzh8/NArtLpVTdz/EqtxSee/ycpnzKEgm2rb7BDzx0NK5MrvwDhPkq/BAnPbRt0ecCKNg3oAHaxkin3ASJR37z1YdpO+MMknSCF3/TJuMMVeuzZcl8TV9FK0MCoRdlOhYnOwMsXLMuDoxgW8klGlLWI8sY8kgH9ahoX6U3NIS3stkaJbVxmlWPRk8oa1VVjc34WbAQFlOoobgO7tsz9hYulg+/+lHy99+nuOIZ9hCyfHAtokGDPTsBsG3pj+bQn6cCTB5743wtZHsk5sg6QJbG//4+ntsQZouF+ZWy8tvnscyc7RcZ2KzjvXFkEd/Ar+KoH4U1J44tpytOluxONnECsbTIYY9Hc6tOtTbhYkcR8wWJU3hx7Cq2Y7ixK0/KFhYOdbHiVR1iC04i1gIuX1BHyeAnP7mtgaPCXWrjqu/hzl9R6sXfZloceKqrEoUJzjDw/hOwyplFTimMGWXrYqeYAjaU3Fbeyo8iPzdOjBAmw5xNPUodDHOJGWMEw3ZVEwaxm4n+2ShVeoGzoJHUYRo/SS7ckKWFWFwY/d3IrgmP0KZs8TW1p3xA0zykTf6NdfHeoFxdgdZsQ/lgphvL5/Tq+Q/Df8NXzdc4jWcW7bqeKrOkZHylc8+huLk7tBNiCA1dIy0fZWRgNqKSswk9rXyXXsdn/kXS1Zq15afcojfLlgkYuImC57alzaF9Gb+3Bi+TFIAAEAASURBVEVILuuQIJEbfm5pRCD7hj4t03Qp2V+fmy0Pjv95Bw6ADq2Crzr+mQdq/+2B8uqrZ8r3n/lnJpjTLO4gayGLCcHONtbLDq5O9FN/AsnethYfcp0s8vmNy+CAwlNrraF6xXYlN/5sGcs1E/9MSFwyq0fVyi9qWmMLsfW2PPs6qs9YWqOaTjqtc6QlMsR6gi1mLCbYAuPwABc5t1ECWoyWDMmPMUdhIQsSgSZJ8xSYLE8YLQ981fEHeA0V1LS/aayHNQPyzHXM3RBrJwcgAxJl/DVtXu0jfnHMNkfkH4AzS+E3R0UMiavWr+akv70+FIS720vl/jsO1a069xxD6Wdk8V/5a+CjIBXibtH56U//mVN13ihPfuoptqE9ilJ0DEU+c5REZBwCF1nIhK9a31UUnm6LUfksDBcuXCjPPPNsue224+Xprzxdjh4/Sl+rcqa+TFSGrCP3e/qWigmVG9YrMgswZUsUWXki1zJxtThxq6ELzG5TVKHhHKcqJitOjOPCp/l7eZTxGNZxonOb8cQxVUuXOAOG/+m0ehWlsW0l/1OBIx+pOJR0WEpDBjffTRXWyBtaFY6M4AqBsfnSBRQnn/5kjmdXEVyvilfLFL+2Cz3Cx8CRe414y9+ausnjFjHe/+XmkL33zlP70N5vkW8bVKN8cMT3f7k5ZO+989Q+tPe2sFvcEwWG673ralO2wQ027Z03IbaNWRO3b+29K8uux87XPHTebhHDb14fR/l0f/mAvd3823rlrbKGSkEE5DvxGoYiseXoTh7U/LoCHB4RxilTlFF4HPFK/Cb8BuewQziHRaRJmfJRt2TEsiGsDkZDxw4DRTNuXj3MxPvo9A6KFr8FY9mCeak4ceDuReAYA/ZeJMh+9rKNL5wrMyhOBvlmJ6gVEvb3X37NfLqr/hYjFaTmfFRxUDt7fd5wfyDxNSVL9tRB1CULiyC9GWTAILQzKBLHT92X6YTRY9TMwRWkdRiP+SrkXEXAdII7O30AwQYhh606awg0a+z9W0O4Eo9Or5zUOul2C0U99g5oeI6diUjmm5m6SgJH5gmcWwhto3+VtD9+Xzxu9ROfOFUee/hUmZlUyNG6hqTEc1Dq1+IEQck2SV1SDPlQ3wjLZOSCie8V+0QAhw6qGdFMxHcpZQsLDFzhlR/96sXyziWY4cA4UGnlUk8x0by7D+F+U/iDKAYN4MhWFUZQ83BgWkbQO3vu3TKOwDnMil0fipNR2mjYgcJBAUhVusQMNPk0ihPADgrIE+AZrKoT4QHu2wiZ6xxpvMbzivsrUZSMcbTwGCujHm+mQs/j1Zgvcb68zl83yzzWLjeo33WYdu/IRJk6eLT0k8aTdvQdwiaw4JBShYi60B7gC9DAPxnByMVZiB78eQzyLquzRw4MYXHyCU7VuQdtGrMkYjEdaZpU/JKKm2lVolVcQX/ma33Bk9jKNq4MFnxII9hQDEyWI10AiCcYvH3mPIqTV8v8IvSgcksqEjzuSSfSzNxsvJsn/21/28cyl1A8KRjZR0xbhaQKY9otHRt6pcysoFgGEkEUKgyUEXYUeAAxyhSlBRGlMpW0UnaEEdpfrFmILR3QAk/b/yhcvJBG1i64e5f1bfFv3vZYLnEmTvhX8Ub/YtCnIaEDHJwxKXgC65+/xTnsrM5hwX8fgNoqQlbT13IDUEKT017Reaphtf2JyYpWVkrNAoHRybl+j7TymZ9HkcdKsLizL6YixMtwBbzmkfItNhlaAIEiJOUbtwni1l7WMdSRb02ENhmvdbLYBpiqee60fw0ydK98XprC5Jvyz5yqQ4U8yl0za7cY5CJh5SQ8iHfzNzMfP4Lya0ZplRakFNv+/KX191QDm0LmKClLsvLwYX2cHMXHSergj3RU6c8nK+cvQ1yuGsJjHjpv9WMTo56q09CuOMqqNLlS8A4TzBWE2ZdffReHj/PlpT+dKZexKtFZ6QgWbwexTnRlcP7Klfi7WL1BL6Ff9faxLXEYi0WUw/3wrr4d+GbPEgL3E+X85R2OM17DakVYMXWG901jkTaLxcnUFY4shv5H2bo55MQXRYUTbSfDQQL1EjWugDtuOP6Mo3CeZFzbRsGytMgRxyjur3F07QqrgwOTh5koz5QVLFz6sD7pn5wtIxylS5EoDtxLvxLrPtYFwh9UoPgnsU3inH2W1UkV+UsoFlboH66sKm/o92UAGttmC8lQP3v8qd9TbNN56pMPgxP9ejCGA6PNMIBDRslV7OeyQZsAHxPJPpcYaXQeuYdPwH943OL7e/DLl19/F6u6mXLuykp56Y3zcei+jOJklQoNcVynkwC3NPWjkJ5EcaLfGcdjj+bcxIxeS5Iht5yqOEG54RYkt/Rcw5TedqhbdZg8qGDJii+KExxECsoBtuosskVpBYXL9ORMxlEnO/pLyak6Kk4Yy9yqo7KlWpw4SWHyxVYdTde1TNHPQBQnWJxYd/FSJ94iSYSYxLEAUqKRXXCI4gSHweOr18sEipNBlGOuZjsuKm0MQi8bKLZmOa3JbcVbjMs50YKG3EJxEl8g8FFapWwxKVshTt/EESyfkBM5ac8xVFrs7be/QcPd/KdpCkgiTRWYJT4ebL+qlIUuoTcG9nLHkVEUJ49iWXsv9agWqeGjlB/e6kgijVHdkEI6uligoNS/4iGYMJ5l8DUlNtFCPyGeOor4PSyRLNrLqFHyE69jIUOY/TqsEL6Z8omYsgQo3JKSHBhDf0lAvCp3KQc4Riaq0VUCEWhPFiW7dKTNRnHyvWc8VWcK3Dr2koZC3M6sbXfSm5wymlHQN+oIbBadH6KJD4Ug/qdoMxEuOzA3LWoZ4c2Id6Hg7mXFyCP58aj440lzCecurpw7xKYniyoVR2l2FCpyFhWjngbl1vtxLVsJ9wh0twDFKoWylA08NWgDGCODilvLDgh+d+zUGlL4aw0CAwXln7AE58IEvIzFQEIVhZWyvEM3gl45vVUgf8JVhogKGCPfhFmUkDow5IWmQXaw/k17VisdFu/gtzubKE7uPITFCafq3HuExWMyEy62Zzv+2GRem8y13jt3ofzzz36BJdHr5clPfqo88MDDKJNQsoKPtFHgAVRwpCiln6RF+Msyp80cPXgIy5FBrFqvlB/96McoYA+Vz332c1icwYfB8RZyz9WrnFIKvx9k8VafSW41zIk55O9Mwjpr9SG/WYS/aVGic/IxeRSLJCrjVNyo8FCWclFNhYZxVbwrH46jMNXZtZjLf2C2hfTvqByhwkSYRe8olpGOa8k3fZVWAj5PJ3Wh+hrjQXV+i6RIhQdRmniC0Jtvvl7+5nOfKYcOHogFi2XZMjZjbR0yz/8GuQJiQO7GaZ5DiE2wQfsfu97a9Inyvp/O1zx03rritWHevf46y2ceY2u1wFZQb/V76xhtaHvfn1LmXDutzWUHbC6it+1Q2+nW6dvo3m8dow1t790pSPMB5Wf1vu3MMp4wrpBT005NY7WNtq9860J5iRJ9dK2XlAhjlFnJHC8tcBwx5sS/5ZjTYSxOHBAjYUHUuCoiHhYk5qNCBiaXfZKSNHk4Se+Hs7XNssXAueV+V/07ELOXSfIo5fQxMPddwSksju+m8HNRD7EjT/+TrfUU/2Gn3gkX8Dr54N06ECcT2EzI7IwyYyssc5Bx6kgJpkb9RjVLs565kpgn74ZxT1kkJk5bfkzk+B7mySehqQxF2AraXLbjoAhSSaHlyTWcuKmD8njBYYSObY7w2kAYXsfiZIu93u4HRjVPLjqMyjBB3mmRMHOxWxVO1JO61XfwR6sIp6ZveOkKrL0Kk1sr5dHHTpZHHkRxMsU+SAQLrVisu3joQ5DTBE7I64SHx7a+5EmO1JX6U+G6zUJYCKMpRW3wz3etkPT9sYKjsmd//vtyem4NgRPFyQ5bQ6iHDJGiYLBQB4NKBDfSOCHJYBXwrQ2KE1Y73z3HVh0mAkMqTlZQnKC8UHHSp7BkczA6CbWDRk+Pp5NQJ/EPnA4stprm5/1Yqqg42UGzvQk+0FfjJBXzRPKZ5EQKLQu20PCbRsXHKIx+WgUa+S+sX2d1cb1cXWfPPZOA6YPHSj+mhJvQr/t9dfxFidSt0dKnPQiCtsgudQyOgNH8dRy2y+B5ZHaIU3Uej+JkF2UOEBNfbT9pof8Ab79xAAHC1NdPPBo3v3zecU+29SckvoS0OjGebaxwBVy7nKrz9rvnc6rOwnUURtBjhBeyrtZkSdHkauL2Mq0rWxTA10UsTjIYRnECBSjs0K6ShsK2fME6Rlki3coHwCG9GQClDwRlBKBdBkOP5pXuyJB8+ONdnJITf+SREoErDeo7IebNt7wJMlGdwEtAhqX9jeYb5RtBGosFDoJQgsjPL04SXPntoU96qpGKk8dRnHyRrTqz0+JbKhQWOZ2FmUp4mufmteU/fo/yx3DjWX/jUp78JSdiqWQD51Q3/b8eR6zixHhOR5OSu23Oc8ry2TK57HCEWde8Wgz13+N/fJbomvIDj41DFh/M/5NVza+m5LeFxLza8gkzb2GCJuVvEWqZLDkx9NjTASaHlh96Edm0f/r4R1z+x1V/2yYol+b5V53DMoHMqTrNccQtpkRHQ39t/6jtkAhBYdDla9CZn+YjQZXdBqWVxqVPRXF4LBmvcfTvMz/+XfndS29xHPE221pYFRybZrvLIRTx/VghXEPwZSskSo6hwYlYKazCVxTYbZ8hJ6LbWCpcP1cOHTgOP0XZOYDQTJuoWNxiNVDFyQyKk9m5N8tRlNlDCLM9tOcazke3dMYHTaaPksZjx7VqcXvG6DDWDvgtcQK1hnXFGpZ5Wyjnr+ELY9VxZRzLiFGc+k0eLFPHb2MAxEKNPr+4vBbLCU2v3QLq6qBWfnCqwA8TRUF1BOsLhG+E5FWULF7yhg0aR59eKhY2cVR6aHqkzIz3xNz9gXtPojiRX4NDJ+9qZLbgPfYV0c6fdGiPro3hb0PLFpDL77Z6bSfTrlPeBXjei6+fRUE/ha8ZlChYnDAzRbHFqipjplt1qnNYToHAse0UypIlcKIliY51t1ScsEI6hCJ+gAnKsgoWJiP1OGIVJ9soTti+lFVat+o4KRmM9QhQ4xz2AFuVUJxgcZLjiAm7wQk7rgyPYHGio1i3VB3B4kQHvhtMOscxc3eSswaMQypOsnCzyko0JyYBk3VL92yRY5XJN9THc6/CCe0/zLg0xfboHEW8gdIcvK9j1bPGtp0NFCjWR4Z2bPYoJyvhZwNLJ9Fndm4d3mJsWkKZtgmv2KRO+r4ZO3gbPmJQ3vQqFyB/QFN9HCMd68lAISzmICH7t8d/hNkOGnmBb/avLcrUubcWJ1/6m8fKI/ezEEGejqHpf85++d+DLJr09YWwlhYqrydBp1Ql1VuVnwybsSbxKT/8tzP+kIpCHTEitxkOvDDLyCiW6aKKdbBa/HLVukp3kT/ybphykGACixUxtjDzTG/klfpRN7eyafW6tdVfXnntbPn+D1SczLAARH7052zFYq+UcKgkIIOaXfK0hHpVy1/jtSHefdnDvwtRrZW4X1UIWH/jBX74hECnb4Fsxz+LrIoI3psy3bolz1RulufaSZUBjL8NP1DJ58R+jNMmFWWjOIE3ZLId+Grd1rE826Ict5FLE4oPYoeAtIGVqfKPd+CiP9eyhDYBVA8ZG4WF2ZJr6lFrYCt6Ne3FN3N3nK7dBtpN/Q2uqdWTeKKSJEdlwU9dJJbuerGqUlqtW3UOl69/8Yny4L34OKF5xQ/ZUhfihU8po3C6DArYF//wB7bbvFnuv/+Bctvtp2K94Zby+IWh2Cw6Uvl1ZOQlxmH9h7hodWT2IHOLkbKwMFd+8qMfl8NHDpZPfeozKCZGcnKOjqL1qaRz2REsWLKIRD3qdh/pV+Wnbea2ShyDI49rOeLR6cMoZKQn+65+dHzW4lRlj9sUtXozv2qVwtZS/pmPYruwyf+kW5XN+l2SVxrXP62KOgszKF7c5qNFi+Vfxy+X1n3jY/BI+J9b/s+cOV2ef/635atf+WJ5AB8n41gAtvJH+L/NQz2q/MW48CFX08I3xWhD2/v+zx80/5ZcJHcpo5LOrdN353brGG1oe+9OQdbSTkN/Um/Ka8r815YPXZp9U4km064a7YfkA97a9GYknO29E72N0AnY/7Dvc+el87A/8i3e2phtue29E7WN0AQEoTwb7LXvMy+pQ1eoCE/cfLDjGommsCCZDBFyGgqc0L17ipZXUJxkqw4r2cMcgbbphF8+DxOQsaM3rQ1rxyF9judKIQhktGovHcKmNmhHxQl/WpHBkkoPigzsFDhRBfNQFCdjOLGbWV7IKSkRNEkfeiFtrlCJ8FK+leHq1Innlv3Th+hE1IdSZQ52MicBHo1lrWUsadzESM0JBcbggk9RODAQ5l7LIGotNIXz3JRvJ7dumqTF/NBBQZM3BB+vYVbiejFd3WQf+OYUwhPmzNsIIznmVosGcG1WWe2DG0d5QUWSK/c6Ra51sVBYnQDywXQaUXNhQbHLytxDD97OVp07yoFJJjfAKYOUWzuY9UdxwkBMWzvQVPC5+wyusnJB+VbPLRkOZracx43ZCMYxVVXuoCSi7O//hK06713jmZUlTLZTDvH7YK4y0WzVIUMH3xxDKZ0Buhp947pV591zKMs0cWZQ6HWrDm00BM30wmhtZMtvLU4w4q5tRB5eDjyemqMQqOIEyZTt2a50cowaDHcRgc62mWWAUdi+Pl+12aMMDrMM2ONa4VC9JSxOriC04t2jjCEgTsxiXom2HT0LeGFQ8Ng/6q623gEgowMTk6CXEgIO+bSU7pacHSxOjh3E4uQLWJw8ei+rECgHgTVbddKcpgI6B5o8qXEHw7StwpIt7ZHNNA/hClH8Bx9VcUAKG4r0WcWjbdxa9BY+Tn7L6Vfzi1i3YBWkVYgCif0wk91aImnN1PSBgG9ygkr/Swjwrv5mD71pobMMtAjRVJm0/CA91O1EQintUo59jCwjsEs8Ch8MkNmDDT2lP4JL+cc2sClAUpDV4kZ9+Kv1Jk+LkG7yLbVMu/OTelt1H3mpD6mKP9BN0tnv/aqyCxiiOEGxhuLtEzjq/eLnH8Uqy7KloGSUvDr9X3yZj3m05QhNU74ld19SQP1HNWBukG8mt6afi8UJdCs04ISgCnbwb2gto+KfZ8oQN5lkcE/7N4XlG3DV+puTNRQ/9aohewFp5nytdezE7TzUuDVeF/+znhBmLNVof7fs6CxukBXAQfqq5YdoKV3a+DjK/7jqb9sEdupkLaleeIPHEbtVRz4le7X9pSFJs8YkcvPO0y0v8WCeuXxxnCQf+4T9ptbJu+MRJxGgOPkv//WH5cc/ewELB45vnDrM9pzDKBjY8rB4jW1ec4DK9lacsY7i02AAhfvVpcvgf4uJ80i57eihMjG0Wc6e/hP5D/BthO0ys1hJUAZ80zFkBN52aPlCObFyvsww+bT7bjtpwVJBpXImN/JYYNLpthTlEZljKAv6mKWsoXzexIJS5eomfPUafHrNmRLWJr0omUcPHy8Tx2+Hh26XBVYtdVy6AR9m9Ge1UAsF/Bnwrl+SCbbnqIgbQii+zkroCj4+7GTun1dhssHzDmUqrI+O9pW7Tx4rJ2aHy8mj0+UIW5KmJkagP3DJRFre1oOyWHx3JouBHsSLe7/YhrazbUqYfSl8wTYHD7bLKnW5wsLNC0xIt/omywVO1Xnp9QsoTlBYUCeVB0MoLxTul3EI208fUHGyyLOKi5lpFkQ0LdciC0sZV4CXUW7EOSz1vRqLExUnez5ONJc3ntYjgnoEi5Nr8F23+MwgL0gbK2xdFS+drTpYnBw5xJYe/ICpOBmbwFEuldlgAuNKsjKNJ+9o7TrFqT+Vh5B7OqtIEiPyayrPFbkMOWV4c4XFKhYs1vAvhuKkH58JntRxg4WYVcbiDcY7T9M6zAJQfKqRjyvBm9BgHRCcHC2ypQmHn3xbgeanjt2Fj9GDyBzjTMygLUtlIaXyZgpPowWM/NTXyn9soMCeMRDuzMecHONWncNj5UsovVWcRPERerV/WTnbWInVDs4VHkpYMMx4lPonmHGmiWvBXN3lm9V/l/+SBkmq4tJyzSQ/Na9guCmvfqAE3g03WiZB4f/kAyG6HcUyrYM8XqEkSiBLACDp1C3DG5y89eqf3arzT2UQi5MN4zG2KhNIC8q21iWX5afYlFjLNyB/xDCYy/j+1fGHX8UurSJkFMFhjZpceLfrU1CtTCdtV5wmY30QMfynPEjDrDLRFtZYnECnW8hto1icKGPol1BmrENsZUPldmVSF2YCc8qCN0AXKhLs/8LkJD1O68m/ci/eeRbHyraG7bBtzEMt6vgLMHwXW36VPojGr+n4JdPaFmROZcO7m5ar4SgBZSKk8L/b5fNmHsxt+tlOvru5WO6983D5tzqHRXHiNMl8VZxYTV+UUVUs29fPnn2vvPXW6XLk2DGUrTPWKuU6/qqsMMk6C3vyizWsBFW8TjGX0D2A38+eO1t+/GMtTg6Xz37mM/DucRTuyrooVd0WgzWaY0fqZu7gWblI5fYqY4EWIW6j1Cp9km2ccRAO2lW8i2XLV0Zz+80yloSrKIn14aS1nc5qzcd5hdYo8YFCKvGvIsQFZMcYFSaeBGYcrWFUxql4iUUzbX0DeliC/+1ShqeOjWPlJ/9bRYb//Yu/L88++0z51jc4VefpL6FcHq/yJHCBdspu7sIpkr2C530PCf6wnzZJ8rM9zdcC2quN0L7fdN/3ufPSebgp9vtf25htue29E7ON0AnY/7Dvc+el87A/cvMWxUn3lxp9f6J9Ye+Dqjv1rZ738soTP+mgt4pK2L6ymjj7wj6C8ltGGeLphoPWrkoVSoRoKzB0AACurMTvpJYqmsHHyvgvXJHn6hwWixNP1fkXfZz8CS33EabtWBbIeywDfyUqTqJ4kclQ1LaEyzcZAoYlDNAKcMSVWTCwbZOmzyMnGHh1AOpGiD5WU3ov6RyWvdjrVzmS1vWprisVrNgzNE/8tPhvoA70fodPBQalIxkiL9S37gfWvNZOLH6Es82jUyDRHTPgzzDIquUWTSSo+fjInz8OeS3WXIHX+ZtMIUoaJ/IIWT0IQusjU2UVzfA65W7BFJw8x9s8E+NcgUNYgFf0Ub6XMITh+t065E8xhG+W7Gq6EdmGsLO9zGr6nZkYzrIy54BLc8Co6iS1HyduCiC1vlbS9OTCqCEj9T1/PPbKZAkzKDTiyJKr3mHnNnT5h2d+Uf741hUEdSbpPZQJPK7guqfZ7TLxBt8MAsidIS0Vc1bSPaM3aPd33ztTFSesZPaxz3qUuowAZ4/Cv2VS9jZ5Sree4uRlU2id6yDscc59DBC9bsUA99h4E86+ecyIr2K9pNZ+wlUNsLcwd7WsoSCZwUz82IGDWKiIUwR7zN/nmRgusLLbj4A4dcCtOgi6tP+OWj5WGXWwa/2cFWfVGCBi7VGhDF4rrRHOEsM2K8PHZgfK0/g4UXGyy6TFjhOfG6SRHpOUeqR/0F98smI5uhHY4vfECrtkAaRNg/BECmFXCQmxZFBjq85bpy9gcfJqmeN4jhwpSJb1uF6qQKoM/OAx5YX+HaDJ12f+e9V9/GwXYADNYAiNMMYCu0IpVxuVe0BOKG1k56ethE2aSf8ymfzHOPzfxCHmNpPCXYTA+NWRfkmj4GQh0psaBwfV2ocpjjD/LLjiNyCk/i0+VAzV/gzu7EPUydVAq9aH0ksfJ/oBcv/xE4+eYtXyEUzipW8FxKbizS3VaZ8t1svyxVEe2vK5Ex7QwKm9MjFUnPBAtVKPefhntur4TiuIoVAzvMnE1tm/ZN8EtIosgpOnwbcuv+Z10+fka36WlZfka/vXMMtv+R8Rusqv0S1f2TBOpRWYgNWjUhVk3KpjrqEhcJfVMN4Dv4B4cW/L+kvLrzUjnya/7gIMs9ki6FLY/5f6Z/UYhHs6gO2P7FYVJ27V8VSd8C8LTclp49Cf5VGSn/fKr7Uz9s1XUtflxZqT1lDm4LjHP7dxraztlv/7v36/PPvT36FgP1aOn7wfB+wTKBUW2eI1Fxir8gFYUUzTWUhzPVt1Ds/MlPvuvbMcmRkocxfeYmztK39+4xpbaXaZ9GL3iRA8hFA9ho+T46tz5cSNi2VwiRN5mOAzA8sYU1cgZUsqReg7xM/+c8YosbqOMLuKX4+ssEIH23y/Ds++wfPQ5KEyeOBQGZg5WHZYzbyIj47rCOBaQkYZD9L0aSXtSCM6gRzR2hILBZ0Jbjg+ppfTK0ijZ6htxhQXFNwCrELiwftOlmPTQ2x5HCnT42x9GYX+hjXPp7fBIHZZhRdO/9Gb+G0u25eyMwYyBqcBIJj98g+wIatsIYvM48Pj96+eZiSYKGevLJc/vnoORdYoW0nZzsSYMoKSynHNk3TcqhMfJ1c9VafCuYXiJIpFt+qoOKGviNsptvRcAy9a3ExSf51BrjLxUHFvvCUWV6ThKE5wIpvjiPGvZedbRpnjEZ4jyA1ae7lyfNjjiDN52qjHEVO+TjY9ItyV4mzVUXEyM0UWFSuR04KbWn9pWH9OjsP9bN0dRLE2yWLVCIsW4yj7Bxmn3Oa7Dp2uMIZfxrGkJx5NDeBQn/FQK5g+/pxUOf5odu/kbRWauwqMc+By+sRdZWgCRUuvvs8q/1BxUntQ7dkd/kOjdfpvYthHiAvNtErZ+MVBwXPnERQn2apzD7AgyzhW2qds2+TiqJEBlb5j/b38dQBr2j/jA/Ti+CuTrv8rezG275RfoSTAyyyEyQdu0ppXxjnCuksyPDgnvvko/daLhG2bECpvtY4eluAiU7IPVyW3+kKsOqrUBYgh+uhg+RNb+777zD+xiKmPE7KAJwuX1mL+C4CC6TeBTV8QilqjWuXaP4zR/O/UP+lk/nwwfXK0CoGp1sV6xydYTW4WKTYPKRL8uR2aD1FyEBZe3YzxWmBphaCF2Rh9REXcBtsllUf1tyFvjtKe/tmzCRw2SvIAIutjJSgMcozsQ6LIIkKdb+Qjt620j7Npt1t76hc8RoWUNbPvyqeyPT6DiZXgIuv6aiwxYID0YsUo0Dt/yp/1ckyHJoSRvD19bHdrqdzDVp1/q3PY+45HGRUOJVy0k7m6rV8nux66cOnSZU7VOVsOHjyM8hXZSDkdPhPLDPJeRQGxwrY+FdDtMefjWARvwovF1euvvYYV0g84rv5A+fznP48C5Ujwa7+cnJqOslfcagEi6uS3zlHW4MEeC+xJN/pT0lnrKP5VrH+s84FDRYfWI85t5FcqWYwbBQt937r7TbSowLVfewzyMsqVNfLOIhzznlGUzz67tVzrEi/5CgFYyeEPi/ha5KlgHpenyGcA9houD37z3K/Ld7/z7fKf/uN/KP/r//Lv2coEL7IZRDn5+NxecUbMi0HiucaoX/eFCbBt9j987eWVJ37a8m+VRY29l8Y4+8L+SsqH1whJe3UB3HnsPCRS91tbh4R1f2iz69w/9GMnVoui/QXtT9v99peWnxq3VGOGzSXBpMP73hTUhkkqleFDvGSQlTUYZawTZNxKJBCs0yuZweW5RnHCqTpDeNGPqoM8K1OygyMACQgDpPlmK4GF8u7krgdhyGHL4YYuw3cYB0KqHb6PgTtr+ZiXeqrO2NzF+DhhiCB9EqVj+FiZmEyHtFRCpYJ1rMIzz7ynbpYUpghzIDCCsgyS8nRk5Hv8NyTPmsZxwj5kfhm0HB9Aj5PyOlTwntzDjjsFOUHaYsLuJ4emKE0iKPIGc+pjFWiHlYFlNMAec7uJQLrlxIoVHFfYcioPKTNOOSfvaEpsGwcn6hkBACgYRBzGIn5bHgDrMIwGQ8MNDJzi8tlP3V8++eR95QCKE3k8qcmdjI2P5iKTf8yp1eR3Jh3Eyxie+hNPRDrieTU3HwnlR4ZnIAMQAf/w3V+WP7x5EUEdK6Q+TAFRCA0wCXaSuo4w6R7VJAT/ZmkaBxzUBwxcrPYxIMTHCauRHoXYq4M6Vrk47wb/N5ZFGtJWixNf8N3RAJV8xKH5wnA9WWcEQnOLzw6D1w1oYH75GsL/KgwdixbgWlnGeR7MfwpFymEGmWGK6APJuwjjq+DzzUueSjFYZo6eYMFxEjipJ9ZBCvWbKHJUdlCic3yUag0uqJS4kTKFzTdPmtqiPY4cGCxf/fzj5Ul9nDTHyWXwIKptqRl8BmAkgErffiBfzJqpRPKrTstQCKWilARN2rgRXCScPr6lsRvFye9e5Wg4hF/2poPFZGfuWrsEvtwqnNK8pejx3gfpfxkfJz2c5jEKzdqpVMoYzwixbjGi/wUFWNzf7IQkz4YZVRwEJsK9J73TJGif4yq3WNGOFVLylm+g/FJgCk1bdbATOJMZ4dz5FjGuATp4of0Nl2aFx1bI8k5ikogwrYRUdvVEkNvgaGh9nGBxgouF9C/bz2SJXvMCbW1FzLS5BIKITf3Nuz6TnvCskBlCn3T1L/5fiK5z2FVWhb1i1ks661M5iXyXZ8L2yidAwFJpEyVpjdBVfsV/EzeZN/G4Na1bA0xPtDYs6Xhvgv3UVT7x/GCbArvCVQRf7iv2I3hInMNarDkCo3ryXC2cvLRldcL/gvI/rvrThdOF6pYZLQWhfywLhrFkOMpWiLA+kQJNpUo816rVWu13Dpsa3uLHFMR3XYErXZYJUnhuiI1+w/cNfJz87//nt8trby+U0Znj+NGcLgtsAbnKFh1Pk5CGtF5zBGVhMH1kcESnoyNlHAXWcUyzT55wGyIrdVgvPvvjl8uVRcZShE63TgwjyJ5g//nBtStl8OyfS8+1K2UdAdW284hhHcVaS03hHVkGGJ/cxqoifQU+uYECwLhwQeLAh3HyqkXBNgqCyUPHyhCWfOtM3OaYrF/m1IY++KyKE8da4ReHg0wMxtmmooWMk3n9n1xDoR2HhMoaKCZ6YtXHRjrSDDLBOXH7sbJy/Vq56/ZDbKcs5c7jM9TzACuc7tWnwaDPbfhxHI/agYLkevO3pT9XOq0TAXthNI1dq7YQxdPxLl1bKX/482msJsbKu+evlxdwwOkpcUtY5awxviv0qwhZYvVU57CTUziHVXECX5medQsOyiQtsmgTFSIrmKe75WDSU3Vc+aWOdasOihP8lWheX0/f0TksPk7wVbBI3k4eZvBxIp04Vjl52PNxso7i5EC293j8sau3WhO6Eu0RyCMqTpgA5VQdtuoov1nLYKcyC95pb4mRS8vGPtIOsyVnbPlqmcDiROewA7x7Ao71XkKJMoc1ST/j8wEsmbbhY054FbWlTf26DY8i1zBR0mJgHnqZW94oh++4u4xitTkwyBGmbN0J37Yz2A7iHxDIIrDJ9zL8cLdNwoxoX+WpTJgJqhYnbtUZRnHyGCcH3lMVJ2biABqmZQbQkz2L91rPJlPooEMTJHEhZofFCL6G70qGlf+GtOqPROJl9t6N4IvhzSfLcAjKFeDrJ8ejBvMd0Nqw1J+8OjGQ37Q6MUvlDFm/L3W7DzzDvCyb04nc5vonlHrf+yFbdYbZhIefJDpYQNJ6wzzNpSt3XytQ5CMvDya68C+/a/GfMTfKXhJRz+QX3PoqhHXUijWbdU8bGVd4+d7c5W760+sPM+Ut9QN3LiQSR4sTrZlsV32cKL9sMaF2fJS/7qCA4BW+puJE8PlHHOcOkYPli0AT+Ub3AJTNl8AonG5d0oKuNirzAOi0D0Woi7qpFekDk/kR12czrFQnvq1T27BknNKEi3Dq7HvktoT7vaZV9kAFxIBSLU7i4wTnsFE8A62CY3BM+W4n1zpiE4XRW2+dKadPny63nzyFstX+rzyBwgIeogJDHqI8PcxWwBmOpR+E/3rQxS5ttYFy88UXXyzf/odvY4UyVb785S+Xu06dou8NRwkxyCKo7VhxaP2tO3xDZS/9VWWFPNtTuUZZLIs1GbLSJkpeL3022Q+1Zlv0WHnSz87OhN8YRyvEWI6QR+aE5O1Wxasogn0/MDOLdRyyJPTp9qOKQvAHP7UN49aArT+Ox/LPMdrKk7qkU4WQJXy5vPzSH8qPf/IjlCbfKF/96lezjce0tY/ZJkBVb5Rpu/liDK7OY+dhf7BR+ASoNer+aIm79/OhH/ei7RX6V18+fMnq713dVex+3ovRPtWvHx6njdviYS92nvZeOxG7g7qfOxE6D/Xrh8fpRG6apElDlX2S2cq3QisGcCVGpa58k7AqYfORl8rIZRIwjaTJUMVH32USMioUJ1qceKoOWwCGR49mi4a9Xw2mWwjYAEP55iMjrenCh0gbE3/3Xib/KpyF+TGXh90zmdF8HvjX2GIxx5GKCxc5Fo8j/2CstQLWZO+yOkBHKZbkb43WqX8TNZYVrmIDiIoSe4XkocZWba57/oIuw5uCpB74mVVPphl4Zb6dq5bp0FxL5pe4wtNv3WXGjEAOgipNNE0eRHO6jVPSVfxlrAHPJqZ16rurnwuGBQaftAl5hCljoqPyJBOvBp4oTtrGBd5N2svyHQRVYIjVXuq1w3HEn//cw+XTTz7AsY3sBbddvWDAJq/t70BMOv7X01z4zgdrVuOQG3HNuw6UFTv+pnsBUxyJmift+p1nfonFyeV6HLHHDtKqTq5QQSNoAxN4yNXQgotO0kY70VZ7feY9TtVxpQ6Lk37M8rQ4GaKN+lScCDd47WzVQSFiraRdFQcRDrkPwpD7GGn7YPirTPyXVZhAmxue4EQervDJirdZnXPbgcqAcdpiFA34No512YGZE3XmOUZHRdcE23UGMtAiyKg4gRlrcq6m3LSeGiT9evmr0KAAXumMtkRxsokF0DFO1VFx8njjHNZ+4uAprnWObDtaF48XdoXMwWKHPfzJNRUlYpCg0Gm8vTpHpSd9R3wgPQL922earTrX9DvDgGlZ5KMAIrWYV7IllXALR9u2NbxanLjCNc6WMr9Z3yDR1NCF9BG4pRGeGvEzvCIOzVKM/QEIoYUoimjDrB6ixIjiMGmtLZlBIuYbgUd4JB+AC8/gWTiNVi/gFw8pmTQKb/xV/mNMcArdi2brZjK3ziizuJ1tgAnB44+hOKGfHJiqHMS8jOmvODJdLZT7TZcpqugoDDVNUoNktzkQxH+UQzyKNuFQcXID03q/9CAYVsybMR9DD8mhPkJD6XcG1WAjNlcts22/7vJv5n/JWliaurRZ1TS+BdQKQhBNuQTTTJ36yzcVKG1/+ZpbBwYR5NyOYM3DI5IRbUva4C0589O8/2vK/7jqL42mnaibVdff1g68ZhBl/okjB5DLQYJIBgA+J27uJvKVW235+u63BoX53v4kjXTH5cqolFMndN4JhG71UfBf/p8fMZGbZZvLbjlzbr6cuzzPRxSXCL6e1uIEZIC+vLa6wxaM2XKQbSu98LUVBNPZqbHyAIL51IjKj14U2b8tq6xKD2ABsoZAO8ME/xirc31Yct5442X4KwIqdR2gf9tmMLXwV5ULCq2uJOoSdA0HputM8GUeIwrO4MPJ+iZC+CbxhhDQx2YO8d5frhG+iHP3Re4jrBZKK5pbOzlSiT7FKqbODsWZAvWiyhi3jTLJVeHv+NVP/VQ4uvo6gSXNsWPHyzuvvcKqJoqSslI+88T9nKjwcE4R8mh22yjjl6uzYgviy2RHxErIXOktNkzzLYs4+cwPFfI9KACOK3NL5fevvMFEdLycPn+1vPDyacYcLEJQnKyrkM+pOih9FoEdHjbJRGBFHydMBqawjPFY+xWP9nUcA1+d44jZquOkJxYn4MYxxBNyhj1FjtXW5WzV4VQdtuBcY9zSQeO0J7qB7xv4NXEFd09xwladw4dyJLLOGV0ddizztIph2k1lTLdzWGU36S2kJjaCFvilITxn/GHhYghLjkm26ih3TWBx0scWnQ0cD6/BL5cYG6+iPHGcvePIbWUXPvbu2fey5dlTn2ZnZ5nAI+M4rtLWc2xRunj9Rjl4+91lBKVa3wBO8XFoapvQ4uAOnNs+FagObAYInvxX/hpeRJhjv1dVnNwod+Ec9gtYnDzywN3Ux5iJTTrShHnZ4dJ5CbH+ta5wq5RFCOHkJ14afl17eS3H8hPRSGZTg+sNoExfg61Dfc64xWv4n/GbtL5bvO8G+diWb4jv+cZkOw9knG0GJrIQ6udEXLpObCwndQ77RyxOvv/Dn7Kwg48TLJbdVm2tqm2KJdUCby7f/h6ZsQLSwGPsCl1+/ZH4LJrnlG35SbO//ipzM/6SU3RXjOGmSU25YzOC8tA3LcRSoSgTfXILiT4y3NYxwjHdysubPGexEZkhC54x20Sh0uYpXvhjHZZ4FoM8xy3rSwCo7xPbwr+0E/RQ258EWF33aZ3mWGZ9rRPtqUWa6qa8m2XKqN8kjyCpuQlG5EBkQevZT5nSqVgRt9vk7cISGw7Zpr3IccSHy9fYqvPw3VicMOakv8F36/gjRVI26d67eKn87vl/Ke+8fYZFz0+WO+66J3iAVTBfYQEQhazKE60+prBaG2bu4jaoVcbjIfixW3d+99vnyre/848c7T5Tvv71r8dXig5eVdZqZeJWGqCM5Yn11tplDT60ihKUyQsK3qkocqsPP2QZld7WW3waXytDZHXpR/9TE1ql8Gy4eascd+uO/E3lrW4KNmhjT/6aweJF568ep25eKlm0mnER17jXGRMsTz7iWOFJO1nE57sLlW5pusRRxC+88Pvy1a98uZy66054Z5UdQZ+tV5VYoVGbrM45DDfIq/u5hnT/1q8fHmcv/s2x836LxN1B3c97ObVP9euHx2njtnXZi52nvddOxO6g7udOhOahozj54EiyiJZ1NSjtRO483JTvB4Ubrf3W3mvS/W/d2X0c5QOFvZdO0cLTluK7tfRrvfwiwTUcQebBY3UoCNOiI2TWEq09sWAQTlIuc0Tib55/g79Xmq06TCQVPokuo9hRw5qSZN+aHlIqTEWILE1OF8GamDKpbC9A8ImJPB0NeRVDAsz2EDAmMeNlVywTnJZrCWDNb68mvKdW1K/p2DfXX8acCR+4kQkbT3PkdbWo1M+jVs03x7qaPbDVvNq8CaP+YiwTGWpiHK+KVccWnqhk8teywhiU5eqLebkK5CRjC+uSFSJukPWmq9G2F3uz3dvnUcSaAKuAsizx7QQMGRjMgSOYR6/WCM2Wnmiia8yU0Yt1RLZ8UOYuipPPffpBmO895SACtWf1CJPOsjIGkbfCrEoecSOMGXUUZhQkxVFGyFrDWu/UklxEEnAHPZWZKsg/++Nfl9dOL5S1HVZDWPF0a4vtqrNWnTpJQ2Zp3hYX4VWGSH2kFfc1nn3v3TKBc9hhfJP0wkRzqg7pezNiAhcCmasHVdBpaM02DTwwSZrFLWH9MPA+Bpgb1xbQjM9jnolyA9xqyTAKU85+frDikWlrN9TWb2J+7L7w5XJ58Uq1gpjASzkn6gyxz9QhcNMBjxVQ/Qfk9IBAbR1dWaBtqJT9r4qFAEEaMbbbVy1OjmFx8vQXcA778D2Ux5IxUUKXIlKck3/07kEtNAb+7B+1w4hsLsuwLYnPD3lQZwjEr1EYaJ2lcoAtdG+fvoSS85WyEB8nbKkDb7YpVMb3Cqcijf0lirpKaLSVihnbn5OFdA7LYOre45xcoPST2ppDOn3eFapiVcFb284xvWTgC3D82PYxYSWOqVU3ub0M8yey5HvbvfnS0p9Qmi40Z50lnGToFx99zxM0oPDWfhc/ItL61TiASHmt4gTTWE4FefKxUwjfj7ASYr+ixYgqZJZf86p51zxaAJsyc2ueSVXTVLjkbklD3SBxBDXy5T43j8UJEw5zjcKUsqxgJnrQTiYTnfJN1ICfOnSXz3PCusrnvZZOuiah78bwamPWN79IN7X89mPq0CmfFOlX9ebpGnXfMivg7D92BXyICb1l2QzyqOQjXXzE5deMP/r6x0Q+QNPm4LofWveY2OFhfJwcOkh7NFXJjfKl09AYNaS6VSzjY1PfvbthN7053IhS/uokiBt9zdHRhYYNBoUXXjmLsmOk/PSXL5S3zsxxRDG8BeHSE8DQLMTCQr5wY2WrHD9xWzl87ADbeC6W82fPlFO3H4ee74KOEaI5BeyZ//ZiWcDBrNxpCkfkd9x2An44X66+/XrpvT6HQhoHfQi4Ctw63NTpqE4AJ2fGy+yBQ/DvXYTUi9kO4rasUZT/Tl53UCg7dmyjFBhmYj8JnraY7FzET8l1BPgelMyrWHM6ILb75+VzoyhkpvG3ocWGx5xfvDLHxAEfKI7DzLYsO8fdavmAsuf4iTtQUoyXc9RtBSuIXVZuN2/MlW88/Tfla09/qhw+yv7/dE/7mkof8Ci9yhOla/EqLVei5LftDUnEm1/qswSsstqlnwtX5ssf/vQWyqaJcvrsAs9nyHKYPfcokFBcac1BkzRbdbA4YUKy1KyS6mugOn1FsUh9VXZordPPuDOFckHrIScUTmLcjrDGVh231XgCzxKOIe1HB3EOu4i/l2XawxMxrJJHNHsyj9axN1aY6KDkqKfqeBzxOpOSut1YCxBPtHDLjM9OcMR5p67SnnUmU0ebXBCk49cAdD/I+De2PI9z2EVO1YEeCnySttmk3fFmUy5yyuGFS3PlnhMnGWPxx3CabUzIUseOHcHBNicAyezSMbA8Av5zbE06eDtbdZgw9fSyPQLatXhbxslW+A+//rfJbuY/rjx7JW4YOP0TRVwPCp1TR0fLFz+DxQmKE4gjbWl7B4nmR1rHSnEqTHutXTt1y/+kkGzRSDlSRB3La0dt6COdtuat3BJmHqCTOVGNR9rcW6ZhXtSzyaLGJFquW/FfcaI85PY90iJDNdmSgkyUC0Wf9M3kf6Nrq84QBw5ssfXVxakoDJCZMp5aflOid3tKDTA/FKZNATTF+/Dv9nodR9f01oU8LT952C5+aetqoOMvN+AT195rrNqvlN28IvMKlxYnxNhCeavFwyZ3t2a4UGM/1vIh23bN0j8GZxUSyjBZ4M2gAy4sKqA01NTK31TK8NAU96rc50GgkCdpZT429YDGgy/5u7xEfpC6+Z3SU2l+zM/y+bftliCkKYLIppX/fAt1czcMZQQ+Tu4/1Zyqcw+WhAjg1IRopIduVYQMsJXSdn3jzbfLD599lq06p8sXv/h0+cTjT2ac3aSvzXNajlsAPYlm9sBUlAvybZUl1nyCU8+WsdL+7W+fLz/4wffL8eMnyje/9c1y7z33wnNQRsBflelUPNYqMRdgYVLlrTzEyg3BN6bZ2hc+T9XddhMeDv83rpYmbjfcQolz+Mhh5k6cIga/9oRSlcTK1zr5tWks8wpbTFWEyIP0BaViJVYtfO9jLuTCqtxHP5A6gjUfFcnTKOP1rbKlda40Cj3saM3CqTrC9Pbbb5SnnnoSaxPmNmFj4rS2S5WHbTAKCY/h8X1XemZSJJ7fmyRdDzel6kS4Kbw78f44+9+6k/11ls88ShKRiJurU4POQ/vlfff9MdoK3pyP78RMz+TGYxh0J7f9ufz3G6WTcC9qgv7Hy685CAhP4WASjs/++ew3/nzmqr/1s7/+87tMS3KuzA9mKQOCKarN3duqg3NYTtXZZK+1JBuylbFjBVCvGhafBsnWHM2emE35lpcVYzQpDpAep+oWAhfZ3W7hhFvWK8iIYMQ2bQf8lJpaKDCZxoJbRkjcGgKDY2Cv2vsAkLgyK7W3CnTT09MM+OTdIKRNaXbiRdaZVXBL9KPCozchsi4wdOUFAXAAivDAt16YjZ+dZMdqJJN+2DKMxUnlKgKjeBvDsZ6DjqavTr57MbuxVK00MH6DD6AsQMjSbE4HfVZRIVETROFxHLUgJ44yayeEvWW1PPnEPeVTj99dDs+4T1HsycgZlEwPXFQD+d2f/Cetl2V3X6Qjonj3qvWvz4GScq0j8mT5zg/wccJWnTWcw+6yIqJjPwfmXgdxyq7mktbYAs3TtuWZ0TinSsCUzyAkT+Q44sbHCQPqEMjti+LENmLIAd9ZPWNQCv6tjBKBTBIAQ0MI9n3QUJ827QgiEYwwcWnpya0d+cfgnAqQzh2arqqvktk6DP0G/jc2OTp6q5+tTgzGuwx0mo/GcoP4gZ80ChQ7tKXCtBNf20aniA7wbomKkofjEw/P9pevfP4T5fEH72ZVm1MXGCx7FICiqWdvOoPWJjioe8eBxnpBu7to1CWwLXFgA8Qtu+0nBaqRR/FAfXzDwzJ/wIji5B0UJ8/9/s9l/jo0hF+WtIJRELD2WpD2FaekqMICqKTNpHdXga8xwYmTMPzAiHPpR2WoOFAQ2qGs0Ll1NVNo0qaI6au4Na5NLG3yTbk5Ad74I0LKMrE5JK7B1jOXdGsccyfM4BTkTarixf5I/V2dMJpbLTx5yNWMCIaGudLCoDzKKgZeJ8G/ipM1juu+o3wBB4OHOSqanlGLtCToQPxaxgZtoBUCr/iJYBtaV/mhP8pv+5QARqFJmtQIRY1yigoZR6T5BY4jxndOrRM4saz81V/h96kWUX/rexDX/TGw7SvfzG7B/wwOYs3OjiySfcxvzbJ5zG2vfF/FvzdxLB+VLnAKB+/UZ4aOQ8W/35M9bVxz7r7X9DUC4X9B+VK3sHQh5yOpv3w74FAn4bcPy2uHhxofJxSbtkrR1qmJn9h1nPRRcg2JJkZi8WOOzWV6SUKebTjtFL4g4fBuTgrHf3j1fFlY7sXPyW/LefxrcLAXfY0xAl7UDy/ZxJJBq63jx5mMIghfWbiMcnMe+txkKyCr/vjq2cSHySCKk3ff06/TTjl04vYye5BtNDhyvXLxHHL8chlmkHWM3aQA+dCIvA7hVaH85B0nsYpaLefPn8+ERhpzTNtGSHdLiCuKnvoyg4n2EfLG9KicPX8R2YByUUQPMYlf4lSGTHDk/dRrBgXLkaMcWUzVVcYs41tK/uEqqooIT9bI6RBgwoUMHXiPYOV2HSuM8xfeg+Uxqe/FkmPhvfIfvvml8o2vfwbFFn6XbL+GNMQtvYwybBDxyofQO49GysDpM38AIta9Vwr3Xn1DXJy/igPC1xh/J9iqc60894c3oZHhjNer8GoF+0EUTssoN/pwwj2FJcl1tuq4aDHNXqKNTEZUXmG5g+LEFVedyLql5xo+Qjzu2gmiEwitLPUjo4WPFideB3MccePjBMWJNbnB6UQDlOnWqfg4YZzwVB2d0urIcUyLQOrsqm3i0Q6u4rtVJ6fqpP7UUBxZ7QYH0p3PjmVu1en3GGL8m0yzRWeC7aW9WJ5o/q9TWC1OLrt9lpXuQbeFgIse7yFnHKnruwDcOPlZha44GAorJCxtZw+X/vEp5vr6lSAyg4CyjM3WXoLDl1w+Sw/y/DgKhe9sw389echW05KGfWPlntsmylc+xyr+Q/dmcuaWYK2X1rBWFe/6XnHK7uKV42MfCsJeFKO7m/B/6FmHvfIuJ3PpjzxbZiwiwhjgB87K6Gc0fpi48opKK8d4/c0YzYrYN6Qgt/xoRSAJxn+ZsiH5q+xxZV1FYXglEfrJdws5tPVVoXLSyaELKLvUv4/ttT2k2aV9w11dDzFjCtxlkYzuWl5762L57rM/Y5KpxQljFWW5NS8ydGQconvJ5OgLKnbCfwiqlh+VFIjaubrbwlYJHkm/CV2bun8ImKAXJ8Yeka01QXynKcMyzsovrOuAvsv4t4ZstMOEt4+TuZTFhUXfUZGpKUz1lv4yrOMEykV9feRY9AYi00R+pi2smzg0fxX5jNKRP5Sp9VNivsomOindVKkAkiIH6NuPf1YzLIGGc9iKvxgePB0GwEMPdTyrNCJ04QziL2lrHobG2T7ykPVwNTJ5N78pi4LY9F128Rt0351HOI4YHyf3ctAANMMoSm61B7gtKdZU5PPWW6fL937ww/LWG2+Ur371axxJ/GmgLuX06fdiZTKJtZ5bHevWRhfkwJ71gC4mUDRcuPgeR/UNRLPAAABAAElEQVS+gXPYH5WTJ+8oX/7yl8sRFBx9WMyZj204QjzHU0/+WoTnqMh1+7q+HrWMl8Yc650H2v9sDxO75XBh4WpgPYJ/JfHXyh856Un8wF/Eh32kHjncEyXuJP1fy0BdGdgXbBPLdAHzKtZ1S1iniT+t/z0hzK2LXtvQVvgAeeo7ap3x7/rStWzp/zecqiMfzWIz8KVrpJY8R36Uxui/XlbeAvYefLnl1Ymar5VqktT3fR95CfK58VjLb7PcF7Er3U3hbfSu+/4Y/3PLB2/B3M01pd4SuGjYe+qCuT7uh7zz+ebg9r3eu387SXhoY7VZf3zlW5KEbEffu7rLD2vnux3WDkfcMATiyFDT+sCnhO8340XaoyNBuE6YrnS26ryKkHSYXDxuk5hGjcVJHUDC3IVHakoZluXEFpbjIN6UXz/BtIA5FhUUjW9Oiodx8bGerOHX+tcSsNDJcNN5A2ltz4pdY7eX+RAPAK1rNNXc19HSuk/fbQvTrMa58thSTBhrCqp0kl9/ZJLcZHlt+XnnJ0oU62oHR8jwck8ihWaAqVt2wBNRXFVTCF5jNVHlQiYf5OApP9Zy0Akn3NUBya0amuW5JUaG5B7PHQQvnx0knBiKI8vfRohxwOlhkl52b5Sn8G/y2U/dw4Qdr9Pg3fplsh0hyoGD9Ai11iE/Ldq8J9APe1f72RCrWvGQMYPou+Xvv/eL8tIbV7A5QpnG6pJn2kM1/BEToVVnXLRGYE9q20VY0i4MMQh5Z949Xabc1oSPkz4cYI3CLEdg4jqHtXwVE27VEY87KE6st3SisgR2TF7glrr2wpD1czIITmNtw4pClC0OrtKryKA9xVwqw9087AMb4PwGOF/DAmWNwX9bpUn+mJgzEXawFj0O1H2ZqAIXz4EQBYF4df+mCkeFvW3gcIXs8Ex/+RInuDzx8L20EwoS9+fTZjsMJls4rx1AgM4ATH0YrXimOnzrcfuA8AJPVQRIX+x7Bb/UnLv1cR+sCgJWq6AvBf23z1wqz//+1TK/hOKICb9qyGTDNip7S6UaG9PcfLd/iGSCEFIs09MdBtg3PcGKoXSmMGhvtVW3pEH+tWbgFZfCRF4SSO7cEmKft8/SDinAEivt2S48hg5M5mtwWQPzaxsRXJ+5+1QhNq4CAQo2Jn2aziuMbjPRCn07USBP49LzQjfbKE56tuz7mzhQPlU+/5mHONqU1RIKpvU65euzwJTSWi0feM0LXtH2/+zB5nP6Ifio/Jd0tIeX23Hkj8gftFM9VWdtTYWpOHK1ysx9Bj7w3dY/7SCezCMxEquJX8MsIfRPqPe0X2JY1/by2RzqVWOZb4v/Jve2EL4ElCSpgf5KCwqwrSM5V74UXkegb7EBK0r58r/aUrU8Qz+K8s2lwh7AUqOPov5aGaTC1oJCVJzsonjztJZjTEwd11IFfiwv7Z8aqWyEolJva1jrub+2RGjC86QSxELSyAimPCvgWSN5BXJ++cVzr6PwWC7Pv/wmp4DRn6FIJyhOvgRwDKXyBELowQO3lTmcql5dOA/LwFITOHu5ry1f5ljfdxGC2WI4fS9Kk7vLIPEXsLxbYPVvi4lEJoxAv4HAPASfSU+C1g+iXHHP/DqWCvPEd9sHYDXjkvPHwZh028JHDx8tB7A02UBAvnDlCoIyjk2ZiLpKrNCd7bFU1fsMDmNn2Nvu5akK+rJSsRkuRrn2LwVkfay4KjmlnwPyWsESUOHbrSLyy8kxuA2rt1/70lPly/glOnSwHo9sO+hsVNz/5fJPTbtJv708d608/4dXgXGqvHvhWvndC68z5gzhvNyJGKNbozjxqPZBJiQK78tYnDgRnfQ4YpSynqqj4+SO4oS+oo8TT9Vx4jAxoYKF/LBiGcbaJBYn5Od1GB8n4mmFiY0WJ1ZMfydundJC0hVirUkOQZ8rTHw8gniSLViOOatM8PX74iTIVWQVKVPwbmk3GUlzUhz1DL+GvuQ7joRjyhP4mxlf5DhiFCaj64tlZ+UqfsGW8VuGvxSEjzXSe0pPL46Md9hSPeLiEJk70R/UHwHjzHWsataYSK4yVgBwHAYP4EesD2WeY2e2BytX0g/25M8KnlBSBPGADlqyq7jCLP9RDhJ2J3p9wHNKxQnWgvfdexehqv4BBGXVDlt85f3KetkGBg1ug6PQBuOaDn1jmUvfz1a8FE16+bowCYREKROmXC8XN/rx6dNP/lHqMNjqgFWlhgtb6AUon7QoEV2oMU4usugljdsTdskj/vbAgRNIF8+yBVK+Ch+15B2UQtYyigNwqYKmB3kvYhI0vmk9QJ3O6pdXd8s771wu//jMT8vwOD5OsFBZB+86Ee1TVo1sJ/ypUe6Ov3UhlKcE+0PfEf4mqvi36i5muVYT2MFLFBP0Ua0jVDypkHKbn7Q1kL4PLQC9mbnAElnJeGS8zWJgL+mcuDtaOea3cwPlFp3QGz7uRJh/KjKUVbV2CITkEUexyHFuE7S8Dcd4yu7NYgbyJRk5UXYLtVZc+moLjEz8oxBThgR5ptnZwcKCB6S+4EF6k8akN+FQWS06pERFiOCGN9NGliaWx7zHMoOYyo4izfrXlEKtxMv2SnwF3Ydz2K9/ieOI2apjDDBLdHBEGiXj+EMEkZew5vrtb54vb771VnnyyafKHXfeFf6rtdswp+G4Jc9x13rEYhw4XbjSN5+8RD8l1+Exv/nNb8ptt92GVcanYhVSZRngp73srzqXXkHe3kKppTWcSkAVveI4jp6pqFbwytv2P+dK4l1EeIqPlh7OSdwmBMlhtUit6K8bDGIeo14tWEqUJqPArUzqle2+1FtlxybjmrxMvyZybxXoqR/lKms47qlgMW/HFB1p6xR3fv5yee/c2fK1r/8beKUO3G0XcxDz9CLgEvteUWp13gC/xuDL3lMidv/Y8G0GXeE3B7fv9d7925WIcroz2yt176k7dp73J+l8vjm4fa/37t9OEh7aWDVsr9S9p+7YUZxIy3aEfVG68ukuKs9d31Jg3pNBd90rMHKWWyC3G4j/2eV3dEUNEBkoAn4V9IMLOk3FiTW0A8v2/KWyAgxj8m7VOoOanRvCV3lyeX4J8//X8GyMc9gxT9VhEEoWEC0WJ4p69SINDFTGonY6w1Gy96myljBOTekBSKaVY70ovy6ow2wNh3HaDAEoP77xF/wLJbBaoRpao+XNalASn7YZyBT4HRDqBZNhEIvwT4Sp2QNhtikoXNJYlsHNDLhbAlNy7gZCUZYf5up7vRLEo/szvVwZtP6eHGAO6eBEV5kgzDvbfOc9J5A4SJg7+LAUme7aBvsI0e7L7JIj+OwXp7zsyq0ClNpbmb6DARxEIYNJYc/uSvk0zmE/99T95RCnEKgokOFbB5uZh9Stz3ws19823LshCaSQ3A1rryaAPJIXr2bz99/7ORYnlxkq2M6iWW5jcQKLpL6KNcDHRVWbdHXAcbBRiFtFUD5z7p0yiRAxtMYWERzRjTI4D4NPj0sUTtOSbaqOZiQZBW7KjyDooA0yPAK5VyEXfDDE0lQKFOCA1aYoTcitbr9iCKNhFOb6KWKbOOv83aDdNjzyD2Fwm9UdRpfsH5Zm6/7pgIOAhKCqVp74DtgODAoO4kP6d5DZguH3M1AfPzxU/vbTj5THHrqTjwpytgl1YiVsG8VJlHsMXGkcG5RM/MeDWVJ3tfAKPrQx321NLTmkxfwZAv26/WWnZ6S8pcXJC56qw8ogq1cx4yX+IAoYe6TpYxrMXSuTlEt6y1Jg0cHcNVY0NRGfQIAX98KoiMoCIGkq96DwwEnJEdjEhUKPgkzlSZZGWRbBX9OleLQ97f+1j1cQiGA8f7y3kaV14htonvox8FstGUGAycowk4WUy4CrQ0rLtLNsilvKGGZVfQ0zd5UdfQgeqE/L4w+fZEvbo5iZM8mR95Gjk9QIJ+LYPkPaCEg2rYIX9Uve3K27bwFZEHMpoPlHLNrCLQ/9xCVqmbuymAlpStJaJymr2G/9e6C9mpl360gdeKSI5seUXPlWH/1klJsvy5PPSCnCmHjC28GbtaiZm2v4/QfwP/GcVT/AUWjO1gEEYX0zVNDym7xbOD7K8ptCyPqjrX8YJ/hIO4PTvgHqieKEOW85xpYJ+0GKpHFsaYnYf3YS+0nlaG2N999r/W2bSiMyccMkSyd1Kk5UdlAqn1CcwJv+7vs/Ly/+8WyZW6SP9zC5pr+uMRl2dVVrg+PHDmWytb01jLXJPBNjnHUyV7E392OhuLp4oVy++CarjLeXEyefKgeOncJfyjJWG1iPyIeoj8KxCgJPOZPzCp1OGQ+hOLGcS+yzXyGudA8B85V/0G/2x9PPXPU8cexEwi5duMgq5AJ4cwUeqxUU3ZqPe4yoVncH8HsxpdIE3F7Dwav+PmCWsDb6JTBM4C/EY30VcGfJ9wAm6IOsDrsSOsepZ+u2BdYDm0zkb/MY4pnB8im2Iz3+0EmUnSiA6VvK5NlqCJzdl/2PpqV+Df37DPnUMMOlf+mp5WMMn/CLS3PXy+9efAVcTKI4WSzPcYqgFnurOO91BXsY5ZWLIjlmmG2RnpazyDYb6zSNVYnbDjxBwgURFecez6s1gZYpVxc82Y3jo5utOrbJMCvArvSKBwiCdsB3Cbhyi8/09IHArKzi6v4ox0JHcbKKc1iOn15C2aKTe/0HqKRSoVItTuB1PGt1oqVLS7uZVIgHCBHMpW2DJ96HHTNXr0ZxMskq+QQTvl4sOzZZZNpkzGHEgh53UlYfihNPJxtAUS2H0UpLnzTD8IM1FEzroPUGvG/b8Y9JWQ9br5hxM0G0TQAg7fLB/CdtFP5L61GOShMVKSrEM26iSDt1fLx88amHyoP33wXegAI+r0PibbcukIEr+a6Yux3CvgeYaXyd1vtNa6tBFgX0ASTda/HrJFali5Ym0r0TThe0BlEA2Xmd+Pu9X6WGMgE4V2lkV6FKkdUsqIbXxS9M87LtYB1l2ABpnJC6kCYv2I1sSJ+E4dga2ywAOFak0Z3Uy2TMH14rTC4QbMEndvvYDrfWU95AcfLd7/8EK69Z2gj8yMPIXwWCW6CDx4b+BRDWXfsysCrPe8ywcpC9QPRIfxVP9hqKVinhohHx6nH0iEL0dS1sSJm4jqdOml200DeRMomLOXVBFQ7DGA92gbkqwZwMBxaSu4gIIsvC1XlkWqwmoFUXZ7Zo58hGtF/u4KQPOcTtwlpA6eTdpSqZsBN/LxWXxo2STRxnKoIsyCKloEaZi0wlv8zxyuBDZ9s56Uk4AMrFoeDcxjREBIozMpAe6qTctqMc8rV5bEjHbsuQl7eXGOxncW+HY73vP+VxxE+WB0+hOCFyHfGlMbHM0eYoJkyrMu7N198qr732Wjlx/HZOH6uK7El4qPxEhYPjr32CZuPdPMAr9Dt3Cb+Q0InW9M8//zxb+Y6ifHkiil6PsJduzWMJJdXSkj6ZCv2VxQ8UG7oRqNbEVWGhzCf/1xpKizmdTsvzJrAmVCGs1VwKpmytx1TwCZcniLnNULyN46/GMUv8u11HOIXXPxUji+S7ytjUC32pOJ5gW+YAVkk6Et9ksVHenr4C/bsdyVODVlcXsTY5U1754yvlW9/6RnnggXuxgtSB+R7eazvYLMpfSji17Sg+bdTCbZPZrp3f+mgsLl7yDtT7wttvhFuhD7ns35LP/5/Kj+LklnW6FRI6GGiq+L445tRV/c73zkOD3K73WxaebG5CeHeaf2X5dEJJZN9l+1KEpViwHb9dDa7x2vItWyZKN4Zg6TXGJqHMCfKCAlxhnsPi5NfP4xyW0zqGRtgHzcq2UaSQHpxgOvDUS5KUKfJWC08+8okIZAQaHAGSu9Gc2PukKZsMyQmy1gXGk/ztenYPwWsJEtafuPB1wqxBoG7SJLswmVqm9anlbqDZVxhxkqdgV/e6+430tUASNxfvVaOsfjpZCCZxBSSc02xTfliig2ne7bjiznzIl3tllMAJV2fIRpAdwBybnIhfzRcpiwo6QKy7io5g6QTUlpVX4bKPsS1QZHCWeQJcZe4gLZYzOHDr3V0un/7k/Vic3IsJN2a8jpjk008CB1dxVc12SWs7WaHUCJgJyoBASC7ebQVxm9qaJHWq+PKVKpd//P7PysscR7y2WxUnUZyRiiGMX1c4XI0gn9AW5QNTHacZEMFGnMO+e6ZMI0R6LGKOI4aBqjjpg5lavu21FVqzVFYNKiFo0JJn89FiqhfGPEDaIYUeBUIGb5LmlB0tnhS+9MWhZ3UFDtvRKqkgXGdwWmMJdxNheRPBwSOIXVXaikBC2jRkah1tfD+rKZ5K4VKQaHYSwOIG8g7QkG6AwXqA7VPHD42Uv8W64ZFH7qUsBEAE7AitjuQM/prp2hbWJQoYsc3/rFKFtusgGWoAxqpNxzIFoUUisz1dFdP8dId++cbpy+XX+DiZuw6ecLi4g0CS/FEIeRdlJEx9Ko3WNraNJDExrkm+W2AmEfq1LFHQccB1ouKKlW2sksA0QQt4MI6DtIO7gmdIhXIiABEvES2X0qO4afpQyiV2+h/le/EmiPxW2AyuOOZBOJu0Gwigg9RdhZU8zv4uDCoiFXZVhOiodYVBfRjhYQAlyTaOEB996Lbyxb95HKusEfoGNQYRu3S04NGyrJelK1hTTSeewaG31CAPFUYAN7b4sM28surGIygJ3PVUHacgfrQ2wQK/3ikAmPf6n5jbX39LE4UG50GkexGWVbz6oYbluYmaEHErj08uTZw2M18r/Hvld0UhXVUA0f4gYoWJoccb5lQdookjy7cfBZgkrbAF1Ob9Ly0frKb/f9T1VxkkzmlgagDd0K9YysS4qwdFJ4oTmQbN4k/4n69BmbHp3134r4RLhA+4nL/5NdvVMmDZb6qCz7FyHbr7P/6v75Wf/+qV0o9/pd5Bjq13WwH0M844MIWzv1m2ScyjUFhc2sj41ctgyciQydvB6RG29HD0OsqTEyfuZLX9RBQwlzGvXmKC7uowETNCa1EWc2/qPYaAewSHpAqrV/A7sorwa9+2fgqe+rLRpFrLhoMoQmZZaHDyPz8/z+T+RtrdVW59nCm0i0sn0Pa5I0ePBf7r7Iv3SF0FZqRhrC2YpEIv40yq11ACTLKCOYt/EBUM65wgcw3hXp9X8nPh3mALyX13nyiP4Pz23jsPl3vumCkznKoT4RiLPBcSskgjwwyWm0agHpKlTWb4B8s/tb5uRrgMfp9/6c/0VxQnnKrzOxUnKK1XoAvrV0/Labbq4PvEo4VVnDhmT8AnN1AuuaIbixPGshWUReGhOF28hpWQW0OqjxMsOLRMYaXVrQVOaOQfR7Q44cSKHEc8pWyCEoI20WRefKm0dLvKYSx+NHGP4gT8OfnXf1L3Vh0XiKYoNyQcDPAUeq99XZyIQ+WbAba69OAYdnLxapncxkksfwNsKXVcUq7bArUqz0bZctMLrWoFmlM6QK4nathX3F4iD95gXFgH3nUmZt63aNdtxn95qO2vQsM2qb1I7uFVYZL/SBtR7sjs+K/lTnpoLDmYdm4ulbuPT2J59Fh54P47gQv+5PZXcGk7mLNKFp0pm0EvMDhBU9lTRSHuPPczYeuTZ0HPbgWwbRxleylPZ+bGcUuVii3bRutR8/a7hKUyYYfFHS23xKGr5L0qecQx/N0tC9Zz2ONiVWqZmrYWJv1EVP97jJkofBy/9a8UnGT8A0d0xB3qnC0o9lnTaVkMnlWcvP72pfLt7/0YWRwnwiqqiFG3n9Bn7Qv2Ywd3O4GpISZlmXAzByW/+xl4PajBxZT4SqMt5Uy7LBx5xLEWNFpFDAsDcVXwyBqVQTLWN0oKT39xDNAa1oUyT5NyzNzEEbtjopZlhokeeUH4BnAvzLHdkPepabbowTtUqqp8iuwrTonfh5LLBSdx5jYP84j85ANM2fRWpo+2sC/soIiRJvvJX8qqDvflE/z9v9S9V7Bex5HnWdfjwnvvQRIkCHrvSZFylGlpRj0TMx07szsdG/u+GxsxsfOwzxuxO/u0068dY7ql6W1521KLFEmRFD0AEqCD97jw5nqzv9+/vnPvB4BqtdSt3p4D3O87X506VVlZWVlZWZlZlOP0qzLFTSD7jMxRooTkmAfEmJTp+LM1XtKmt+IsoQaj2LE881iLsizPRSwbyZ24A08ZHBbFyaefuAf+tZp8XvaznVxr0TWyBsPtK8eOHCt7PvgQvrIQZTRKT9rmSVv2jbQgfCqqbZvKCi34hljLjGNdZgykc8Saeu6551BCLy0PP/xwgk07Z/ue76uwGEVJmICxKFwdM0ALrlEmUpeSsO5Y0r0KbC0QdU30lJ3EB6QcaUFaUpkib9PN8jLWIAaYFSZjLM1jXpGuatlu8Ko0VKmmS+FQ3DBdB6lg0c1P5ajz8RT97fojG4m8rQvjIMpnMWX/796zuzz3/F+XL33h2fLEk48zH+HeRWdOyz9ktE6vaVed+rNBd/OLb0ttctt/zhFtSdM5W8/8Pf18+qaV1vZ7+r1rbq7L0p7w/3/9bYqTCtjV4LVQ1Z6Y9rUl5Lb+9tOrQW8b5qZx2PZmzTz9OVNG8/503umbJnNbQm7rbz+9mvd/Xf0ZGeay5xnZdejTGTAyJyEHrAPe8rJj60jyt8Qto5Ij+Lo5YNjyI8uUmNROm89TdX6FxckrHEc8C1cdlqbkJRuD2DPRVW9Yf8wxeU9idsBbTmt9n5Z0AYiLknCi1O87wsM7Sm2UNMniJgwqb/Ax/U0+AfXKV72vb9U6I1AFck0qqYZys5lLpoiaCBialKmpXIQwqGA4XT/FCortoDSbJzh1ouC2MsiKXZ/VehuImFRgQM1LEbxhGBYY31U4rOahmoN3M8l0c0pLTBmpX3w4QSq7TyKk6ss5m52pKzAjJ+oJG8LpNQrrUQSRr5p+V6ZobIzUg81H16SuOjeUB+/ZgsUJzJfydYXKLgnMKUtwYLFP7duQjI0U/zY3DWtrf6udPozwyQ1ZazPJC19mAn+xvKfiBFed0jGb9gCn02/oAfw6qfgOYNr/zFrpF2cw6cWdtEMoThawo+epOl34Ws+BeUfxgZDGGxGO68JDUGHSwmFf+RcKp2UAn90whNx+TdN5aADXbnEPjm2bE7sKpDEm2XEX1c6iwoGm3fgmo072ChgoTcZJj7UGNB7zSmZPq5MyGncdJzcLthwnE9saE/vsviJIYT20hlN1HkPQu+OOm4APAQChWXPdYe4VzJxEnLQsXIHMGAAKb44qBc3EMgF3UXhBUzGhZeelU/9d6lUOMpBrdbOZhVB1vLz0q53l5AWEN2LO2L4a5Jm8xDmJYCrqaJeV5jed6oTojpj/LqI4kWZcEDiGosH3GTBKg35nJFCEpYCAlJWdHwQZFXYqo8SNCtE0jyw5+cIxQX7/6JBMqL6fcvykY1NHyszPPLO+SpngiTsTFbQdO+5sdNOHUp5mpDmKkOpdmEww5tzFNYhZHzCNDZ3FP35Deerxu8vKRZpN696jcgkYpBEEPSd7wdZcVTgN9giCakOoN3cgoZUSuFQGy2OEU1cdj24NiinI4LCDCBnVXdE2kosvyxTvLv5aP9Mu22H9XsGzsFFubb+8tZU/z2su8WahNT95rYMkhT7r8q/yf0pu1Z9XzMTvmfprLdYf2qA/Fe605rnEWO0Dzx5zaP7mvSjzfg/115YBX9r799f+6g5KqdCfSHL8qDjzVJ1VKxa36N3arZsLWjIf1CYoaXvwRXLwYJ6rrgpzPpkWFAid3pxDpF5+8WJNHKLeP/2zH5TnXtxZeuYshxcRaJRFUi+C7dIlBGHFHeMibhAGIh3ETF+hvws4x9mBVbG3ce2KsmU9Aa27hzDPXlw+2neBo42Pc2QwdUBXxqTRRUalrBaY1m1MgcVLsQghz8lTp7Jj6EKyKpRYTMPLnJ/8pxuN7iG6EZwmDoguGSrD+xCyXbToBuSCxWNHF7JYXwLM6g/PX8IsnHEns9Cdx0WRJ+2w1A0WFLCXEAywl8XAIEqG82dwE8INpNvFK7xNK5YxXDdv27q+3LV9Y9mydlFZt2oBSiQVpXVRpTDuWE0HpUOaXgHPtEf6Tf/wkb7IQPAZr0DztU9VQWEuT9veePt9eocYJ0ewOHmrZXHSuOowP6lM0k1GK0AtTi4At+NCV50x3VpQbmhtorXMRd1skDM8eeYcliSeqqM7kkoYFz2z2JU1MK47tcKynBgn5zklyQDMixagpAJM3XZ6qdNTdVwkuUAxxon16r4xhwWQC81hFScsMI1VM4JyZYTF1DxjuFFGaA5SC/VChJXfSwcohMkwy/kRV505xMyZO3SJ4PycbAdtTaBQCX+VN4G32fSJq81sUMRNAts9aFde5Lynxe2kmw7uTKM4GZL3yjeglUaOzMmD0EFwD/rDMuwZ+Q9fUbgDq/xflybxbenSpRYGY+w8b8bixMDe27A4cSGnS9d8aHQKQU9e79J/Ciss437oztDrJgh0YrVuWBkXaIrF2jh46mE+cL4VJ2YIaC0/PGNc1fhZlIhSxjKihIHODWws7Rv3p5eV9Dj9ojJGJqtiUsWNC1CIE4tGrAFQlnSg4FcFMsouemQ+cJTgz5k3gNlvYJRue5zbubXP/FQ47GQTpJMTigYJJr37o+PlW9/6CacTE0QYd7K4zdoO5ILwKD4TWy6/lPUoSXq3fMtzbm5NDPV4X2VB8gB/5Cz6UdnHeZXWcWITcgdjfESFH/ixGDeN7K8szmVu1O8mhesO51PlEY5gAnQ3H6AScKFxp4pa5W/njHOMeft20ZIl9BP0Qnlad2q5EgtuYOphg0ohfgqcdjOHKxONApcCflyagDlWCtSnBZT3buKo+HFWEyondl2jPQzA8hSpR6krblMqAGDQ4j5rpszFrfmX+sFURRmtltZtZ90EJJ13wlFpi4+y/sFKqwM5fAwl300oe7U42XbjGh7DqyCwbOClP5DxVDiHBrvLoYOHy64du1GWLIWHrohSrY/ArwZhFX+9wNmt8g38XCYYd6zO4JGLUbSo4Hj/gz3lm3/5LVwvl5QvPPv5suWmm6BD8I0S1ZhR8p8++lVlhVa6kQ9I08JYGFyzGB/mMvxH93lp38DW/fApGhaFn9yUbBlXpqnAvUTZNn4OVinzCBqbTU7wLI0oN+he5dzjpov80XHhqTzz4Z/KZMpYuh85/4SXgxsVsrpx2fY55NUid8fbb5bv/+D75Q++9Gx55jOfgUcSikDcUU+QD11nk076tj94wl3rebqw3rcn+mp7rjyrGfz0quV7N/Nic9d8+/Tqqz5pfz59P33TvNGWkNv620+vf4j6oziZBmP6pgJgR4dh83P60fRNK0/zdW36J/5uElvfV3+1VfL7rV/BO+ilfukFvgGyQXfa2zSI709svwzVF/jjue/l3RQiw/UPVx2O6nv1TU/V2Y01xHIYDxYn1gvhGoAjihOLYFTJambKEJZW1dfVz7u8kwyCCQOspO6ySFiSmLswdSfXPKFa74A59bRypwmOF5vjcyfOMPMUBEOtfpIGh3XAeWyXwiDVXn8Jiyh0+MsphN1vC/fyJQUO6xEP/M7kLCz8VECVgdiOako7yKQzFCHIyVPBa5aB3YBDMz3R2OtkAszuSBh5+jQa5AUwUU0hxzi9xro02XMCcqJnT4Rv4JDj018duIX0dAzmRJ0H77mJ2BoEg1KC5bkTvvhwgeaE6aQaDKcdtUlXfYpDm9tcNrv9t03mnwuQ7/zgBVx1UJxwqk4xNggzr6V3IWzE1FPhyXKkLwo1Lof4ckpzp8KI3YcOHuA4Yl11WhYnCIX9wG509/QmdXvUYQIrGghVuMGAjDLtYPqKQMc7XRyL2O8OoRMCgXNAGRM2sNLXKkW0+Rl1lwZXGU84URExyaQ7grXJOLhVseK2ghPLOESl8BczzsiatgEK528UWBVQ/e1uRzfCjv6orGoQzmg7QuYU1g0rF/eVJx/xFIAN+MSfjWC7ZtOWTGjip5t2R8vu4EUYySDmthN4NFNlHyDD04CzNsb+h1q495tnKhrBqWazSAsIVcfKCy/vKifOsIOsAObOi8qToNJ3xBxZaWaGJMXaPxEeuVOoO39OM3NddeaZk7zimUta4uWMcAUvC8k7CiYIUTx0d9J3Uiq4yS2/amWUBX5dwEehab8Ae8r2LcrLHzVkt833KEDYQrutnMLt+EvwX7tLIYgJeoiYC5qY9iKUuuM1m8m8VyUlNNFN2b20yWMCb70FxQkBe2f3egzpybJssUEpOUEIJCm8xd0O2OOzDl5dADVtDUgBOFikXcIPELaVXXCb7/HHDLd0kcPzjKfqsMMeqw/6IuBTUHDOe0FjCvaDshDkTAM7rd9+kdBqPwXxu/WsKY1CZ/h/K6uvNFc6uyY09TevNlmmvynf6lQiKdhWZZmn6rAbzXjJqTryPyukkHz/Hur/fbWfLq7opHNsp91rDKm+JjhscC02mpmI71aauMsosr9D3K3vq79qV1F2ayM8JWWxIt1CM44hFblDxN75wV+9TJyT98ulEfjQJLvhLJCMUTF7Di4PCKEnDh9Wx0E/4HrHe7ppjnhsLOVvvWF92b51PUcGo3Sn31745cdlP0cazzbAqOMR4VjlcXVV1Ke8j0CkSymjqxw+fJTxcpHxQp0qa8NvbZU0ilUKShuPBdYn3iCw8uoEyoZvE0ob+FlAsgvNUGPxivXIskUJOHv86EmsKtAYyc8ozQWFfMXdZpUexu0wpoeBbsdRLl/A/ciFgPlVKMjXxlgQO6fduGlluZnTKbZuXlm2bFiChQrKYPCnb6/m3QaxlCNlTPhFP0C9zQ3PWtevoX+Hkgspg8O+8Q4WJ1PzMQu/WF5PjJMe+gfFBEpPhX1N2d0J7SQ47HxOrbsY03cUIihOEhzV44jBpQoR3ZPcTTXGyXlP1WGBoMLKhYOBXnXn1OXtEsEXvZahOLlAwMRYnMxfkjRN4FUe1OCwbafq4N4zyqLJgIpaAUVxwriU38VVB0WKix7XsyCi0ro4Ey/8hfpos/Mipz2XDk4vmn3hTJmHa5TBYbvpk0msJZ2fXPx0QqNaV2hSGdwyp0xKg/C4cfsVPmCsNTTVBFaHRmn/CLzWzYk6WqpyohlWraEiaBUkvprLudA5xGOrHX2N+6R0YRy3NUtxfX1ge9l+2w28ghID5ZpuB5ew1jGG3XyUcfaVQY9tsvMJzYSHs5CG5gdwSXPne5TFocqvhSgQe+kL5zb5vziTXqV/42mcxBpLSyLTxa/xOHR9mNUP7aLow4Erbua6LFEINK48QJQLForu3Ls54jGr0rqbES4SbbeKoMQEcV7XqkILKgCWX/QwZ2mNAmaB3fmFviJNyx48mzmO+Hj5xl9+lxgpKCmpy60XO9lg9uOMTfEc2bDFo8Sjl7Tuw0oJpJFQ5X8fmCu9C/7hAPRplBU86gfGERbUF8+ezr2Kon4Up8654reOc9sg3+DPcpkAte7s46QywXADxrp16RMa5T83aKTf2QlybBdUa6Y+2urUomIMAgSHbkSgkEgzcKm+jDIBiygDnmr9KL1ElrcFvNgLHpxb/Wd9o9DCBIqyWX1dWeCP8mwE+axXVzKUnHjjVZk4NEI/An+lm4opWiPofPLXkqfs62ahrnzlw0kYMhTAzSBKvHoc8TMoTm69cS28zM1E/xg7ZA7/h6k7rwwiG7z77h5cUd4tW7bcVDZuvLHSMM+1LLNsaUiZyJgfBodWUWv8qSXEp5LQX+M44j//+p9x6s3K8odf+1rZdsu2uE3qWqgC1nlb18zgmnbIj1SMqejzGGF58wWUuK6PjFnkqWGOI/GpHBRlH+uHuNzABy4RMFoFi3hx/M2lD7uQuyxTZa4yg+W67rJM+ZqWXVrkybf8tl3yTfvK335rgX75Evhj7PUjG8/FmsYNgvf2vFd+8Yu/Ls9+/vPlwQfupx1aH1G5hUBNErXrKMe646pJ5mG9KDvA8qv1RttNk6nJy7dFN9f0C62E/G4SW99Xf11f9j/i+qM4SdPagGza7vc1TW1/NH3f5JnJPf1o+iZ5ZjK20tsS/iHrpy5rrvBCOf5q1d+QVaWCmk9aa8BTISFrUAivTCk/uE9qtM5OsgPEOHnlDVx1XsdVh12xMY7SlT94VFgsTtAqW6hMM3U19bdQUq0Q2urnXQek8GUSt1quMHZS3DUIkMkhM2/aSKYWQWfRzH3rp2/zZ5m1/dFCU0fNQCpt0oTVoEQybU8HcLK0Nt9tammBTFplaiqDZIcVNuGuoImvem99FkEeM1FmLF1gVO6wjWDCe4oTBY4Qx2PlypUoA1h0Y4q9duM6doUWYXlgWznRgUlIs08FK12KPti7t6y/4UZ2EQnQSX79a0XrOHkDi/UBgAoD6+yIq85geeDem8oDWJ3oqqPixMnI8+uRMvOe2ltSWi0W9Lb2WwGXKWlUGthgJol5SJGUF0yX7+Cfv4sYJyMo00onO2AAZ73xJwXGCYQVcVxzK0bxcnyeo17JpHHo0H6CwxJvghgnnZgNa3EyC3gT4wQhRoY4bj9kBkagA0xItgWn5Tg5iwP8r5lo+sBf9myYqHxX6pEeRmGout6MurBFyFNxgoaCHSh2MWD646wA7DMDydpE4+1QQuDPooc0FRX213kWEpeBtZedDbXzMv3ZHA+4kN2TSfuYxXoHpqrrVswtTz5+V1m/alH5cM+7yJ5j5elnvww8WJ8w+few4LBBTjQd7LqEEpGeOpjk6XSAqIoo/XAlOBUIxrXR0qju7tg6kSHMnWX3h8fLcy+9U46fJsokcVomhY3Jmh9pR7qU7BQT/KVvqN8ecri4UL6IAO+uzXx2kM3VpCfGjnVxCU69EE5oiwrC0AXQpIMcet6GmCy98hqS6VdGIA+lXXeifcVsts38TnxRetp5ecK7Kaf+FNeWPYHS1h0oZ0std44dI/bCmVOMGeKetHYON24iUCaCmLtNsxBQOzBF3751Q3nswdvZATlVXn/t1fLw/Zi2rlgDGJSrQEudMc8GHpVi0lazcBYcx3nAsT8ACVBsFHlso7hEgCCPMrEL3tPwTxc0wTULEZtsP/hm6uS+Kcey5E+tUvOrecaPZKT5ab8vzTyrbyZPMgAbSRX/SQ2cTf/7PfPuzH3gaT0Rz9KDrmfuetWYC+yysuCzbPvLMmpj/v7rr1C34Th10SbrblU304a/ff3K45ka7AguFSexOEFxshJXHZXetRK/qD/9WisXn6HKtupSSMVu61YAg5kshESS77kQjZsa9TpH5IhhdpC1bniPcfsBbnaXRzmpZN7iBDrUtPriReJjsCAcVvnXNRfh1AUgaotBFmUQ2O233pQdzckxApASP+sXL71PjI5zZd5SNjigeRfYDgX/ZrGIW87JCNLXyVPHUZBecLBFAI87Gkhxd9zFxsJFxBYhnoaLyJPHj7PA5wQMENXDYmUMHjlBuVOd8Gh4bD8+6qtXr4hwf+wY5Z65wPj1xCp4amjdBRZKIvky8K9btz7Y8ljewcsseFnIMoBJoyOCt3QIR22iYGEDYOHsceKbbCh33YYl5VJOyKJOGhsBPadx1K7hszVuGvonxW7ihdyly+yHVvdIQ45JF7unzhgQ9n340fxyBFed19/8gIUQMU7QDhggsbdlceKCPhYnLKAvsihRERTFCVYaQwRqdIHch7IjSg94r8oSLUnsCxfdfg+y89qPG5auTVqPyBtjcQLf9b1FuuoAZD2OuAaHvZLgsLjqrFgGTbCDDD3MYe5wgaolSo4tZvGS4LAqBbA4sZ11YSwGGNlpt/MHHA486B7cA+9ugsPOZ6Grq04fsUSmVJyATpZVeMeimPYH74tK3beZocoItJPT1lDQaxExwVw4xp/xw0ZYNOleqzzgvJAYU/zO0JImKE78W6RX7aF6Mwz8nsLkwsiFnhsUsz15o7+zbFyFxQknot1269bw6skR4jYwoHe//y7KN9ww79ie+C4qYJyjDOxr/84hloJKqjdfeyMLyQ4Wg6s5Tnnthg1sonlsMjzOuQ0AVZyrdDx89Eh54cWXAQo80f/9WPhoXbSMOBK33nILyhqObka26qP9cdPC4kdlom3xJCZPqKLEctMtN9dgl9CGyk7Zjjv8Liq1funU1RZgxckkFk5asHYwToIsSkvf8dDxoaJ15+6T5c/+/Dulc9Z83sHChxyMCN5jnjefzMYO5hK/zkgpg3t5WeafFv5ZypPYlgPYtIxAamCYsc0EjL3II2fpj4Fjh3ELZOMJBfo8aFd3NS0Y3LRRIRSXHXFHmVrdnB44Ql8QSwOaUMnSuOu4yaLVyQj97IktzrFx1aBOrajmz1sYxZJjxY2ILjaRcmgCtK7lxHliBo2htHLceXKXLRWX1cIBeY58WqXEGos6zhP0ehJl4BLyL0FRpkw3Ar/p1SUG2WCEdyeNaYcMqCIqOBR/1C3mlADFo+1KwH3wnBMaJWDrJofzplbKUY5wqs74xKVy42ZcdZ68B/68lkz2kPKosqLzgJ2toq2DWFQny6svv1o+/Ojjcu/9D2ARe3voQ4su+9L51/405shl+L7KFHnHfKwutOjV2v3VV18p3/zmt3DXXFO++pWvlE2bt8TlWn6ghagK837oU0WkFh3Gk1FuyulJyFBatF2IsrIQ+013IWLiiW/mgupaBRaA1zlfK/FznrTDe0vZgO5nbDabTuLD/lAmFmkGzNZNSPlMd8d59K+ufj60z5QrhEmL7UHoytN2VGr2ezwx/WU7x1hDHTl6qOxBefLMM58iDsxqukqc1LrSCzA3e8py892ifxKmL/OFVlrf0w/abpo8AT652x62bpNnJmN7ar2vjPa6F5tXmu/rMpAw82zm7tp8eXLd47aE37L+GcVJe01t5SX52t/Tea9/cH3KdOaZGzLNWD/MJE/fXVvItb9/bcZ2JE5nuu7GBWFOWGFic/DmaqjDjheJoadWIj/DE/LTN0zgLQZ2lBmZIB3kTCRkdMGgq86r+Py++hqKEy1OcNWR+VbGDsNkAWMkEBmLV2g2oPCb+sPH/bCuVnUypyy+84rMXo5NWt6T+GVUFTS+WjDmLvcytFwWCQNrLS9rGo/chZaRRrGgHAaejJKu9jWLwiUoTtA6a8boQAszgym0QCQFhuQzJtCAFBxSdQN/rSnMRxwaR8QyUi/t0OrEHbwhfACPHN1bDuz9sNxx+63lyvkrBKI7XW7cdktZDpO7jHuKPuPG5+hHOTKXnZVBAgO+8fbbZdPNN+MXzaSOoDOXCPVjMNL4RbK40zQ3vqTsMBgIrQPmUjCzVXFijJPli1k8Ao+TSJdWEJSvJCHW0t5pIrAhplZsx5w/LWby47vmb2UJIsS3+GT3gORvfx/Fyd6TCJxa0KD9h35cHMTaBWQlHoa4oyQn78Q48c4Jnl2qWJwcwuKEXYW+UeLPsAPaj/DeD0MFgfQD79Evxr6xczqgteayL7I7BDxdrqxhzj0oXfowy+zxfZUcrQ5TgJ8ADwp34wi2w/plKxQyUU7wN4L5uxZAnu7EGjj0jegYDGh5pYkEn9NtO0ngRRm9kcY7MP+8fB4zVoS81WvWkakTIY1dKnZMN61ZUL74hUdRnCwtb73+En07VL723/0rgjkezXGh3QjQl85h/k7/zEbINpip/an/rUoyppTUazuC/gwukKHixCfQdYVNYbaUPR+eLM/9EsXJSU+8YIJios5OUDQPIqyFPRsjevymQ6N4gm6d3y8SoV0TyzkI+j52keeOg1X5gkK9k16srPiuk7DlUGAuIUvmOv4A127IRRatPwTEnSKtGiLoCQzClMKyE3Jcn8yWnHxQhqxMgFqY4DfWWfSZu3JjHPe7/8C+BFhcSTBNdxyPHj1e7rn3AfxhFwbHrPbKBIE177h1U3nisbvK+YGj5Wc/+SE7NP+cheKiKH/0txbYS+BAH3nTmd1Trx8BKQpjKFOBBpjkv8I9BfEEA1iJZbhBt36fYYFqsE/fV/i1+XUM8TJjRXoJ7sivQth8/llW6uM7Nw0O/d1c1g+81/I/4aoVcUPx+e07/GzqrxWYsb1+M9XLCPp0Mny04rjZucpuEVmy40MWp4xPrF+gLYPn0/3Pz79t/X/n9rfV395+QRJm+7VajNE++Qa7o6uxxrA7zN/QWYqxI+gbkrP44ate5qNxLcpoUqe/2SjlhRRHw22Rs6R9TlkA4XHEb+46VI4ODJd3dh/AFB9FLPEkFI7PERvDQHoG51PxVrT05O2uHvuEgJ0It1s3rWcBOLcMnjvOLnU/J2odLsfPwkeZR7SUkAB70JrMYXfVBaKxfs5g4XEcxcksAkYbqwk5moU3vA8BVisOLUI8erIHYfs4yshz588w3yjsigTnSrFAPlhLP4u++Sg3FrBDeQ7LioOHD0dR2anSVh5F2S6snHNcPC3EImDVmpXl6KkT8BmEZHk+CNfVx1LH2P5VDplFucuXYYZOwPOLpw+Wh++7pTz18F1sPixk0cXM5MByfqPs6+ifcurVoj/BDux2VlCfulqZ4F8dWBacK2/v/ACSWMAxy5ywg4XtOEp1j1UdZsHiqTru2mqariuRChEVGAr+8znBzlMpBlEEaUWiu04sTphrdCM5j/KrHkes6T0xTlAUeQx9LE5apu7Lli3Jjq+yQHOqzqU2V50cR8yCYvlygsPyTmKcsAhxh3cIXueR6S52ojhBzlHpLa1UvkJLwVV4i/yF8SyuVZz0QqDGFZtz6UxZMI7bCzFDelCgGHR1HHplWU+sMWQN5BtfUiaAErHaZL5BDnHunOB7agr3Kvj2CH2npYkbFMbfcEpwrvD0JAOqq8yYHv8CEUgccLXf5Rfi9ejRY1jlzkKJtBAlyknGZh/Hx88rN29ehgJtA4GKGaeMO5WB5vvlSy+yQXWyPP3pp6lgggCbnDDFyVKXsWrQNcrd9mNHj5ZfPPeLHJFtYN3V7M6vWbsKRTunIoGHsfB5RjLtsZ0ffPxR+dEPf1huvu3WshJZ7SKyicqQ+Zww9PnPPYuSa4jjsddmk+TQgcNl9aZN0AhyH/JLLDdZXKpAufnO28ol3FIMGDwLC4m5LEx7oach+to4Dip+ulFILFqC9QtjMiih7+IqZReKHZBmbwxyOtt7758p//HPvo1yBcUJ8XaMCVhjuIlPmW3rpfQ/b4efkUZJ9kfFP+mZayzdDIwn+sjNqj46qYd5eRz5pgPa76A9Z44fKGeQWVZgheox4RO0awnz6mIUF5PgWysBlVTdyC99bFzYNx98+C6WVIujJDT4q/OG80SCwCovU9e+fftQvsyHlzjno2hEweVC3EDTulspU3ssduRXoFSpO3DyVBbvC8HjUqxF1QrrAm8gXevRHciNK621tf49f36gDGnBjQXfajcvWbgPyhfpdzS5bF7S37zr/YTytAiCaVe+zm2oAXTBryeDb5QstMONZztH1x1xNyYalSqx/pvEcuvGLR5HfDfxeNaQSRWV/F/FTKV1ZdtRBNP9+w+Vn//853wfKE89+XS5dfsdVkbMINrFOHL8qHSuVhvOUyggwJEWIcPgfZKN6x1v7Sg/+aufRHHy9DNP0ze49OHeqSJ3AXxf5VUUgtSXdYpNZPxLYyPg6oJBjOG5c7Ee8XQu106NskorkxwcAC7lQ8LhHGCwa90PPcVqnDI8zShKkFiaaD0yHF5lEFtPz1FW1j1HpKlokx/l6G/w0Jx4auwrrfHmQUO6SNf4crgJo2DXMv3RRx6CjlA0Uka6KWSM/JMJXVqWvunL5jLJtjbXtb+bdDvyqozp2mtSpjPP3PDa3zT/X1fs9dW0yrr+wfUpM9VO3/091H+14iQFtqFCIg+mp6u8+qYNynrblmDO5me+/TBJhsUNHdM8zoP6MHxpus9+T/VnLAKEPDCwBByFslZaCyAHeUzFA+l0YvJVLbSEqPinYBQ9NPmZGEk7aYwTgsP+iuCwvf0oThCKbI7lqev2xI9afziHWLmqfvMm+BT5gyl+V18/gbWMjGHuK7YsywoimHPv+7bNx977IwsVMqbNpvMXRuZj2uAOuAswkusF8zHQnYtdFwIeweiCy8s8tVxvalnOMAGHScRJ0DaJJzgN384+vkN7ahEwFtK5d8fBCO7xGWThr+Lk1MlDHBl5ujz44APl2OEjZce77+GDeGNZChM/fuokE8EJFn6cXMDCe/369TDG/vLKr35VNm25AXNMJkZ2fBayC4XehB0qBBvMROfOhwk5SbHTI6xTWpyMXyz33nVjeYgjiZct0uQZHNBuz3MXLiX+tCEIBjdBaFv7ecNfUEo+yR28hmRa+K/4qTnGwMW3E+PkVE7V6eBUl2pdRDFIzPrPwhpnSk3/K/CKPGnB0yOGymEsTuYjrPewa9eFRr2PSdngsN3ATsbAmW6PEssJXmB4ZF9omhoAESsQIrtAUg/fKm54lGqsTsXJFAuLKfA5wmQ0wgQ7QT+h8srE5+6Z1kUqjtxNsM+t04mFkqiPZ4DsjkI3E/XpAfoMelLjrjR4FKGtm52ItfQf3c5u1Qn66kxZu2J2+aN//mx54K5tZfd7O7Jo+fRnP1e+9b3vlk9/7tP4sy4p7+3ahTA9Wm648aayeu1ahDjwRpniP+a8wRdton4VViozs8vBPdmgc82lecb9Htymnn9xVzl2Cv97BIR6rKALmCoYkYtWWU6alHvL0LIgz8h27sK5CC6amXvFVQMhJ1Yg8gRoyVg8KvIAliEgD0AlQBlOyonlYXmk2U3pLoFLA8gnrMkL3JRb+49ExiuVZdwaFM4J0KanKMuh3d5noJGuOafCuMoXheOjR46mf2664QaEhjPlnZ07GA/3x3z78P59LNQQoi8MlNtx1fmDzz8BHIPlR/TDZz7zefytz8dM9Lbtt0U5eeAAFmLLV5Rtt24jH3CBLy0R6mTP2EHQ8iSj7BzZoD5EQgQIG2y6lktTCAnS0hncAEYQCIwNEcsV8uRqfWcY0tB0M+0zj/WEx5knD2w/t2m/uPTefDz3nj+VX/UCJ9wYyM53Qr7Sc8afL9Q02GSuZGmVY96UFURbB/3OeyqEExyWsdPH7qt10lkpe6b+WrY4uWr+ITlVXVt/yqhF1XZRZFM/CTPtr+m/uf2t+qntqvp5vX3+C1shzU4VBGVm4yP0tY4jFsxmLhEXXsLX0J/HQCaNt6fxT0Ira5417Yg2Mw/AP/QjT27q1RJM643X3uIY4uH+svODQ2WIRerl4QlOlznDogqTZcaa7jYL8GVXGeHRlT14BK4htskS5oS58M3x4bPk3w999pWjx1mE9aNsZ464yILYmDQLMI1eAS13IfTv3ftxNg9c3IfHgQyFapWgCuceIbxuwzp0jGNZJHokZTgRTZb+1VfYBsecgZmXs+BfsHAuyoZDCXA4j13oKE3I18VuLsvu7Cy78N6wcXNZzAk6R1jADpw5xb6FivGse8RkYBDPxh6YwyJn3crlOWr5wAdvlycevKM8+9mHy5oVKk4onD6o/ZjOgmbtC/vFkuwv6d37Vhp0HJP6jIzksEPJiATDYDk5cL68tYMYJ1icHMDi5M0336e/2MQAD+KlX4tIdqR1KVLBpGLiHBYOxqVasGguiwxOs2HX1sVhoxDRVF2XGV11xKeLeRd1mqPP0n2DBZDlCc1SXJdcvKicXICblfBrmeLiwTgDxgcwfokuPS7AtYKZQ+BGF59a0nYD2xzyRXHCXGJ8qhb3BVWMYcrT0jSXzeYf3DWWnV2JcXK+LEBhMhflSQ8n60wwF8dCCl7WhWtSJ8feRqYB2mHmwFF43rjurlpJyve41+UJxy34HAsp3suESZq8MPEdqJdeBgRbzIPWgE4vAZ/KcmUzFU0nUNgtxTpgEQvAjz/+kL7tLOvWLC3oqGg31rOTuF8gI9zM5tID999d3tr1djm0d1/59Gc/W97duQurkGVlt6dRNAAAQABJREFUPdYkA6cGyqmBgbII3O8/dJhnO+H3n85CbinWZStRnugGYUwN49D1Ih90oohxs+bjD94vf/mdb5cvf/WflG3bby8nTh4rLz7/fDl+4nh55NGHywcf7S+fffoLZR8L3udf+EX5zDPP2CqsuQZQkGCVB15c9N52++1RwHy4d29iAD340EMoO1eXA2wYvfH667iqnS4bgPWhhx8qW2++pc5xDDRdfuRFzs3KVRB+uQRv2PX+QPmPX/8uOGfR2oVMA86g7sw1kV/ICzrpv4zc4Fo+lbmHX/W6Hv/2h3TShSVsFwrNCdyfChY93ViXnaftw1fOls3wBpWlH7//flmI4mHJoiVYVQ+VUyhjnYP7jV1En83m1JOdyDurtERj8e7Y7EFetd+vYKnl/O+i+YP3PyqrkINVFvbCkzxFSvcRN/zEnfKLrq/OwTl9BXo/hdL1Em4lc4ipIQ/01KFzLvxZ4JMZJaZyLzGXsFiYg6J4COWZ8HfCs5SvupCvBymnh03JSRRVHuc8hTJ0ClnC2CcqTsSfuJdUc8/verAA+YCpdow8p5VO1VK2pwlpLDTJJuYNKPmeefK+shWLk05wGut1YLTtvp+NFwS3A9DPX/3VzzmOeF956ulnyvbb7wgNROFMXaOMZ91oPN0rihDaJe/Q+gYPr4Dy7u73ygvQ5gro+QGsVnSfUZG1FPlSmhb/niSk0kq6UjGle45WpFoYy5NWrloJX+OIa/pbCxE3f+0LlbNxwSH/APOS8acWL10Cfo2ThYKH38bHc0Naa1QHgUeiG+zbzWr510IUyL3A5EPTnG9UsshLDWNwBiWj8Ai3Cn5m8FjBSMNzbCsbwwcP7CUUwd1YJXnSUO2cKIPJ4/CwL7waV52k1aRWJ8p3fs3lu63H9bYtwVean/n2w6Ta/6ERfl9VOln+W6kf2oA6mgamafXj6qS2X223bdl/t9umrOa7rZSrk9p+td22Zf+tbm2yHebgrjf8UrDzd+tqHuWn1NX0KANYwgzFpRynMN5nEFiY8U1kIrE4aVx1sDhBxITwW/UwibJ6atXvu9fXL5+QzL3CzHk3SheraT0Qxly87uXXdNo1KTXLTI7cmTk3fChUMaCdhL1SDu1TqPDIRYW4hbjqOKHnneS6+mO69IxGS7CsihcHdcU6ZYt4khvTNJmICps+hVHwOIj525Eje8vxwx+VLbjnDJwZgAlNlW3sYqgVP3TkUCZZd+LOucBCOLr11lvLG2+8iXLlpvj6jmCWuH79JtrUU/YfPoZA2ltuwmJlBCGsfxZWATCcSSa6zqnBcv9duOoQHHYFu4D2YyYfNe/2r01Qu0MfuQhwckn7m8YGURUPTZK/6r2UMZMhQVZ5puJkj8cRG+OESTxuJEweKi4UkhTPKtZqQZ54EqUZ+OxA0z/KZHv88P7Sz+K/z8U9iq1u0npg3D3ArCWOYoFClQK7cS0aeJxdmOr5h9BFf2tlYgDeSXaPDHqVAGlMEioQmD0Sw8QdMaZMTIqBC82KwrQChcdDZ+dAQdC6nDxtL/hSgHSM1BQnlA4mkFMctXkmbVSRpv/munVrMSVfUX715uvscC1kApjPjtehsu2m1eUzTzyEm9YI1kYny5e+9NXyf/xf/7780b/8F2Xz5o0E9fpmdoIffvgRFkTrmOBUSnC6khMe+FMYCt6AoVpn0GLgDWkCo0oWUafr3J6PzpS/fvEtFHIEZkRRBGHSNnuh1Y+2g/LsEw1yqIAX/at9Kz0r6McnVosTkmNxAK4tw/wTTMiacEYZR4J5fDujjbIdfw1stdyGdsQgQjT0r4/yODvi+jFrLjqfnUV3lrzsD015/bZc/wJneroBmWcIz/qKu/vsUXtHjhxgsr3MQg6fdyxGNGW/ddv2cvAgu/DsYG7dvBnlDmamxw4SdHJteeTO7Yyzl8sjjz9ZdiFsKxw8/amncjTrznfeYcxtKA8/+CDjG6hREilUii6DHLoocAwJTb6QpYpWS4wp49G4k9STY6072InHVYexzmwf3qDSK24ulkMRCuw01huUL9Cf7kF5Zvkkp4LcJr94rpBAm7xnLrsv9Mpvj9cWd3UnTPBrbvPUzrIu+0vYxXEw3Kqg9UXmzC18214Fo8b9wEWBJUp31i+fk8cLZuiT9GuL7EEIE872+kOESby66mTjo4HKLN5f/e2vSsBJn87Ar2vqbx7xQurXhz/EipAszHTL1TFOGN+18ooZ88h1+EqNdtd1V1NJ893KoMWFXZuXHRThv2CP+ykWo2PwrOdffhdXxznl8MnLZdeeA+XioBacnDTD+Igikvf1Z79CwFXnmZUsII1TcgFl3+ljp8rFMygiTh5gkbKWBfNm+Bx++2xmjOFyMXduX9m8aWMUQ3s/3g+JIQhjaekJN90sSIzVAOtiQ4Fxwxy0dt268NrjKDeOHDvK+/Ni8TKIG4r0MI6wrgm3sSRWrcbtB+QcOXQoi9057Ci6yPP0i9QBn/DqZxfZI4o9uvgyyhzHIpE5sxhyESH/Vki3XGl1Kaf93HLLVnZR32DBQWywcqk8ePfW8gRBttesJO4Sc4uBP9s3R0J3YUSOTejCfmhRTGi1oaDg33FHhsx/LLS4PTmAsiQxTuZwqs5FFCdYn2BBOaSFiAsWhPdeFsLGODFehpagutlocbJwIccR01ZjLBn/pxfXpWpxUmOcaIljnIB58FPxPqxLD7vFsxjrnxjjhMWg4F42NgG80eCwOY6YxYgWJ8IgP5k71/kWuSauOihYtDhhkTWEW4L9E5xIt5TlkJCCbbJl64riHNuHAtzYYvPY4JmNNd4c/nqxOmE5QzZV4vAy5L0x3Vrp6wlN5OkrXXIm6cNxrVnhW861OeHFOZUBErcN3m1VBwwteSDzPzyHB3ZFMrQyyRO93NE+deIUtcvfqI56V61eA94nUEIfqhaDTz4CDi6Xd3ftQHlyQxaRB/bvL48+9nh58aUX2IS4sdx91z0szD9A8bKv3Hnn7eWnP/0pyor95XHyfHRgPxYsuFJ84dmyadOGSrcoUESOc4/X7nd3le9+97tYjH6h3IJM5jz44ssvl1d/+RKWjPdzCsqect+9D7GgO1B++MMfla/94dfirmUw2TXAexrLYhek227bXv7kP/yH8sd//Mdshm1KEM03336rvLfz3SxW3Szb8fY7Ub799//mf0h7bXN6TASkJxwlWLZikbVzzykUJz9AMcP83I2cR1bpwOfZec84aCGVtPqk/g7e+Wjk8BCHdaUe30fWgiamLuPecu50mcs8PY/j2k+fZA7F9WYui1rjiPWhxN2wdj1ywFg5M3ASWFASYl2itdAgz2/ddmvZvXs3ChGDKKNsBHfLUd66ED8zcCb9r3XJTjY3lsDLVDa6gJduVSJpdXCJ8dXPeJrC3NjNyDFgkX4NAH+KoNZQWjauLFvZ4cylc+XIyeMoCxaVFYuXlNkopmfB2yZQOl7G4kRl7zKs3SY51vki43bWPKyWWNQjMqqthbzdcBWLorTyfm6544/EWOBLo62+yWZwnivjwkcyZ3PXBd8jHs/GNYvLgxydvXHdcsYTrirwD634oliWrlF0G5fmxOmB8qtfvY7F7GFcde5H7r85MlEPCp5LjAUVqgmWjEZGHtIovZURPU3L9gv7d777He67y00c17uQjeEF/Am7G1Ba7Ng2N3izscu3inkDV6vEUHklL29kDXmx70rzuurI305zoprKnrnw8XkcI+94NV9iwdH+XtqmO/c5Tugyfo2X8BrTxPnLS8sSSCwnTmmlb+iEMxxt79rM+cN5IjBEpnCOBA7k5ktXLpd9WO3ffecdWJutBD9uWglhyL8h33RUs+7LQz6Cg+ZH+6+rH0zn+J1umrKa77ZCrk5q+9V225b9d7ttymq+20q5OqntV+u2Kk7aXqi3bRmbZ5+QlEdJb3s4fTt9M432mZRrO6appPluz9lK+4Sk37V+hZnKDJv6WoTETxlAwzStsrkqX/WpM5fv85RyXNBWi4QKIPpeJkNjnFwsL7/xEVYn1eJkgh0IxloGjsdv6apTaVjmXuu1rl9bP+8q2Ft3/TS3IMC4mdBrakCiENKcQJOzpvlqez3t7W/So9EG9qb9CmQOTgMVqb1eyACVmbTXLyLjg96qi7kobfyb6hf/lu1C03vPQHfSC6PgtxYnRw/vK4f272F3bhGWCgRhg5ncieb0FNYme5hcFHjccbpM0CdkknLDDTeWN996i4n/7jCMSzC3tes2AgtC3dGBKE5uuPkmmA4Lx0wsTHQId91TQ7jqbMXiRFedOciP4I02GhxW4VC+btMqw1eouRr/17a/IjniVvomfeQHl5hzvonFCYqTsQl2PlCcqIAQXz6Nq1CEJ36JaP60QqnuQK5WPOVkECscFSfEZHfdjPDZyYKhl91ILU4UuLqYyH3XXq+U4MQvXCSCY2kP1QiKE567Y8DvuNnQ/5NMVh3gaBIhdYJJxSWJKoAck5iJrgokiu/Su/FOJsBXKpTDI3EyPVAfhfImSwf6totd0yP4eRIsDZPbbqwNLrNL5okSq1avLjvY1RrSB3cemvFyGYH/tvIMipNBLDkGBk6VP/jyPy3/5//978u//tf/qmzcsL78+Z/9OWaM88ojDz2KKToxhBACRjnVQDPNXgLuGg9EuxcbKpyOiIwV2m4aaKTdgMdEtmfv6fKz599GyMHMVgsbJj2z2fXNZVOuGn/+NpMDiXrOYULvboCLJsefsrKCGevx4Nr4BtK5SjkvX3NZnHL5bSnNn+VWSjMnFzjtoo8MTmZgSk+VcFG+hMWSOyG9s9gNQWEk1qmOsn2bbz/5HyUAE67P2IoHbuiMPvbI4b0ff4Bi9GJZhQB35ZKLlbHy8COPlv37DnKcHeMPZelcuraPHbIHbt9a7rhpfXnll8+Xzz77BQLu7S7nTp/HBPvT8Q12J3Dt6rW49DyGMDDGbjK7cAEIRQJ05WAKTmk8JA88WOGg8FLc8CjEIZQwvShOOBAripNBYiUYHNBdJHGa3W9v00Y+Qs+YGrtdhYtVsGZ91BX+06J3aTumqqaLEmmYm1jFkN0C7VvNt7U4sW+0GEn54rLtt4HZ+EldYNc8ZG7nf/ZvjnMkk3ECwjsRWDQ3t+e1QArdWDxwWU7IKN8mJCX1q/i0Lq/r6I/8LcCvqj+Fpd2BMO9an4JV642aRsGW2Vy1TSlRSKi35m+eQzLZ/aNj04eddOwkyoQ+fOlXr1hEH4kLcgu/9cOz5GiWaLsxOLu6/jzh45MuaB1MReBPEcIjX0kZuupMlO/+5JcoxC+Ws5c7iE3ECTqcntONK2E9Rp6aGdxj0JPuBh4L3AGjNBbJJeblUaJFDmHNePrEPqwbFpd5izdDP724UoyVFauXsWO/AKuNyznOeBh3Nt1HdcNTWan1g/EjVApoabJkyTKE61EsTTiVB4HdRXuUpvAj6U68W6/WmkuwfBA/HrU7SFDUBGfkt4oElSCeHiPGNOdeAsw9KNs8ZcdjM6Ujj/5OjCcWRMa/0HpPhcTylUtjmXGefGfYuV+9bD5xLeZE0bn95tXACD+UH9tzErd49Ou36P92+rMAuh6FNsoSFCdEqioHjl4sb7yF4oTF4TAK4pFRFoos5DzhxtMeerDgdOFgTA2V0vNRnIzRXvHgrrm7wgZlVCaY5+k78FN3h533XWgOezoVC7bq0mMQ1FZwWCziBrG4XMCpOhRLGQabNcbJLHaGDQ6LxQmn6qiU0Q3LeAEuetxZNq5KLFPIYz7dSaR8y+5m4Zmd2fSItOucoRUIsRJ0ayUo7Bw2eGYjq8weU3GCgo4ZFAZb50zm9VGslcSzQdonoR8Dqfvt6Su5d3FG+ROc3OZ481/l+97JtxlDIp6/dvz7THjsB2U2x5oLc1115C/G0NBKeB6uynPmeArRMdxj5pYtN6zDWpWApZfPl089/VhwrhvNU5/6VHn+58+Xm2/dXu65577y2muvlXff2Vkef+Kx8vLLL2Ue+J//1/8FK6Lx8s1vfwsF3c0oWx5NMGYb6LHG8mjh3ck8/r3vf7/8069+BcvD7dQ3XH6J4uTdd3eVRx99rLz33nvBtfPXCH1qXBn7fcumTbE8/ejDD6GdkfI0VgRf//rXo3RctWoV7gaPlp27dhHs8vnQ1Jo1axPrYwFWQv/j//RvGBtaKsJbGaviyvl3yo0ecHSZGCc79pwo/+kbP2Tew7LMgPzBKWMC+IN10RmMVtya2kqq6fy4av5PKh95n6Cc0G0X8eYQVss8rJB6iddx6Mi+cgzX1tnQn6cK9bIGWLcSNyVo7fCxI/CxsTIfhYXWT5Mo727CenY/CquVq1eGXrUsXb4C6xP4wEn4i4qABdCo1ra33HoL/GcJFqPHkAtQDq5cheUDAZcpqw8ZZvgCylNcDXUdm+BvHJyexYqoH5yspkzjz1xGjjxy/Eg5eOJwFMCbcK1agQvQLAb3KK6yF+An0u8S8D8Jzi6Bgy6UnxMoL66wKeJGlAoFdWZaZ4B98MEfvEm6rZvI/lZKrJdytHQ7Ae1buN8GyO/ERXuSWEGb1yxhE/PmnHw2RZpx/zoYM55uGDkZPq2MMAwuVO7tP3iUeDg3EbB4WeT1QZSjw8gwzlPd8J+4uqBw1X1FDsgD4EXBDu7lrT/5yY9jCXfX3XexnlgO7nBJAi/C6CxULWRR8MCDPPbdTV55tRakWnpIO8p/8vqqMGGcw/sdf1fI71jVPWiOeTNHCgMXsoYWLCrVVRRqxSJfMqi1R7AbUkAFi8+dQ7RwVK4ZolwDYitLqlyZiyJN2VpR1rlJa3k3xhznez/+qPzyly8SL/AxFJ+PAC+bigLEZb8GHa3f9GTS60fto7aEq/Jen96Wf/p2+ma6mpmUX1/cfyv1Q++I1Q7+BqFA3rqdxk/T4ObbB8198z2d+bqb9hzt901GpoB/6PoZvEJSr1q/ArVpEfoyuBs8tAYQjCA8gVz5F8brGwrjvJxVmAxDga+66ryG4uSV198jjsMKlggI9imAgYbiBI+1Vv1OPAy/v3X9DdQVPif62l8zLTJHhUzmn+FfXwLmOmLqz/o50/4ssOCCMsFcfLuD51FemnmpkZ1e+JHBen+X+sWDfy403R10UtEiQT9x8T9CfSePHSKa+0EiXW9lF2Ig7hrr1m0g7zA+5Piaw2CWLsVXEwhcYMyHibz86qvlvnvuAd5BFtsDLOIQFJBlDh05xXHF88rW7Tenm1ScTDKRTLDL1DF+qTx07y3l0ftv5qQQjuuy7TAsdxkj+AKXGnsnBR81E4AYmm5/blo4I7256uRb8e9TmdUYHypOdhNXY2yS2Bz6f7Ja62KCkAEHo64qvbNc8OHkJ3242u8gr646Rw4fLLOdrIUCBt0FXnrAJUaRYZ4ep+liw3/2US6/WmUbx8ZaupzgaKAmkXXhiBBHuSA1f5NMjKMuEH0e5iompCsL85uLPLkxX26ok12y+ss2scBiojl4cG+EncVLOP2CXdsTmO8aoM9+PXhwH4KxvqI9mBZPlc89fX+5/85tZf+HH9B/h+Mf/f/8yZ+UP/qjPyrrsVL5r9/4Br7Ci8rDDz+GKfZy+tUYJ2Lcf2AFvDlypxhrtFIg+R1QAJu+5IfrM0F+7+Mz5ee/eBshBz9UTGOrEAhdmT+l8Ok7NDZlkJp2kxA88G0wNelQQT9vMckFxy6KeMtFwhgCt1Zdee6bDb7Av/2bJ8It4XHVusAxQoBTo+4Hmuwax8HF42oEmi4WHC7o3EFxYnX81P7xfX+3YLQ07j2izwlWU9QhFhwffbQHg6VLCK5rYnlyAhe47VicXGIB4q7gIhYxy+b1l7UEmFRxsnxeX/npj75THnvyU5hzn8Tfen957NFHIgQrFK/H+ueB++6rizrqVxEY3IOjSi21H4QqfKB10pU7PC50uxB2NLE+g3AwjKDuiVMjvpsdWXAjUXGJIfFjqXGH8p73gjNw4JMoP8zJ75jQgmTzu+gU4T7Pji3vefrWKGnj9ptlk0e+rt+yOJU/AQaCVbVssX6f5zvvVGJSOcoAitAO+SEQsRikjxR+hV2TWNutD7MtqHgAbl7LvWXx3H/Sh3BkBPvcpvCWl/V6X7+9E8/NU2593/5vS680XHFk+U0Jv3n+ITfZ8wb9aS2gZTo47OoVnGTWAEMu81VQWu3jB+jlaur0vrksu7bflORgjAh9yuArLZM/CgRtGoZO/stf/Li8/Kvd8Ka58NA5LCrEFWbrCMXSt4xel8zFS5YSB2Ehi8WzKOMGyuBFXGuYKxbP15ybeYFgh0ePY1XJKwtRQCzEOsEgmEcPc3oOFlmzCOyqG6DCqAApQCvgusuboLH01xl2yU+j3FWJqTDrZkN811U6M+Z1UVmM+bfjzyOSVXxangoZgzmmDmB2/vXYyxpocHYWWLofuXPYpXKa5y4AXNBr6eXC35PuPMlrEEuL48eP4oI0Wm7dujHHEd+8YWlZtRyFxVzjGIg+xgS8erqrRDiY/s39D+ht/M/+0OX0FIqTN3bsoYR55eCRsyhOPmKKIg4Fc+sQioTZjGXjVxhfRMXJfHfSiWtlOxcQpHEEXGpxYiwY4254VKe8zOOc46pDP6hsMd8wChatCbXcsjwpoQaHVXFyhSOMDezaEbcdlVcGkh1WQcZ79pPvuCCfOwd6of+Ua1SwuJjw9C6tVuexKHUhR1NRVEOzlCe9Of602jQuRZcLMhaqPbhjqDTpZYHSj5Kol8Vyt3MeeJ6grWP4hk3ituSiNcHeeTDOhOJ8GrccxrUbEfIWY5rV8eN4rfVTMTg3f+Xh7fj3WcaJeaA5lU2XUU4dOXIkAZK1sNE1BlChO6wrxs6XW25cVZ584v4okbQuXbp8Ydm5452yf/9+NiW+Un78458Qa2xtuf+hR8rune+hAHmnfOpTT+Iq+x6bf6+Vf/e//zvm69PlL/7i/y0P4D79+BNPZJHuol/XWcemir89KNt//OMfly9/+ctl601by4GD+8uLL74UZcmXvvxFdsD3lz/90z8tGzduiCLllVdeps+Gymc/82ksR9dCQ2/R6Mny1a9+NeNKheTPn38OS8Z16bMTx48TVHl1ufvuexiLLCoh7Fu2EfSWTqPHgAP+JIFyZXcfTDm2337/RPnP38DipBslL0oA+5SKBJv+xmoRXJtifzTyr73P0xb/cZz4q17m9d5azdHNOO+EFrqxHJiH9VHXuBZAyLAXT5UlWKR56uOJwycJ3LycObgLi4izWQHMZRxYknFSlmN5uwMF02Zw42La4KNalTomzmCtK7+Q5t3Q2LJlC3xlHn0yQJOxCqHPjbOh9VS/MV+wTHZRPYJSS4stAzSfxl2nH/64dvWqWM2dw63o6Kmj5SSWpv3MUeuxSli1cGnpucJmKTxNixM1UIu0VoBPDjM2elA0T3hSEcoVLaido5RvlF2dmuXc4qYq/NysoT9IiOub94qDmZeka9dK8DVywdWh1ytl4/ql5aF7tmGZs4yNTQJg05+RIJVRnQKYG1Qoe3TycU57OnjkeFmFzOEpTypTElQVeLRSmYMbfy9ypu5yjvnQCP0wd+GScgmecfjwkfKzn/00CpMnnny8rF27nnGOJQ7WaI3S1+DBg9C1G1a6BKocMRivSlwttG2tFKD8IP+3HpXc5+HxDGyUwMSfwQrGtY5u0vIox6tKFPmx/eJJY/J16zQOlM8bqxQ33zwtSqsXFT1RmgCP6x/nCY+b1nrF0ySrizBwQMPnsHx6k9AFP/rJ98sXP//58pWvfBl60aK+UrB95J+dI0yJx9RKa2jcx17mq7Rev39dmunXX83bPmm/b3LW+uV1Xp+c4x9f/fAI2UEbyIG8HfzW/XTS9E197bf8nH6bG/twppNaT/I1nYvS29Pb4Pwt622y1+LtrFp/lB788F+gqRnqfYCr6ZHb0rsZ+WS1jDpoHDgpkLecFBnjVweH5VSd8Ul2H6yHQZxTdZi8bH/1LfvN9Yd7Ww+ApL6mSsATLIk/HCoYFWYukppmBShhbiV7U+uvBQUC0upOa81lmgKFJq9M7ezEaXHCZGXB9bWU/9vWL65ibswEIv5dVDrJufA03SCVx44eLieOHyqPP/owky5Hj+3YBWPB3G1+Na3Nzi/SuwsStcoykh07dpTbb78DBneZ3ZejMDiOHkSgPodViiZ4N6KZjqChxQl1TnA8WaeKk3u2EhyW44gXG8TLyVTFCYxOXGmva+tDI7b8Gvy38CklzCDbd7jakuyfCJwkf0fFyQenCGTFBIfZaFy4qC1lw3ipNC9rFWGfOgnpIuQEb6eJD5m+u3k9/DYQWQc7AN3Aje6AfO4GAL3104f2M4TnD36l97yJIG6dpqUumDkNjzlxAgiCX31Y6ypJAdLJzlKshIkszWTS9+X6wIdcrbozaPjJbGmAq6PHDrFouIBrzuKYUh87wW4JwuyaNasJBsquqguEyeGyYllf+eoXnyj33nlLOcDCXvPdZ576bPnuD76P1v8yi/O1MVfdsunG8qmnnkYQWJO+6sbtI3C04KqzdAufwR2wAHSEUG61KFIhuJsYJz99HledAXYaFDqgS8jw6sWenZd2z/S/k5p49EPTcidHBX1/O8E6kRlaILvDWNO4iHJ3WRTSpQ42/0fQc6xZftLrR6oLkkU0fdxN5mHMNo2q78J79ao1ESJ87BQeqZ0+shBFkdolrW+ZUspl0Q7OXax5etWRQ/twERgumzaxE8ZYP7h/X9wDHGsDA8SGoO8WovS6cfXS8vDd28ui/qnyygs/L08981Ta7O7kQiZ74wloQn/HbXeWhx64L1WprIgSyKpjPmUbAZOvygslLR6CxwhRCIYeISmqT2O2OsiOemgT2nX8RCnUzv8sjLLorYpTf4vDNvyF2q3CwUy+pl/99iUVIpahKntSZal07SOZGv/z0BuBIk16kWdliJoOXCqtxLVCpGXk5Cno3QWiC1qP1zQ+g9lVxNBD3PMON9avECX/8xmfDsGwa2Fq2lKHkpXM0J/9LrCCmnu+U0kFtt6mMSS3+t+f1m9+Ybb5ltHQn/ctIqzv1wR2cs1vZmNs0GYWgYlxwphbvQKLjhYeyVDLIqvl++G3agevFN+6Md0s9Wo98YsxEvoV0PQlX3YYmX08gkvXf2IR9NxLOzndgd28BSsZy/qv85R+mkQZ0c9ifAU7qxhNs8ivJytcGTybgHxbWIRtYyHZ0YGVGAuJ117fWy4RpXDl+o3wF07lwYrhCnGxxiAalXIqvjzVwkWiil6tQeawK2h/n8UU29gB9p2xNQTSRb749bc7hctW4J5DZ587dyY7i7bZsaHbqKsB87hb72koC1CyGHtjiDlXSw2Dolqm9KjiREsfGZe8P0oT5jjH3cBpdofhAqNXhsvt2zZjbbImQUHXLHcHk8UN78uvVAp7P9P/ItXep9wGwW3jp+mx0F/6v84X43TeiQEsTlCcTE7NLQePcsIOJx1NxlUHCw4UJwaH9ZhT44v0sHGk4uT8Bdwhqc9F3zD8UDN0FRjiSqsJLQaTD4sTd27dfdXFUf7iHK/lmNYjgpoYJyhYB7H0WYCrjm2qp+pwzLCuOiw+tXjwVJ1LFy5HceKiR/rKccSUNYd85jHvfDeGJErorjMns9lTYgDc0T9xecVSuIcFkLHFelDG9BrcVjcdrI66cbfNIo02TxGEVAWKZDuOuZXLSk/MySl04h/amXDcM9iFu+KZUaKli11CnXHfyRggDbqu4+/q8V93ubEeBD9uRCjzrITejmJJokm/R1H3TJ4rG1bNL9sJ8L0AV9iFxHpbu25Fef/D91mQn2FT4rNYhbyCvOXpavOJS3IySol/8S//WdzEvvWt75QNLP4vErxdpeTTTz8di9+PPvqYuXxZuf/+++gPeB/j5b33dpdvfOPriTvnYl7LIelz86YtnO7xdMr+3/7tv8W1env52j/7w/IX//Ubwf8XvviF9Oebb7xpUZT5QNmFEsE27Nr1LvPKHTnm+8DBfYHZuDVzWMyvXrMK64utLJCxYtU1SoSDULAHTM4r4IYTFt/mVB0VJwWLk04VJ+RJPBr6JW69v4H+G/4vg2nmlNpv0IrjClruhBZ6L3vCIYpC4t6cPnO8XMSacym46+udgyxDUFe+Z8NPLqNQH2Fza9acGt+nF4XaQiyEduzaWdYTvFqa97hox7185gqLay0XdDfUbS/H3jKWnbPDN1CwqKQwoPp8rRBQ2mlNpfXFEPP7eVwUzxJjzpgmq+gX5eYr0O0l5CktPN1oW4QF79J+Nn5QYA4OKI9RPwFBFnEiUlc/1ljIBD2MEY7WQSEDXSuT+s+xIc7FN7/F0bSLDuMpShMUHsbsEc3OH5NYkGTWdr5UeuvAugQXw43rlsRVZ8Pa5ShJCKmshawEIW9X6YgLitbgKjg8EfDgoSNlOfKfLjpakXmqji6J8mxjvtWRBVzO32KI+no5Kecy8thHe/eW55//OW5KS7G6egoX8C1kYVzSjm7qMvCryqpYmqCQUoExB1c/rUd1mxyj35XDdMvpxjrbtZEnqUmzKlDso/lYIEeBB/z2jy5tyokGFJ8OXotieDaBXeeRV+WLsrvWRqJSeUOePcL4PodMpMWhCiH5p3OOCvTwCz+AW9xOQIvGtNn5zlvlpZdeKF989nPl88QI9Dj40JOYAJ2KBFn/AplyRyok3XJyz2c6TECmk6Zv8vS3/Zh+m5tfO///N1B/m+LkahRMN7BJvqahIt6+8qp5295obvNdB1Mra31hphdav6//aoqYfkJCO6L/LvVLwO7aWEclkHqXXwLKsyoemFIv+1L/9zyhct+whNouyVXJsmpQZQinzl3hRJ0Pyqtv7IbPrIA1YHHi4FXAxCRYSbTW6uCeqb9VW4sR1V8+dbFR6+fG/8BY8V/xq0lbJn1fyQvU4yQC4EJZ4azlmfZJ7Zf5ORGnZSKYNxPjBA2qPvsLcKtQu/13rb/Bf0zXxBzM3QFtJGqvCSYNzZMv46Zxw+ZNgNNRjsIgbf98Jn1dS86fcwfJc+bRLCNczZ83m4Cpxwimt57JZBRXqTMRdq1jDKY9e+7sshzNuSihoihI3MXXVefeOzYT5+QGFCeYyDkJuOMJwwr6gg6ZP1ADB8ud36n9tku6cJ/52z98AV/fgTJKjBN9R91hl64Ur2KSjrDfLCxJzK6Vu18KAQrgQ7hAaKKor32sImCe3TBtqdB1qOZ6EoxKMN18sotiQc2feShvXMZPWyOpSEy2FRybXzjJwiLOBV4VLhKJnvSr+t/Xgam6ZFiHCOYKwfrQkVTpSqXJCIsKg/RqkuhCQjNdd6zUxl+6wA4HQsfq5bPLH3zxsfLIvbeVM/jfult4Jwvyjw8dKO+88yaKlaX4g15hV2JddrXmIzS7i2ZEc9vcy05zYmBIqwiz2cWzceJZumb3EMBoP5+0bffegWlXnRrjhFgA4FD3Ai9b9Enjrz6BLshwEdg1AY3FiVWgfNFsWLFA/IwzqWty7oRKAuWJ3JkKWnfpIWGrozbVQwtMuI4ROneYSdngfQZt9oi52QSEnGKS9yQrd0cc61EyOI5pq9Vk/LcqcGzp9+zpIAarPYvrVCdWOasQtlUkDjDheuTrfHayjQswwq6qi4KtCDIeR7x6QV/Zu2dX2bZtG+0ZYcdybxR59qu+wpu3bEH42RCseUR1+h8gxF8uv6EJ6S+NVSgKnCgt3DlhwatJ7gCKKNuI5CYyyUp/tfgvDWrtdlkOPxA0oPSUGZRSNkPFAdeqA5xk4IMP84eZi5eZS7lsit21KBCBW/znIr/FKByH/uk/hW5S4JUUZTNoAzIMxdIGhRz9fdjZAsEIXSwG3VGPmWyLx1F/L88zLh1bKpD5lv81gpK8rh5dX8EIyGm/gPLfNpoYSMEv91fzfx6l/bZE7LVd5J3h/23p3NZ8CrbkaXtJPewk9KdS1mpdIE7RX7BfAgIvA24S7W/hESzyBuck0TPTY6nWZoa2wmvizKd1+cssySa8lAu+TRhFOP7L771Ydr5/EDcdTuIibgEYRbHBW/Az42Dos7+UnUUD6J2GxsdRyE4Qv0TeeMtNm8o98HxUGRGMd7x7nIUACjsW/cfYVR8fFmfQAvxD/tmNsiTjhnG4CKsEF+wO44u4g8rDJK1+lIvi3zGuC00ntDSPhcgi4FDwvcJYko8ZY6SLZ+JfHiM/cKfeWCmeeGEcD08OOYWVpZaFuq9qQeEcaUDRccbcLC1pFuGeg1WF1hanmStd0CrAj2E9sZHd2mULu8t92zdx9PIaYiYYIwCkqmIEWNsz0/8i2P6oVzN+5NrtPWSfN/KH7fVko1McGf7mjvdA+XxcD87mVJ1YnKBgGgbXs4C9Fze1i7S7V4sTFKwX4Nu0OsrWYWA2OKYBc21n46qjhc5ZdsGbGCeNq04vSo5YphATQlwvxwXCHVjjmrj76gZEPcWKGCcscAZZ6KtoXobFSXgZcoEnwxi3YZSFUxNEVjcdrU7mU0bFBP1D/1aLE0hK1ICAiW4oGTz2QBddtLEbntetxQx9go0J+IJqpHvaWvrYtCEej0NijL/GotMFJboRimNmoL+cm5rLPskftJe9bPJGkcv7nzT/OP7TX/AfF9DnUfi5qNJ9xR3vYRZbfT2jZePqflwsV2CtQGBc2r5k6XxOL9nAIg93NAKU3nzzLXE127dvP4A6blEW0icPPXB/2r3rvXfLAJYeo4ytDcQWuWHLDVgq7iVo8kcoQLYRyPVO2q7SqDuWvj/76c/Sx1o8aPnjptXaNWtiKWJ/ff8H38OyYGMCwL7yyqs0f6ps3749NHD48GF4CTGJiB/x1ltvRv50jt5+y7ayiHlfyy4Dg+rytRBlwcbNnJC1bg18CD5KLxhQWBzm5BFlOfjUZfST72hx8vUf4rKHqw7KeXmUnF13E+Wmqy97cob+pYlPxr9v8RQCwSmGIP3MgVic9OOy0zOFsvbiQJQTs7GM7sFK4wpugj1axSFY6FIziPyqm4pWcnNxr5mNxa1HSi+mnZ4OeAUFobwh7vG0qZ/Fsi5oKqNGscbsBCgX4v30OVMI/KW6/y1bjLUG9NXHHDPEnG5MqCsor64gf/VAMAuQvwwoq+LS+mej0PR4c8KylHm8M3FhsIzST2MjlAdfn61lQz+nKKH87yIOnS7cIxCx/Bj0RoZy/SEq+KzfwBO5VtTCMAyC7598RKKdmkT5we1EJk7Gk8e0E0stipN7t2FVrOWfp6KBXPL4nvO/v1U2DqHI+Ojjj8teaGHdhs1Yi2GZRiwbLTC0SPGExQRnZZ0l/1VhpPJzeJTTjMCp7nH79u8vv3jhxcQIefKpJ8umTZsDWzc4ANTg3RNrPKloNsqSeVgu5kRIno0wt3dyqtN4q3zld3mIMc2c9/oJ9DtHRRNK8FFkLgku8ztKE8egimAt41WQ6H7oSWIePaycqLJGHBmLxcs1mMriIfil/G8JLqK6K43C28L3abMMy1PDpH+PrL505Vw5eeQowZj3lCcef7zcdcdtwO7JP+BevgE9pa/CQJztZviQddo3V40KEnhlOs3xk1en87a90dzmu46kq8q6vnSrvOpqiphOJOEfU/3II6JARMko2prXjplp6Gdurss/8yilVRRf0/zmZ/PdhsDryvs91s/4odOtcRqQ3Kf9DQoc5A5wqQNYTDZ3xRRvQqjKcRJg2mqh5KquOlqcDKI0eb+8+vr7uOrU4LDygClXY9xMdcDNUyIES8E1HkSrB8xC/WoDm/rJfNUl7HlegbrqWS1X6CpMPhTK1CPzSbIv+mdl/BcffMfkv1W/76jFdUdHwd4driz88ubfvf468JhyZRSAmwUIiRFYWdW6Mxb/b9MATuFRywVh0Q9RwUr3HncNZT7nYfbz2HFxl9DTS8bYWdeiJ8GdEBbVGI/ABNnIotUTCKAy8+EoTu67i+CwWpw42TLRCIsCfzQRtFeFSUiSe68G/8FdTWr7lF58Ywb/PnQi8fi1byP07/7IExIQ1lWcuFpDIJOVxbUBmLPokMnxp7GilOEC3Nlq0GNksThxMZagm7TJRZr4qi4FEhpFKsTxXv1vp9byIs3x08nB57LNaTLml242USoAs4u3WCHZAHDDa1y+41u8RxmV6baLH7YbWJUo8kZdIBj1PccFM5H1stDQ7codebXtLrx0nerkSLpVy3rLlz7zIAv1O5kEBmOyupzFygj9P0CgWJUu9tEsBCCVJJ1q6REiPAJTIcTdWGGGYqgeAIHF/mjaXYF2t4DWQ0u7iXHSWJwY44QtgUxoChl/0/gTF7ZfFqppebU40VUH3CBETqnMArH+JTisE5s0wX+Kzl/FD6/wjp0g1oRZdlLHv4+gCxdxlGccBd3Q3P1egbJjNvQ+ix2UTIjQhvDaYjUH0/RnXaZZhX1HPxgoLAsIJt9ucGdEf4DMDvc4WoR5xIYw99AQJ1gQCFHFySP33FrWEgfoPAHv5hNEdor8dbHIZI+7ne4Hnm6V4Gu0JErWWnM+5ZXhc632S0O+L/5cROgGoRudNm1nzuKqQ5mmqzyxTWRtXY4JcEJbXVB7igqF2DJwaiZ500z7m10WC3EICUeyJX8wjnLSZxV/dlDy8LuCX+sRXvmCL3ufAHNkqGkmZwSz204+FuEeka21nrtfWpyQE4FOvNC3EWB8B9qEX1lPLHRItwp32TrZhQ9BAHdtlVjwqvXbfsDw/zWXuKAeCS1vtj7NS9km11d9s1UGNCd8uRwzyUM9teLAnL4LD6B0FSUwiV4sTlZx0obtqYCImYon+8gGWAtspD5v/84TE3zUVj/jUhzRiRXeVt/WXCrCCQ77yq5y8uwoxxLvZ5NiCFB0UoSvIPQuQgmxFGHfAKCnTnG8KQvcCfj8JMfFqrTYiin8rbeshFcwZlkE7D94qZw4N1xOcBLeufPs7PZI23YC8zSL5Qin8AQVG8bfABwWKVhEMRZVRDgfqRBxbqzzoxYVC8mP2TVKs3ME8jOvPF3+J1pc2Hmkqu+7+7uYTQmF/jHG9Xk2DIzdJT3NQllgvxgLRNi7GbeLUQgZFHAMxYquPPG7512Vz50sbhfimjN25XR56pHbyqefvJvjRymDenXTcRElj/zd5Z+ATzmFzYlL5Y13rrU40VVHixOt2uDF9Ie4iuIEa7zzLO50V/RUoRHyVMUJix1oXVN0eWgsTs4SbwmczgXfKqJ0XcipOszzBsCUXjz++TyLSHmiJ+jZL1GcgCN34VUqaZ3ZBIfVjWGOGy/QrYEbPfEnJ2mgOHFXWYuTUKODg2141bXSv0qkSXjkJEp4mHkWmB3gupe+8ujZbvkWuYw1Is3SUdAt83p4EApfoYWcnD3xoKDPzUOCfxngrfFsZbTB6p2r5OXhS/yslyPJq5Wf5+E9pERhBxzGaHCxNkEbpa/x4QvlthsWl88+vr3cuHEFmzW4ckhvbBLpwqHyZD79Ms6YugjNaf2kG6ebUx7p7akdXmfPDAApp01xkouuBAcOHogyat269RkX9pWXY+H48WPU7UKxI/3gZkJ1P2BxB0yXUXypwHIMXGAXXb7nxpbjw36WFnqBQyWLc7qnjui+JR/1UsGmu5tygMo43RQgcLDiDryKE3+KI2YS6O8KMU6iOPnzH6LQQnGSvpHBML7pmMgxIp4SRL0fv4383QGu5tD+bi1mGbuzRi+XWdDKOLKMsTQ6UYhERkIPEMu09JuLbxSMLIa1plPJEQUpc5ntSdw5yjUOivcqOpRxxY/ysUQiTp0sxJ88yjgZzkme2jOBSYj8YsjNiMytKGCgeafTXmQTyW4EpQqTPwFf5yJPDKKo4IQWcNjJxtwklmtTU8zr0HsX8E/RBwY1nurjOG9g9XjtzFHUp7JfwhXjknaI2L6wI6yI9lFMNh8Au0Xbjhjz23eOHSx24NEb1xoc9uayCQtXgmjVZ8LLi/IUizPuiCcSvf32jvLhx/uxONqek3+UebSUNf6U8pcbRPajeFImyUk7xD1S6Sve9uOK/Nxzz8eV78lPPVU2btxINa4TtFyCx6NsGiJmii6CC+FDfSitlBedm2xbJ+PMMWwcEpXDKrmk/9nI5bNQbIh/eVKU5cFTbbOBgD1tz/WMZbtmMUisJKhCTQKWb3oNotDSisW1jspllR8qyqRRlSS2z+DnKoXkpbpyT0ALjj3lOeMKGhzWjTFdxAzCL/zyf/EvXr2UF6bnfxMEvPXMn9deV83X1z60XAGc/m5luC65STBn2/xv9n/k9bPBDYQVd63WXd1chdoqjF6dPp25dTODgk9OuO452ZL2CQ/ak34f9VumHZvOalUm0fjba7p+8DLDQLmHGSsam0Mh0jecWfMu1JinDEh3CU6d5jhiTFdfeW0PA24F+xEMBIickUXVTqWZThkj/vb/31C/z4TFuqi3AVC6dhD4/jTM3lCkYAJK6M8MlTBJn75aaa0XAzuFOfCbsnxHM1Z3vhTuPUqrsTixzr/v+tvxr8+di18X1rNYTOmreIUdJIUTmUqanT5AaHVRSZ/KXBTKjNDfA5PXrE2TuPino7BSkFcRE4Uu+fVJnkKwePC+rVg3GOOkBodV+FWTX0WBKhiLYfEvdmh6xZF4F1mt36I9P81wDf5lUr7JEq8V4wRXHSxOPL6R2lIu0hr0YE4YnEWk/yjRPhfZ7GQLwRC+2/uOHK6KE2dCJlYFcBUnMQukRK1mVBiJJ8sTLns8xCExWQP5M6FbNs/qP8oiW/NL+sq8aA7SVSelGEviNXGShS10Yym5vEn7KcUqeVKVXkw0XLHmgIn7QvxJEQgSawXz1g6EjTUrZpUvfPre8sh9t5GZcaJQwsSdolJ/VSQpAI+xg6Aw0olZpgJFT2vCqfXTv96k/eDG8SNA4FOhVMWJi8/d+06Xv37+HWKcYAKe4LAIvuRz7sn4B1Cbee34ExHiR1zpqmP8kASHpd/sUZGgLA102TUYZ0Kzn5pLvqpAYcrMulMq8R9XU3/6v/qQe6KSCwytazTXVHmmefW47fK/gAJ4TqoBxQ3+hdEy7e+4VTC2tGQBIPCndY0NqfgaducGulKAnUS46h4fKjcR5f6+Wzn6eQHWOChTelEADA5eYKe9TuKOO/GV3RIaYzBN485Yr7UKnG2sjfKGPxrvHTXDXxBsUJR0Mtbt3bOeqiO+Ql/QL7iSjuVpokP+KytNd/LMwmv5FMZNVXzMPK/THPWRt/7VstJPvk5hNKH2s/3iXwpkAeUY4z79n/ZQh9UkDXisn8tvWzSB9AgGuGenT1cdaNcAmPJu+UrGV8Z50xa+eb/yG7lUC0YXBDTQkdVOf7X91sfjII9v3g+WmxvSr+f/PqywcpN3LMskU7310zYkFx8N/Qc3SWe8kVFZVouqXnwDV7VcdVKW7zeFpX/ESdBXy0wd9SNV56MtkVt5lxA5NToO7YaqtPEhymcQ+PZuTqVhuD73yx3wwgHSFCrnY4XBcZoqYsmq68LglXF2/1yka+3BIgw6W79mRVm/Gl5fzufI4X0EmX3vA9wIh+GbNKy3i515lF/1VC7i7EDHWjWuWLk6sJxk992jLp1XXNQ5X7gQt9nSkYL7KvIa1+nEcY5YZ3HoQzGR8UE+FQQqP/tYEK0j2LU0pkXKBaxYXMS7AHL82RPOayPMZZ4Mt5J4MrMZcwrfF7DK0Lqy8nlKZ/x20e/9vZyYNXCwPPvMA+Vzz9yPhR5xPVDiGcjRHU7rrtDQ0y38W08dq6A4T+vH9fJPzZUYJ+ygN4qTA7jqvIW8Y+DIIRY4Wnr0z2YRj5LGdmn6vgB+dYHdbMfiAna5Hd9a0KlENzisi/gon7VMgZ/qqmNMEtuvK41K2cbiBHCjEGkUJ7rqCHgdb1ickNeFhMHgjXGi25O4isWJ/cWCSAWD+Yax1tDFwRgnFiKf0VpOurNvHWcTEL1KDxGmzpBRjKUJnIrFSxfzjjNtlFHSLDQ0BT1CDBnrjmcVL6Hn5KQAnqmoFf/NeKMafpBAHv9FZgh/tm+afHnBLIHLG19R8eE87cLPZwa+dgNp6NJp3Cz7yx987u6cHijvEU7lKBVX1q8CoxsrmaokAa4W4wA86JUgui6UmQd6CHib+ROYPBFPvFTXAnACTblbLn+zPMewinmBSewH+R3vOZ7tY+dB8S79C6cNdPEnvi0rshy/3SjzXp7iYtg5xfgPOYLXDuK/PRb+QHWxzhQnCnnOL8BujJO33j9e/ktjcZIYJ75HnvRzXTRajpefFFvLBk7ngmn+28qTx9Yt/uSD0LKuW11YdfRzOsxsiKSrR8UH/IMxYDyQTizidLnupb3OHTyJwsN5o4/xkXoZmlqiWL94cPy5GWacE3GmAsuNodmMC625tL7LxQsqWARI5ZlxilXADNEnKsLcXFB2k3dTUNqmEtfZJhZQeYaSgWc9KJq6VLzxjxmYORmlCeNzvAOFmv0BrFp/5ARG6q3zoPMUfS7g/Cl/VKttE3xCmdQtvXlBBZRsv1kAeDF0wSTBYVGcGHNwC65kNiJuPsjyWmoNqYQyECplHDl6NKc1ffzxAU7VeYBTcbZlLHeSNzFBoGHrDD1wL20Po0i9gkJo8RIPE5iK8u8nP/ox7l6ryzNPP833GuZMRjZ1aa11CXcq26HC3NPNxsGNMpzrEjc13KxzknUT9yz8SjrxKHWDeyubZkMHZEQ+Vd7lfRVf8jn7Uv4+F6WUMhygBkZlWAPDOo8YJkH+prWjrmhuZhljzd/izvmH0vOeCkkDx2ptQhPKPBTuffT/ocOchrhtW1mIix46k1y1Cyp/sXfsr6wD6mNBMSmX47WRb9rTW4+nv657dk3CNT9r2XyGHprKWqW15/3HWv+0xUmFuR1kU9p/t99/QgtbSc07NTefcpxrEBNu44C57rq2jvbf7fetFz8h6W9TfwKcynR5XygyeP0hFyRlRohoVZBMzTNfg2ztcYhfovIfI4x3VZqQzOR1kt2YX3E836u/er/MmstAxQc7ixmIF3ZKNS6hvRzglGB5f6v6A4zQCki9alLg0LTYwSTTgHWRh7Kdl5JXSJ3QZgZDLabWL14c5DVXZToGUZMxyBQ8RUBGbsUpLh/8/Huqv2m/u8XjLooVShB03VlwsafG2Ak70xyMJX7HThY8c9KQuRn7xDbIVLQ4GQP+Hsw41f6OMAGrfc8kRgu6ECZHr5zBJPWW8tj9txC4S8UJjfKPyU2f8rrIFCNOSmL2Gvz7s3W1sMuvFv5DI6Smzy2WyYan3/w+rjofKezrqtM6Vcd+0+zPOplUxHANsEXNmVhAM+aXHZgYynz3HzmYCVFtupOD5OciMAIa78Yfnt+hZ3CV/pcWbAnNy11ot0UcwCqTaj2xMu75M6/PoA2baolWJnwq/SgygqKTaMoO/YXy+O37pDvJIlwYw8bFuYFNFZ5kAU4g7riOcTR0N7EIdNXZuHpu+fxn7ikP3nVz6lOAC92JSMp059ieMH5B6qQM4TCzZVqnl7GEbHh2CXzI4Izg5/hllQ8Y0EpHeQ/FSU7VOaXihOCwEXQQ4ISfGvxfERis+SNP6liov+pxxPjMY4kRKqFciBJBCSi5V2jXb1nFY/BMWhb3lNZcDdaspVUDd6mcT2gVgUpZVJg9lUChkh+MDxaGWbjpRy/9KCIp8DX4b4HPkyrQKpzWRVkVHrXxoE/sZ4DpxpLHMd+DkOIxhl0Tl8vN61eiOOHkKY7Y6etkEQcMWgO5S5YxwoviPnSHgKu1k0J3cA7q5bHps9qcwG2/6tLgssOdEJVgXSquwPw5gmhq7Sby4Uqk8SJ8QaElvNKfIg44OqWl7KKLN+g0dEKZaT/fdr00ElZR2yhNVIGAdyhIWOSLZo6gbM3W4W81QiKGGrWKsf9SjzRVk+tzCqoLfFuKYKjFCQKYZu8usH1fvNVYFfVdi1UhYz84jmPeTD5rEO4yNuIAAEAASURBVCbf8T8Pcy8WrSytCNHXLCTwxPpppN/NJ3lsf16vObg3s+W0ip6+makj5bdK4TVIub5T+ROwIQVPomDr66uKE2ky2SvlVTyIf7W/9X/qy0db/TOJ3tX6ZZLeBZfpBH6Bm7QcHjjMYmHXh4c5EWKwvPjau+XYaayiUHYsWbwcJcmcuLpoMm3sERewjslRgneKY83FlyJEcnAO+Y5zUtwirEMnykFcTToJDN3DjmUHLhY0gz/GBcLxXNxAs2vOeLuEEuTMmdOUx4IEaxUXfS72xKnK/lkoCwyibpDXKyzWTxw/Sb2MN+axYJt2ZNEJvoznsRgrziVaTuBuo2WKC30br1DtTq6KFhc34tdTedxx1MzbgLSjmIabz/LHmC8NxOpu47xZjJfRM+UzT9yD1ckdKIfYMZa+6At3pxv+Yz3BaWhBHPMXGqn0F0DMURE//UzeouLkpDFOds5YnLxJcNhxFlW62A0RP0H3pV4WMh6n3I2Lgi44F4k1ouJ8gab1wGo8GAM5GuckxxbDl4zzcg6lkMp2Y0YZaNdAmbOiOPHIVfoVaIyv4bGjjcUJBJNAuvJFT93zHU3tV6xchoUCbocoS2JFAeMfAcYaHLYqTuKqg6VSbb+LHMnWloIF6Nc+cM7wt+nipgsxrguchTvxwGcGgyWjTymqzom6O+SN0LJlkROcQor1ynggKfin/IxrH8L1xP01+De9Gf/yrBn6EzL5SOVZvjY+dKFs3zy/fPkz9+LuuiWKsyjKIUerA0pirTm2nE9tH43NH3eZD4CcgtzQ6SSGyHhcRMitAhiZywWpO+rG03A+VFmi8kbZTP6ljCD/EU6S4hJiEFMVCgZ3VQlQGZ31wavIXMtR6XwZJSixhMJbeA7tupEmPahQszymmnoxTidRdjanTrnw7ADvuqwMYkH5Fqfq/Oev/wiLP2J1tE47Sk/IH4E3PRv8y0X/Zvqv+KdauwjcqDjpg5Z74DldxAXxBLpeNyPQYhoniTB75OXkuwnGtHUoAzifM5dYSK90DS9Q3pXWjD1muVqXNYtWcSMvMA6QQsA8aFV+o0WOz1TMU0MUXBPgwd5U1pXm++BJsYYwjfqMgaH85cl6E5R1Bbmirx/LbXAxhYzdi7WVp0sCeZmgf8ZQPk0xjjt07+XeYMfG51GODqnQCtsh/citpKWghsfjcVtlDudepVFc3sGBViNBBX2rAqID16aOycFy44bl5YkH7yhbNnHUL2mxgKfffX8Sfqgrr8G6Dxw6XF544ZflY1x17n/woXLr9tsTDNexlbkeKJpA2saW0+JPizOnpBUrdNmfzPHM3//e9+Lq5YlPKj1cTmilNgB/lT/MJubLLNyh7YfEHZHmpEdgEf4hFMSNe450PQfLXV19pFEtAN3AVUErTPJ2g8HK/1WWzMVyRPcc6V0YKTUKQZGn0lcZTPcbFfFzGF+RIXgmXzSzm4SOfYNeq8hXoeimsS6S3WwU62p1EPzcc9ftCQqdkw0pvOH/wk+zSLHHQqRAIBTNRYbp3+33reefkGRpvjP9KY20F+mrTcWtYma+ri2w/Xf7feuNT0j6h6ofXp1reoDWigWs1doAVyH006vBw0zq1Sk1V9tnXuTDXvqEy45sGMQ/SP3UJ7GkLr4k3Cosm8bICJg+gIWaV80iHM236jv1299qUcM9JAZzwMzUjkZx8joWJ7rq9HNUFhYnzW6wiyDEJ7IDRUuQ+Y31I/REkKSWoJHqrLFCJcCye9lGgOd3bUZ+t2D7/6h77y8/j+Ne8wUwwMwg5wxCDGCmmCVKJJWoZNnytfdv29/2nN2zx7579/rca9k+VrBN27Ip6SowSLKZxQQm5Bwn7vN86u1vGAwokhZ9tC8w3zd0qOrq7urq6urqJCQ48wnj+T23Kj9kqDEsWdD4IYVHp130uFvKtbnfqvNJwS/cpT8dWfN8mK6D66wm1gheE0zkHDxcWVM4yuAdJuIAwZnmmNJ5ModtyQmdBcieV5k5dagm372PKiYcKldhijh7+RTHn93WPYZ5oKfquCXCCUycw8roFR7gKRlMrS/p6o/0596cxlYMcCeG3w2vG7+k85/buHTE9Vd/66k6KE7m0eCz+uEKtRYXtj2VJwBPPg5BsqB5PIx6mk5ydOIAU33z8FsxHczeTSoudScoBxngyagjJCsaGRiEKo80Z7GDjl7D9p83aCWkUqB5rKcrAqBmDsYOTk76k4WwgBnFibTyI/8Djjh5J46XQpETfmXLnFpE3gpFk0xAPD6NLaPd4tWT7HNd3/3RNz7bPfxpvOUzkq1mcjavEwonsCAyp4Mw8HMyojUH40XKPKFnXGGKAj8O420iW0iVaChyTuQ9LtYJ/vMosf7pX7Q4QSjB4oFRJyRzj/ug/y/T/6Sz/U/yluIET/YI+l4KblQCtFVYYLBDsErbTR2LL3lnkpsigatl8yJyHqss6d9RnDlAIlAROMWExKEuCl9guFVHAcqkYtSyMpsI8KG/+RYM63s15VQJoZ8TV/HWQAeVi5r8EsD+W1aM3aMLn8I1XXf3pw5QH4c48WglRxoiUNEvFQisEtucbdzsPZHJXpDtQAzgg/o3DGHDvK2bwhOcyEClXB2/R9tQuYZQdY4tEwoN9me91xOxyguQCOQppP0TXRsAY3HTf2vlJAigwBwpf1ooYKtLyC/5x3v4rwn8Iwmf+av6jwKTeuQ1E+XEJZFRhvyXpLRt/kNThGZCFGQuQsdYJiDYildTMgVH45Nn8iEsIPn1m+1hAh4Y/mID6uEDkqS8EynpjB+kTeOX+k08fgb0NywVxM3kphFZb7zX+GMq4/WpgdHa/wKMSz5oGseZKICom2m36niqjhUROnNLevMwN9KRZfib8Je5rhn/qZzqWYWKTjkrUxVLrPAxeXv6V692v3rpne75l97GPwk+ZNZxeg7bCGhAOV3ElbfpKRyLgrcCLLvbKST0RsGwlW0C05Nz3XuH/y3Kuen1N3TTmzjankmGWyOzzQt+o8+dTcTdd3AvQupkLFhOodyYoJ/ozyenKhDHNmYZXA30hK+NWzbHgeMxnGxOaeZOe3Al0VMTXI0XN+myf/8+nHVu7o5ilaLSxImUdDSOJXe7hTgouK/DceDBmz/FaWpvczrGsZqEKLw72YKs9p/wRdrJrm3ruy1rFxjT7mCy/ClOxlJJSp8loltbiwtaF1XXQ/7vt2oTofxI/Vv51eKrKuYwdz/CqTrP/eoF8uVUHSxOnvE4Yp3DanHCtstpBH6F+/P6VkCh4klEZ/FNIbyN+OFwgugEQeVPTsthUihtDXOldQ4FmaflzWBto5WlEwiV7E4+RH07vnXOs9KrjxO36tjQLqiozIRHxQkWJ27VwWeEyhaVJeuzxZDWQF17jGeOI6Y+nFRtYGW5aAKFoKP9xPa/wATYVu/KvOHpf4RLR4wK+JU2hjjZ5BkyrmSc9jKd/cUFDZmV7MjIKk/lDwSm7chHvdIXuPsqn2Q0h/Qk4FkYXoP+Hzh87duffdN21XiM+a/A+uGWvVMoTu7vPvvQXaQ2L2sSvoqssTryJ/gwzppW5UQhCbq8L0B74cuL6ov9mfI4w6Q8q2HGpiORsXPXGtI3rajksyqb7fq2PRWOU0x+HXtt6ypelPGMmzEAZaWTTmUFFZNasbgYaPuzXziu2pfmaRtuqdDnj3SxzsVXTPzgUd/SaYF+ehFLj+ewOPnz//79bsUanMO6WGU8+oR8JhP68F9Tp0K583yd9h82JgyiyJ+VUVZp2YGcvMr2jIXmBOOqPkhUnFRZsRTAkiOnp1AG29UcC0UuHKzRkgewF5CBFmlr8rtawIS21KEKDJVxNkjbh4cpqIRUhppHJtBPXRRFlMZ6hN0w28BiFPq5VUi6Z9s6cF3EsVIy7yBnxBq2g1zBEoutupRpRqsZeJ9t37ZLZZVDY3DMdh36sSc9pT8QPAExsgghxfr2l8qWOKSfox7jpwdaKJf7VXaOVMcf/0J36lIfJziIve2mHd1XHnugu5Wjs21P0lqeTGGhF+0CmW8F7eD48ZOc9vRM9xscFN+F/7ubbr4Z3kC9QosopMBfC2DlrMsoTfTx4pilc+gNKDi05nF70vd/8AOs+HZ3Dz34INDwW8U2QP2SKZNv27adOcWGwJ3FGtZ+FSsg5Ga3JttjTuNLUWuvbONEqW2Zbf/yONu01qbyf61HzmFpYtvWIkjFn7xHgrg9x75q3soKWmKd1tIE/i9f3MTR7atXsbhIXVr/8shYX5FvLJCjkOGUr7Vr4uhWZeRFlNUn8A/46quvxFfRwRv2we+gnbS07QI4/NxX8lyV/g36BuYiwMtXG0Pi+yEo5z782setyAkb+0mW/FiZy1zheYOwRCbW7zd8+i9UCwXyM3y0gMuXc5miLyH3krIvTXBN8H8yfCc8XjWGUUjfB4MTnTvhfE+DIWIeJYakMi2/vtqxYeiyAyfzRtTMEbaH4uRc95OnX8Zj/0ssYu9goGSSTM8kGMYDI4QZmoUOHX8r/OTcN/Q8m4TUPc7mYyf2ClpBDhyB1bd5UR5c7TFxTdGXH74Hflo1GEN4dnicWKH5VLu9BSHPSYHpPgn4EtUOndHW2rGTB5q/H6P8FF4NvxPUDPbJjnyguUxe51MKFp/FMewjD+EcdgsesB3pFThg0K4LWF92eG+hU49R8JS44Gy2y9G/b2ApgiSVctb8d/rjiOc5jng+Pk7Igz2x4iSMbN8iboQtyQH+0QhbrypOsDh5ExM8VzdloHxkgAYAyDqAmo1tzdVQLR6CYNqEmBq7fn0rIYwvRHOSb2SFgURyNM8Xv0I3ccu7vcB8uZsv/1UIefcmTdqtFCy8F9D6nljmJVUJy0COyAndF7E4ObBzsvv2HzzcPYolUImf0sSoinuKpmZXdI9CCXhOboOKP8QNBu5Jh+qpRDsDYbZtlSmJZvGgl6fqPPmvnKpzDGdoWpzQByRnKbOIw5Xy94Uwb8ud1f1khAktK6Su4ujky8wjFFrjKnxI7UBadcst6evdfJwcxcO/eQop9AcKeGY8oR4UcFdoh+9lEYVr/1DA51+4D+8RTgzjStVTboveo14CMJmaXpFXXFSg2f5c6RRNhVnNXYW/iIn2Kpy2PXTHzd2jj9zNSr2KRxVSwCA8yPDb2n+6bOATJlwv0Qmges4jZYw/IcMtrwlN58kIXCdPspKMAGE7S1uzJYgPYVLJVdW8k8bpRbz48ySo5GPMVu5R+AYbh6uChWvcljN5+G5OKV+i9jk3+CYhfBn+m85Hna8izGmUp114SkrMp80VukVoNzm5Vjsu+PY/eYTwCU7b9+5VJWvw+/IvAz/xB/Qn39CUDCxSf9WjVC34g/ITd9D+jNuA85jySn/5PwwmExgUbpNMPvexVafRuuqGhMYNgfnCa/XjQqChMsjeD6G1qXmEfuKR9EYizGCpo+PgK0y0/uff/LB78qln2O64ttu2+1Pd+i1YdaLzO8XpMhfPewKLykRW5ugyUxxxfnXWiTNm2vgeObB3b3f+9HvdM//r77E42YUfjYPdCgTkC1gnrNHXD/1xBmHUrXC7d++ib6zgVIrTrEIeTx938pPJCrNdJ/cKvRvZ5rGTU3wUdI9iZaJVygpW/Rw3LY1KbFcHdUSqAkXz8M1Mfo4dO96pYLGt2CP11+Qxw14qFUyzZ8+e7sabb8Ts+nD3Hv4j6KDQnYkM/fYqOEeRQ3+9jMC/b9ee7r47D3U7Nq3s9u2e7m7at4mVVMc7Wh24aPXmKqfXNe1/rP6rDhMxj1aEJammFIsTZJxnfvkSeKzv3nj3dE7VWXSrDtZ1V7A4caV70q06nqoTi5NyDqvSwOOIneBYRld2XSn1+Gcnz3UcsafqaHGCjxMmG1fwTTClZYqKEyZ3to8dOofFIaqryRtRmtltc4oVypq1LKJcJG+VVDlVJxYntVVHGUZL2ingqthyPPWEi+RBvmlrtjeLnBLza7+tRki4dKD3QC/HavtRL80lnkmUAmslnbThLYOcCJaK9nQvfmW6xCmFMq3AgvDn2Cz3yzPxKp0pCj4JCKtcXJAwXk2gq2adaK7MNst1KE4e6B7CgtM40n+VizHcpaPfiv8Wr82YHdQcbwoPQVWLaZyIdD18MWj8P9gEJ/NM8hSRbAZXe/y47U8YA37Es/SY10ccspEl8AQ35T1XNaVgTtWJxUk5h8XTafCVn8Q6R3x7Xio9ruH/BFfdFOaBL91a+aHfLNs6mAF3k24H1EqTRblFj9QjbxdTldsWMT2Jgoy+yJJF6BHQ+tzI+Gc0xg22raQe+aY1j9WjZbFb17SuEr6KVC0h5T+xwuQ5UznadixxUSDIPy2f94ws5FP1KF2gD23adm06nehbRpUaE6RNG8AiJYqTKErcqgPW/JlGhRSAAnsisElINuHzwiQvnUkbybulzRtx+E94LcyVpTK9ZxUKsNkL3SEsTb7+hfs5Pns/Y430MwxYGRuhM7iqAHGx7vA773avvvIG/HEfPJwFK3CyX3sM/QTh+jiJU1Usz6zezShu9Z80g0yjktlt/H+PI2NPf7rvwfvg44u4WEAxTlqPJlbhPtMvxLpI5zgQ3MHLU7lO42fF+t3GgvJaFDLyasuoUt3tnQL1wACt58KX4Ov6M1GRBQmhDPUHL3dMdVHYJuuC1Xl8+6gAU8krzjXW2larziy/lLZ8+iXS8fh6lM2WTfrnxB62Uz///K+773/3b7uvffUrnJz1TawbzcuWUJe5RI6DprYha8jyVRvJKz8f7jKdl3kmk8GLD+PXWFyDBkD//wOfeWJY0EjJUor+faSI/efR0EGZq+Q9AUayao8kcrJSLHeYygSjXyv6KASfvaiO/vNoqCH1fp1AI3gRPAo/TENsaDARVI1jS5aB9MzQT5W79xH4PEazmngF1yaXtGEWMAry0uLkJ1ic/PwX+DiZ2oFn9eFWnUVP1Qmjt/x0SPIcwG8FNc88ex+FXwOrkL2Cclr6ME6FVGqzXlr+fABoyg/ODb4d2I4VeCajo0ZxQme2Y23BrDlm5gD9JOCnMOQ8aJFBrDp0AFpEvmUV1nJxfWD5wV+x1fgOrGES5gljtOwrWPFaxK/GIw8e4o/jiN2qw6DmdoZo840T8ApJfYmXwgeHvruHrFaVlzdIm29Wj3ukHcE0zvRUHS1O5rA4mcNRmYO7QlJt1XGIURniWEEa8vDuQEbJicvgwEDw5ttvIvTprArLHCbW8FnCnVxTYpJRldQl73JJ0o62/0LK+ocZm7fAHBAR4AoPvqS8ZJTL93oap79MWKZbKwdBNpkZtzFB0y6Fb3Df/oiXsoHHSgebubPdfkzLv/0HD3SPsn2qBmLiiw91oSmodSo6GdyhqXf+9ygWAlabPgqgBOE1eNexyeZjucE9wsXK7kWcwz6Jxcn72aqDxQmDncpV4dTVZ56iDNtfBJVE4ZQnVgg0vdd5pJeDXZSM0lShqec0EVLMVzSAkXojX1fYGs2t6boafO+UQ7zpn7bFClFACSmFmDylg82GX2UNfitOq39XDu0CilfV/0lJ/VW9kz9tIoI1kwozVHEygdO2B+64CWe993CqiCt7KHoanzAzsfEu3Hbz7p9B/KkETD0RxQ8EUQ3UNxF8HrQ/2rO4nWCieonVYq20FBJKUDbPIf1NBrqhncJpi+N38xQdr8DnhxL330Sqb38gmSLkxxT+BaP+mZvv/edr4CerFmg0nhF2WRRK/euwL87pWLkTvsDsMzQQ8KJVAJ/H+u4EqgckplozFiYfHn5D3XTm5vWhy29l9fCHqZND+AkFgMbQ0XGCjrmA0L5mjRP0rWnvFCzAzEVy2v6qEfBu/TfkRmCMfhVS4CovE9c2O0zCF95Zl0NBMtv97//Hf2ObyBvdzgN3dJt37OUUkVmUEBw3DE6r8NdwFQfak2zX0DLqCqdcrMU3z3YsMTau24QFxnom9EzGzx1lm8j67vU3znenznMaAbnPM4HQP4InF3isq47/3kVAd6udTqzdPqajZ1dZ7eOaXW9GaN2p+Ted+RTCtFspTZftdCgCdHR5kS0pbhtdjyJgBxYQ+iw4i5XJGU+IgZ7yClcKPfFG6wctv9YjZOufw1VLLSbOooDQL4UmVjrxdEXT1foZxjHjbMPB7Gbyv+Xgnu7grvWcToaChqPdN66DB6hExmRCqwuPCZawaX/eJbsVZv1LcztuLmvSy7D+1seL4gSLk6c5jnhxcR0WJ2e7p595BesMncPq9FLlFYoOeKlbpnTgrqLI44g1k3erjttqVW7o9LIpRCz/Ro5l1YKvtuqgOKGMWo843q1B6XTRLVig45G0OoT3GPvNm7aFF6S/QcfR44h3EU/azWBl0k7VyRYGrE8nUZxcRZnlKR1axFT5q28WFargNm0V0cP+D9380NMvowVR9VsisRhx0n5DTp8d/4vS5pRHx1/lzwoq+tvG7Ti9epV8zHSc/mPyD7mZX7qu2TKJVTnrlW3Ncxe72/at6779jfu7zzxwR8Y1oa5yhtzDN7bbVZKHL4xXliF4ck9uaRt8hx8vhV9clVgjRUw2+SGJRaKQKuaNkg/k+nHbX+AnI0klfj2t9QHnu5UFvtLcmriEo9RnXtbHCRYnnqrDCTcZG5H1VMSqbvko7f+a8ktHrANWIL9g20FubDEnb3GRbCoJlFq0uiqejjJAPAf1L6HcuoIVClYXWrIpu2mxEksg8lDxKo/wGGPbi5YTLprJZ2xyTuzlP+ahf5uJyIVF6uBAfUt/aZ5aoI1IHXU2xIr4kDDT81EY8sKrBHnyjQoYXRDMxzJB2YKApAW28gPxXXSMHJMcbbll8WHzneMgBptcYAS2fYzvUMuFzZWrUDTNnOsOYXHy9S880N1xy17kQSuSvIFP0WnbOrQlV6sL5fW7HJ/98itvYVHGFkYtfcFNerlFybEpFh70bf2trd+IU1UtmKiXySmtOi6zjeWN7n/81V9jLbKpe/SxxzgtDSfnpF2L0l2L9jiJpZBXsS6Rhlo1idFlLEd0UKxPQS1/FIUst4sj0tQ6kTc7RjgW6CBea5FptmTpdiBzLGnA2GH/Vz4wrgrcU2zZ1FpcP13rtIwxH/K2brP9jbhaebk1W796llcL5E1Yy8X/CXlqZe9R0r/+1S+7v/vB9ziO+FvdN7/1NY6pxiJGuBLeyqNQ1qPVaC+oqw/Li89exh275as/9fk6gS0WwU3+rE99/FCr4LeoLcdhPJ9+/+DTjpXW0pXSkEPFYSny1Iq55POHfh1L74uXtZWrQvM7FrGFj5Ny+PXDP41ly4udoHo693qpBsB32lGeRTCsxYbG9wxghvHiP9PLBExQEzi7lAxFBlOKk//1cxQnT7+IgpvjC7uyOLGTxTv7CqfQJnegXY7+EqjyH4dvqmuvxCY6WVUBvHN5K8zymmLLOvOQT5XIX3omjEN8vCw3ihM6qFpuTfy2bN2ajlrh478B+x+E3/pz5SxNxHMp/cfhtrfrwdfiJPUFFWoiSX6sakcwicXJhe6zD9zK36Fu15Y6jjiaYwciCa82HSAMHQWqv/kSmPnpnyvG2K/B0t9LJslcvfvL7/4IHydHkX/1cUK7iKDiwMZgm9gyMdsSkaF7KQv8QitEOJAZx+IEQVJnWAyXGVicPFp/7vYp1MmHAU+UxaMwsHnwFmKDWZQLfmMAG2koBidNT3+pYP5yi8GVCObW490AtTsR7UWWu+AL1pwc+fqMRLQPXaUig5Xh/ayUqjhx+5T9xSj0qkTT4qTlVXvByau3qqmJWmFXcRReFBMpmzAVHIrdRaiyLigep+qc7P7hh891R46xVQdzT5Y9U87U+Qe0P3mBl3edGSq8RHEivtRD2pEwFYIKLXC1NYaaVbBQQ+GBeJVdH1OaSXN+6Yceg6mCrLWlijTINTRZcKXegvtjwep/ypLyE4T8wvfiUYlgjimH7c/AWtXR4sSaW2QFTausB+/6VPcljiPegsWJ7UhpLfROBQ1yIqy/gN2KI8685rKM1f4UJflTeRJI1g1PCJdaj5w8iZNMJjNORDyFQlpYpiK5uZELL8mLnNI3BFhBde9fl4ffsKtEeeNHdlwho/S3/RlIWAWS8xDU4FlAxPP0o6wSIe3pZ0EfLFoimN6WGJxb1UEQ+4fZShurrYGourLM9pgKHIffyt/jkkyCQn0wT/4+bPmXwjdttT8yFjB1rQm8z8hz9FWtF1Z1e7E4iTAofPDMrS+HVIwq2My8Esi9vefDsPz5Lh/wwTysD0FLJWnFh8vA/b/+/K+71w+f6aY37+H45zUcOXwJ5QLOw+V38IMJeLyOSa9ijn0Fh9P7b9jKthj4LcqLjRwTfGD3ZhZTrafV3Y9//JvuHbbprdCUGafHnjqymy03TuaPcmqNyhAnOZ7q4f5369atoE78Pb1FSxOPqTx18lT37nvvpr48AcbLLVDSX3Ns0+zeqYULx+1yRPL7OK/VWiUOZilkHCoDR38EWq/tJt+tTAoU0t9847XQIg4BJYr1AJVc0NDc35NRtm7nCHGcKu/auhbnilv524KzxY0I1fRlBgVPmtNJq1Yd1djMwgoZ0n/Y/q7f/oU8R588iuIkPk4W1neH2arzi6Y4oaxXsHTQQmQSWGdR+rTjiN2CJ9/NqTrs69dKRjpLW49f9fSLDaweZ6sOliY+u2VPf0pRnESJxFYdLhVQ7u3XWfYWj6QHsVJU6jyT44iZsKgUcUtPO454A/Xk1gn9AngM8jT1ZJ3qkNcjoYtvUnbyShMNz6E12vbyLOS+88rb82xMWqikDD8sfm39GJLpUTIrsvOpfyCBvJ20g55sPD730XmQ/5X8YV80TDAVzfytOz/ya8Ugr8gN5c0qAFfMX8Sx93T3J1/T4uS2jD/2x1Vst0y6zEgtjx3N3Myvyhf+7yN52/+cuQ4WtfgcHAqReubb0stg0fLuJY+pvs1D0lZp8usnvvlsYErW0/wa/mtcYsUq1zZshSFHmSaP+eb3ld35dqoOzmFXTrAda8ITj8RKRYMlL+lO3JIpt3owPTkmjs8hRUIHP3wLCfuBw2O/V2oxEtticKGv0uu4lywWHyfAjf+1EEOgwjeOYFW4ypfYpkneOiR1Ir0G6zW7fLapgYgKwKa4tf1oBWc8ZVZ5jv+8/LW+vecvxaBMAMufNCC9bUH4Sp5eNbdBocvHOh0HHKl/y5J2mNzMTHrbhkEO+OkPrV9ErsWqgjYU8oCbPqZS3h6hvJrHSixOZs6zVWdn99Uv3tfdieIk8iDxoggkImJUFONafrjg8PY773Ac8WF8luzFqg6FmNaqlF2rvvNYYqhs1YHuBpy1uuXPcXRGn0rIzfNYBD399NPdd77zl/T7zd2XvvKVHLPtiTWecKMCPv53kOnikwpllorwK4wfzof02bcVhYuWP1oFuR2t+Hj5ZpR/K/95MpT0Ujmi0kRnyPIffQGpCGlb1JTpz2sVB19yO6L+VnSYbf3nVB7ooFLHtqhixlN8TCNPVCGt1Z75OZ649foCiuJ/++Wz3T/905PdH3/7j7qvfPmL5ImVku3ZuoPmkJDL+rC10Hfq0Y9j13U+j8X5oJex9L54BbYPFZrfsYiG1XWdzy34t97H0vvi9R+ET/+wKrjGcs+X4Y9RILhxJPYA5jDGdZ+G2aZrXj/tMOK1ef2O4bciN0CWJ1TIING+1r3CRssvor57k2H0TCo4yophYNBKixMVJz99+gXmYrvodHQY4mZc7bfqJBu4wQB+HsZpfC38Rv8WGTx8rFx8qDoSTZ/5CxMF11qtzuf6qcABfGtIQazVcRgNQpCMwk67Bad38VGR1L97+K0IA/jQMs1u2fr/EPCtJtLW2M+A5GqMQPqtOiwBwvQvdo+gNPns/Ti+3I6PE0anmKNTXgU8ZOSioeAca00+Wv48+2MACXqa5s7XepVZOQgxiPLhr7JVh+OI5xHkMRvNtgPbEoOug5ADVMCZJX9RFpDWFS6GHjTTV1CcHIZpqr0uLbhty8juz89gxGhZAoFDGWGp/yF+Dpd+01EcoZTN9IWwOEszX6WX+dUECvgG8K2YL49ZWVFYkDj5mqxcsTCf+pUpw9Rp/KXsMD259HEcrK3i1NPV8zEx/+Nv3ofFye0M04aRd5CEFgomqQhKRZomkNhmglurIwFAuWyB6cUBW0Kc/5LQ+AoiDsJanPzDPz/XvX8CxQkO0BYZ5Gz74iXY8Is0RJ754Dd/g1fK4HHEWJywP1XTcuO46uPAKUWcrIRX2PlTOeZJFj7zKStYPd490QKhoBiXyvQYW8ptGUMngPjPfESt8DFTH/kgElz54jMP9n+tBeq738iJm/S0LPHnA066ZHV1xUSLTIBWoTh54O4buy985h5W4lU6iocC1AB4wUnOwx9h+9cwkRY++83fVcB0ZS5v4mjBqHH5Z7bqoDgRx6I/gcQxWuVn2cmRzIaCXIWMAOnh8x1YH8z/bAzBpAHwpbJqRA6sFmxkP/SRjOmrcaFx+Cj3C3i618+Dk7S0/wJCFVX8KnNlU8Aa0IJTdU2effzR8hfAEfhBZpieJ75IdWn328o/ki50MPWw/D2XSDuRp3rc4QIKjEksTvbu7hUnJiBMHPPoa19et721+q/QJb8DOvK9V5ykrJIJ3M1GuPrpuHhlrvvXn/66e+fIhe6t98917+Kk9NJVlRmT4ZFXr2IFsnYzjkjpkzRjFSa7cDh94fKZ7iSrlAf37oujzA6FihOWp378cvdetulx3CqLAzu2bkv96CDQE3HsC672anXiSqfCpyvC9vedKEIUXs/S/7UKuYSiTJP6KU6yUHh2v7r74z2icuee3dnb7iT+yJEjEYAVkj01RMfZV9hKNHNlNkezetTuOv17MP7qCPbcRRzgMhjB3rNy7YqqSjmtNfbs24P/sa0Iy2e7E5SPXSrdvm1T3WfuvYlyHmRc43hltm2wKB5lUrNIaDVgddls0xa9j1wVNt7+jDJQnGBxMs9WncNanDz3Mn23P1UHHycqTpwwaC6/mjHXU2tUnDi581QdncN6VLRl8JSUC1iOZLtjFCdanMxl66OKpMtMhqSpkxrpZ4OI4oQ6Vj6JxQm1efECEyPqy0mKihPTSUv9rFxhcqPiRL6sSX9zDqty5RJbgTZiyl48pviUZbf/iW94DG0vbTJfi/84ttquS8oQLePSWxwX+eq/PpfiP43N5HsRu/q4afnjk+lbvqbN8cXX6/8kMVHqj3vqFj4vX3Ryp+LkDhQn/+Wr9+LHga068PdMrrONpOAXLwLbjE+t1IaZn3n3SDMRlg7GGCCb5/5doCbry+HXevUD/Me0flhyJbrfQ4jxwAojcGn5rQhyzxjU3+Vz8oigkbGVsjIOc7pu99yL73V/9hc4h0VxsqLfqhOcyMdljdCvh184Fx7XhT9SSHmjywDBBVqt0F8Hbcz6jy8zT9QEjrJfUzKUOtmWAwTCFnEea/m1zHN5JEfdK0MwEVYeXYUCQL8ujifipOWaymr9poUfMcGHO8byhQ/Uu6UIhYhPxuDluKpc1q48QixxyLjVyk+A0oYWroC15kBRGQasZUCmALfIV+Qb2YE4xiwZVTjEM76wBU8YycrqhJitvwQn+wv2fgvzKE4+tQPFyf0oTvZEVpOyhR+JUV4EF0CdYovev/37v3dvvvlWd9sdd3GK2QHogcTLQuIZTiY7hwI58hjWhetxnh+raFKnX7n9B979o399qvvu97/X7di2rfvq17/e3XbnHYQjqzK2ibQKi1I0uy0GCxb4zkUUvS7+rYWPaGmYQw8kEvSoUxOrnFqIeBJYTk5D2bsW67vV8Hq3Acl/IhvSRnUAbB2fQOnuWOEx8273tixe1q14yPu1aNaXndZ4Knu1qstRxipYgKffzAnwsNwzWNS8/iquIp7+WfeH3/hmd9ddt2FpU1avqXfyzp2iWtrI0T4sd9k5lva/5eIt800YlS1tKDW5TCQ/DSNeG+H3EP4yW3V6vIOsz8tT84PKWTXS57Ns8j41N9rg8hA+QfjpxAIewBdJXsRF1HxbAr+PUYE00do+Yfc3hD8HF+6xOOFeipNXOVnneY6S2kUKHWIRQybiVh1YSugUxUmf+yiQ68IPGyG/QOV35OpxH34hwxSo0gyDfRJn/hvFZ178p8CfC/j+y6k6CGgy8nIOa/MPBZOqIve/QwD9h48Gv0ej0go/uIhblbWy/wjlp3DZhpAcUD+YlQxAmjuwavbcK05UnuzElHsVleRABeeyhqgvy+DQAz4kEfpo+c3ymmL3H0JX6C/FjKUoZZ7fcavOyzqHxZS8V5xkgOgVJ9H7A6wsgW1TxfBpOGRTipPX8So+jZmxq9kOzh6tGxGAsjkXFVdYKOUXmcI5Hv9pOTFh5XuzvjCKGOYir7Kgah8N4Y//YhH6pSz9Y4REQ6BZa5SCzGW6+mgug88tLHUsvGRfg/rVi1icTGFafC8+Tm5zGEtKaW5bdXiwZhxOy/TUAZraIcy/AOnzk6bu3VUFEgqkbAQiEEgtxiM1HLVVB4uTpjiJxYnCivkFfv0W/kVLSxUi56bi5EwGNhUnXg58rgCl9pk0FJ9IED+Vk+iJc+qhGtcgSAGwQFdrU17xdCU/+q+eWn71tQgQpBMvGYQmhbOZK9AXVq5e2KptJ8YkRypCA0QMV7JlgI+0JX2coDjB4uSLn72n28HqNdMm+pIqLZLQN/rS5D0//YdgYh7G429wt8IDjlq0zfPP+qj/2qGs6E6cOIOPE5yhUXeVp3EK06II7cD6NN8eRgBU7PwmlPZn3mPwE9FElBf8DfNSUByt0wqo0JY+8fhpuY5+L0QoG0Y0EXwQXnX4555mJ8iWv1Z8eIK3OcZEmdMyJTNJ42WYrCpE6X/7oKC1PHzjD6+PWn4zHoU/Wn6bo3SmURMH3PQfwJaRSU4su8biREQb7txtq9WXGm59hD7PRG1B3lWcGJbqNtQaNA08kY/uO//1K+90h9+72D39699wwtgJTLj1I7OWPoiSA8WJE/ULrODt3rur275rM0IrW27ee6M7dexEd+ehQ93nOOp8YRYBlJOB/vXHL3Qnz1JpTLi37sDKgxU8nbW+f/QIsNhCw8Q+ynT6tPXqRHslwunePfgpYX+7W1Hee++9bCFy0u72GeOpDHEbjZMMzcm3bt+e1dD3sUrx2Fwn927ZUTliP1Jwdvx1C4p/+vao44zPopzBYgmljdtXVqO0VNgWjubc+w/upx6mujdefx3fLDinRPW5ZvFC98Tj93dffvyeKE7S78k7jm+t5L7TeBv0Dp8ls5S2AlIx+cljBVkTKFdRtB/T4gTFyQKKE53DDrbqaHHCVh2tOVaz9eY8ipM1KI+1HjnHVh0XXzYx6XB19RICvn1DJUYUJ0wOXG09fZpTdVBA5VQdFB6asesbxrgqTmxT0tQV3VKceOLfiMVJTtVhCw757+wtTlwtHmzVQUHlarSnH2WrjhYnKE4sdi7v6YCj5R8+t/5XnaJPAjGd+CYT+FoIuHT8I7jaMnkBTP5jM89l9oFvO+eZsOJ/NXokyM/8DeDznHzSYXhGXsnk0DjIMiuYjN6+f7r706/fVxYnxHP0dMHENm392yZyAU9HqXwhS4hpY/AVLKrWjVe5S3+D/Gt48VhX/6H4j63FJ+99/P9g+zMv6RPUyCv/Ioc4zgMt8g93IshfL6A4+SXOYf9scBwxVr5iylioNIFbWl/GCyM9C4Ax8yj+PgV+/80vjlc6EzWLkncc0/LCB2uX/FU6mTY0ZeKe+iIlfbIYHXHIwpPYtBxwG4h12bbfqFB0Au0JLuY4jVJVmVzlomNKtqcob5DWGqoWQx6hD+/cR2lmxcd6iMxST335E0fUySd+zyJzGEfFjEWUVkoLlkYq1tc822T8Z6Ny2xRxJUs1qT4et1LGWe4Ky1adFW6bP9/d+qmd3ddRnNx+M4oTaURr9aAIxx0tOfUJaRlff+tw9y//8lRO13n08S90t96G42PgqijVV5Jbp1QqbNjIKYe071kUucLTYfUc249VfP/jk0923/+7v8vR8X/wh3/Q3YHiBN1M6ByLQviOC5PSXueuHjnsVseNbKPZsHVz+PsV5kZay/k3gzIjChHwtp6cO+lAX2tTeWEU35TI+i1KqYBmcQWLQo9s10pZpb0nw+kUWf6XkygT23SLUd443rglfyfjiYunWrsY1zbgYRHVLq52x/CH9fKrL3RPfOlLKI/ZHoplItUBmSBE6p9bX/+MDISMXNLeyPUzElCP1kyCrwnhg4HtWjZSn5qbTWX5KAQmYNnQgFg+BMD/CfDpv4PuVACXYNMXsSdD/9Y+tnsj0uBeAS243ZcEj+fp2zURl37qI7R47T7IuD1UQAtu90Go7dZ2ksZBqLXnNRaxmNwIdRJBxpCG7wBJoAKP/zJqGwYjNLsjJy90P/n5y90vnn5pZKsOAaaTs2t1Anw70yj81pBkS6PMTtRE0AYegVuUzcr7kqsYU7Fui1nd1Eh+I2eBLFP++Lig81bWKSmM6GpWtsRzK85h7fOfFHzLM15+aV395+OU37R1WKOpVXZRfgqH4Rx0oyAKtQts1bn/1lid7GSrThxWOjlAgHNwSPzUMXlQX43+PZEKOUk7ci1Hf4MdYN2Y8Fd/+yMckmpxwgDuftsMMoSq8UisGmRtK9adK1ACst5WLqzOStpb77wTYXklq9nm6yQ9x+o5yICnA4DCe+3Htfx8o87jRE06MPFxwm5dr8zIxkOBAQ7PhLhK3C9chA7GdqAvutiSiCVtbMepPNP5zQG82pqtqQbZhEjRlGO0/aU9ErwKD7MLMxe6G/ZMd3/0zfu7xz97KLgkb/M3X+/SSbjgrYgQoYSCWFeJZgEC30G3VC0Vr3AVExdE9MCuzPjCb072xxFz4gNbdeIYmCxWk88Htb/0F+FS/ihOGGBVnFh+6yI+TsgHMajwAtv8F1HQjmAgzqnkwt0gP0k38xE+v1Ga6FQ6A5xlMxJh3uqSFggr0oV0RX/ziUqLMCJzLSp8+EjbiOAkvASBhO2GIGGuYiKj2bFCxsT85e7+bNW5q9u5ieVshLvs5Q+yQwzM32sc/qBZGcIfmZuOtp32HilX7FTkEAxetufjWOw50akVO8JTYMtY/E+kXb0DWLJNEUjuNYRvf7X9FeT6bfD5KEALn7xNaX5+83lI/7z6CSyH8H017jA0+fDN1b7UP4LuFVa9PckrW3WSp/QlbzUJpocOls1sfBV25fgx4Isi18cufw+/zyX5tPbvvWgOntAd9khfxeKE/eJLFSfGk1LSXfrL1bA0H15jcEZe2qOKE6lAH6m6M6DlyLGdTAj/7TfvY+XAGPvMC1ienOX0CJSTCrlaQkF3j+pVWL35xls8CRTB9Birekcj3B664cbujtv20oxPR9D9+dNv5mSeSfruFMK2phmnT+E7g9Vdp0B1/COjhiuLIsS1gYn/NlYqFVo9acfJu5MYhWydtcpblBM0zfbo0I0oA1wd9tjhcxy1vQqFQlYT3dYFr9CqQr8l5mtcncRePH8Wh7Sc2gCM1awW2mZyWgjtRTNyy7cBk32Pu76IxcRRtv54yoXHhV88/U73ra8+wt9nwJPthwwIHjOei/TSttovtPXdq9E/L9dvf/pAmiPNESx9nh05jvjpZ/FxgnPYK4yrl7Ge8RScKco03KrDqTocR2y33YQjdhUnl1GUa+K+homHlimurKs40YJP54yawlt+V2+d9EzBn8/p4wSUnThkq04sTlCcgL8KKU+PcFHBE608FXC4Vcf88ClG3WjtouJkGvMcjyz2qFC36jgUMl+l7uCcuVd/lDResSQDUEimnJD+30KNQTtVKUEEW3HjP5JbojtWxOqSDyVHET9xkzT5AjYphR+1MtkXhKH8EfjkJXzhmJeRXOiBhQZ+2iPWgof2qTi5v3uYrTrGlf6rJlCgK8z1/K/4GgBTnsI221zF14zzZ2nk0wLqP3lfcg35TzUzu0xF6xOJvCCSj3j3mQimf/TBfBr/KRDygJ7/+8GkytM8lJVPrzgh38ippFdeGW7V+T6DOj4pdMhvuiwigIrCgJefgC8KgTQCXyh+dTytestryGWY7aVqgjbhYg0ZSeuVMJ9SZcA76OeqraRn4qT9kGvgl4yAXTK4U+tM7BHFUZ7CPSmPJ67ocPTyRRQMvHvEum3Rb7YBJ81+d+Kus9ZsWQQvyex4alsa4gRuoT/xQociurRuVcETaRnnPcmG58gw+abioo8HkeO7KvKpH8mUW5rQvDSlzKSeIzzxeLNibP8lP1hvUlo5rbbqHGKrzte/qI+TfbFGXdQ6RCEN5KIUgseL58uvvtp993s/6N5iAfGJb3yju+eeB8y9O3b8OLRgu+XUOpSiHOELH7UfzKNwtoa0xtDq2K0yP3rqqe4Hf/f9bu/e/d23/vBb3aFbbo1liVtzdOpqP1Ih5ilGbgGUX2kZp0WI/rO0MAGV0F1Z2zoQlkese7qXhNikggte7ZacLKRVwwwvl+r6rHFbkWn1keK2IpU20lhFkdt6VIREacZ4pDPw1D98X+s5dwI4dlC0WNopi3vSlD6yLpw7g3+uw93jjz0KzuuCQ+ubNgzzSf8nserDFKaagqTMZfmGn/q39rHd+7jDWwW04HYfhI99GHkZeWxxxz/1b+1ju7fIg3sFtOB2XxLcv46Ejjy2uOOf+jduzJ3sTlwtRru3lMvcK8r1I14bsvTL8H3w1B7afRm47VNFuX7Ea0PGv2TA6JuD3VbGkPLLTIZcPjEq3GDz8JKJ8DXMQqZrIj87CCks0shhaMdOnctWnZ/hHHbN9HZYA9syKgqTtt7iJAnZ2wjzSPaj8PkkAwrTM/tEqGgmC85EcPJYDrKSWX6Cs3l65eZP3/zzOHy39ht8mVgmZkbt08VDvuaBdN4trK45GfT6JODbmQf0B35N0u3UH6/88oW2p1T6KRSFDlqcGIiPky7OYTlVp23VQWi2fajRlUE74lgPIYg41dOw/A4CRaxB4HL095t1GcXJd8s5rD5O5uMcFhjCZMD1aDsHWLO0aansoZJFoAZDlCxu1Tl82K06OI9iIDFPm5/pFMKTBen1vl2Te60MGAScpFHHOfKONppETDhyekzSU1zB94WU7k56soWDu2XQaqcGO+hEm1DoQA6ljgwgUuL1ebQXc0mmZC63SdssIFo9RJAQX54XWaE/sBcfJ1icPPbZW1P3fcykyzHewYRXoUIPCWW20sv2IxoKCoV9X7AMDNKZUOjpSp90cnR88TVO1XGrzkm26mABxOiaQc7J3gf1P+HY/1y30seJk6Zs1QGzbGWgr6ReLKH9TJT8pfy+5R2k7X9NmZmqrighmfDFYh6fSFoeBeWktRzkQYVU/zeeedkeK88QQsLkgRv3HC8tPNqJeesPQuUSCJKcPx5nua2eVOiB2vg4WY25t4qTL+Icdtcmjxi2xYkkcKR9y38ACrxSxwHZ/xjYX8LySnmkgznU6pR0cEvLCVazPTJU/z/JNkms4CqNaUJHOzWwEpyfym0woUn5i2opnO8DVIhrNg2AnvtChwLi96qpQQmXwAd34NsfBgGkR9aj/rU4wlyXiZzC2hosApIbNEt92f/I3fbfaNXab+FTbfgjwS8sQovk4XtQ8yeF5J178K33AXyj9P2nRR0tv308323TlHECWrmqN4UCYE+cw5ofF2G2KwXlAl31kRXUimEk/ir+4Kk9eIfxWm7/t+qqXIteV+m4T//6NbboXOl+xsl1R05rxcA2HXhh2q+TYjrvLrbRbN26m21fJ/GBgk8pJvT6PtnGXu/10/ifufgeW2c2dsdPs7Vm/bZuAgeCs5TpApP7qygivLS+mkNgrv3lYEH5tjNh12JCIfnE0ROlJGDiP+vKH9+0jMgYAgE3s39+E1tUVJqcAg8n/bYNLURUzMzMzGOejWxAmTbiMFXHqJbeVUUtK7QkVMc24wTJ+uFvNeboCuTiMEE/PoWAfPrUaXCfwTqMVdU1izi+fbf71hOf6b7xlYcxRYenQRjN2CVqtY1G/+p/qZLR+k9Ma8pw71ZMXY4J8k6tap/lVJ05nMPq46QpTi4hyF9lXM0pOExAzqMAWo1CxVOKzp65BCa9xQkCvr5FJjlRR2eMTiJUfOU44tM6PsSiBt8tcXaojxPoqj8UV2dtVTvYgnMOk/3zbPHRxwlVE8WJFl7S1K06jpU7WW3V6sUV4FicMDHVOawKlikmWFexSrlMfW/aurH4dQosnRyz6Z+WPuSqtiwVfJIwss6yCPWrNOrjFpHbJ6IWFYk8oL80Het/DGUA5CJ3ANqHavwxZl3tXvCJ048/KbwphcufNydbK2dxDovFyX/5+oOlOLF/Miit8thbnht/CR930Aw/tSEYUnCLw4uYvDiQ61c44DnAri9i8hzlicSwTMNr+N7gJ8qHbH9JbWVbSC6pFMWJ+PndMkjIjAucqoNV2XMvHe3+fKnFiSM38kN4k5kuA7/on1oVVND0PqA/5Zee0qVgushBqVpxlU+UPfhzEhxasVhgy4oVGIqV1JnKE+CbzHrJoheZqCS2TB7FbH2qoBVVFabyJLfv2DadmJvOMUdGsdJByMvAXHWXPFVj1UZK/jce2PflN4n4RyYFT5cy/PVStjAH8U88viykPNYkYeTjWG6viVN8IhmeZAC3LNWChvmtQLbBjg891rnu1oM7u6996cH4ONF5fNGZFgjt5DnOCRyLXnv9te7vn/zn7o3X3uoe+cLj3e233x0YnjKzHmWrvuZipaNjV/OBT6o8jqKJdxXQv/nNq5yq84/dvv17u68+8dVuN1seZ9kGqkWcdHLcvopVpXzYQzImUFDrsFveLaPw2OHpybVYFaqokJ/UGODJOCo9tCjUIaz8XpoJuxY4QYXyRLmBIkwF8UZwntSq0G4G7VWiWwcqzlSkXQa+yhvrWEuadShujOAYnDZF3xWOvFLl8TyWwnNYybz//rvdlx5/NAoZ5wPmWf21aqEsbqlV2x9h1l/u3K53VZTrR7w2ZOmX4fvgqT20+/WA872iXD/itSFLvwzfB0/tod1/G3wqwqgjV0vZPvfUrF7SsE78FrMSt7d2H8ly5HEQmofB2zIxDPP6JODTdNJAe/itwQSazWrkg88OAhlQiMCj56zb7LJXVBZhdBq3PdeJrlYnR7E4+V9YnPwM57BrcA473zuHlYG4B1JNqvnmOOIGjrwVoq+BH6DiIXzvFaOKwLsTHqrRr9UdjMifVevj4DK0WF9q3XIZwThmQWd2wtfg2zRqqw7mgTAeT9XxJJJPCr44tPILu1Qm/cRs0P4oowMT/35b+WU4tVUlQ2tIJ6PQQ3m4MCvpOY6YU3X0cbJrKxYnpImywUEJejixjsAKMOu5JrjAJyjwJZ4vRdAiZLArWjfSE4P4DITc/+Z7T3XPv4TFiUdUu/phG3LwRQsO64MMTlodkLilPTHQOsGy8aDJ1/nVO5gqesLAShispHGF02OW3eNo23b81M+GUNvETsXJgpK1dxESdQZ25SULoxBYkziom/LwPZfMmfIQz2SRA3ioLsF3ceMvd8MllJmbKdewRdd7tiuIAHRz4BZLM54wQ32c7OFUnW9+Gh8ntwIKXCKdEid4QwvSWf9JK8MnB2FYt3zkvfJ0dSeFCgSgpACkJR+juXKqQvCF1zlVR8UJzmFXMCguIuxHuCSnKgeRTWuRchd+n553hbYzmJZrVumKQeDTxrI/mTfx85uDvnimvOaQ//yAh22OoLTX5C6CAkn+JXgokJWgU3lIzdYeTasSxCQFz/KSU33gGxG4bIHBgvpK+/KNSLE+Mgq09cSM1fiucELAvgd2O13o7r/zxu5Ln/90t3Mjg7tbDcEr8EmfNiO+hTC/Ag3yDSzvXtCKf0GJxPLR1BkfXHVrOah4Psl6EgngAABAAElEQVSK/2XM6bMaZ0AGdiOaDXkAr8pPOaFL2qth/WULiII8BGg5m1za2Xb4FkRIQJwaBq0jBVEDWmb9c+q9/8wnvxZ8HqxQ87Ou7Hz0xawSqThB2FGAc8KXLERX+OZAOZKTmXGlS/Rt2/zqs7+Vdz6MJLkG/kiEpOrjevOqnFr5/TCSmY/8Nf6bFxORStrkqHKblII8EZEpowyYxPfOvr34OEnm/hSvqElnQbYkTv69eoj9w+CtAlsMmrsh0su6SR+GOMFtgX3eCKU/ePJnKDpnu397+e3u9MVZeLUWGewhT90tsNVlG34A1neXLsxx2s7pnEZjdlbPFH3z8vmj3emTL3cHDuzvduy9tVu3eWd3HOXnJUy6L1+i3aEAsaxWqwJpOAoZrMex7LadnOBCmKubMZum8HEkSCyF40rD6TBYTmzFKkWh+gyKAB07upqp0795BN5ZnZMzrmrOrXNDBW3pp6XEWfbpm06eoi8DTcTValuGTWwl2rp5a9rVBSYKp8+dzpggL1uBML914xRbTs92X370Hk7CurvbsR0FO33ZcUbcMskaNB7pK2H447Juq+X1Hwy0nRSzT6XIh+ZR2L6PcvNZt+p0bNV5x+OIX2XyxbHIKE60OvG0B08U0vfLar5LD0/VEZirtlcpf44jNp6KE5Qbbu3ZQPnOcISnSqjaqoOPExQb+pJxVfa8/l5AaRtbdc7hg+YiCpfNG5FNaDAeTeyqcE7VUXHCWOlWHZUtKkvWrYfGEOHKZZRrKIe1THGLlPnrs8A2piRnMyouG27JC40/dAKwj5KEf+F90Fb+wyeiVPtPLze+FAUv43vZl2K1MkZ/O5aB/MnjC1A+JAsDwjy8G6fBN3c/8U+FJm/infz5Xlt1PFVnLYqTB7qHH7gdGhnLP+WMIJFMwi9NHxAFoNqCwcQTrqOxdAi/tfyGATCJjFDY1NeW2nTtqm/yCvthEcX0PCd/H0fLbzrDCRxpf4FPerNIWuhvra1QCcFXlUnm57jsdu+LOId97qUjKE60ONmIEmJtxn6EP+rDhYOe/sluGfiVWcGyHpaWX5j069DTPPvsUizGLbe5tL43j/PXyKTIesHVPhn4yjHUAmVVdrPPy0PrVB2tSPSzhOIER8jW/4Z1yBn0d/0jaWXkeBMRIjIvEczkmqvRn7vB0hWY1dhTy2nP1e5pI9BP7IuY5JlHxmkfaUfZ+u23WM9DawKEYHmLCET0Eg55SUbTVf0Lt6c16bGbYOskipMbd3Rf++JD3Z037yUYWYX4FC6rOZ6ApBzg9pYjR4/hBuEZjiN+vTvEFpu9e/bD/9muiUxsn84JN6S1XyhvqoiyP8zMXOousb1PermN8l+xOvFo+M898jn4L0fYwy+yaAraWnLo+0gnrB4xvI7xxK2RYq0VnCScWsMWHKrOrYSeAqgVinLeWvBQweL2G2Hb71V8pcQUSR8l8jGIkNPa1oKz/k9UojjvsP7LiqisUi5dwHKVvCyXyl/nMLMqYiiTSjXlDsupElhFj+1ijuc333q9++Ljj+FAd3vGKPlDmpvVkXqxAm1Stj8vkKuKztvgOfxnJHj8ceStpe+TL7kNQvMweBuJ1b559xI/nn/P4DNfo0ZbBwqiy/+04oyHtq/tviTUrK2pai4hQWKM0KHqafn0o7ktH6N9bffRFEC9DnwnKZksiJoCmYOVnVJkgm5fWa3SKqTH36bf4sn8ZBZG8JtMsNjhkVMcR/xTnPM88xKaxO0cR4zTSedxdIgFT9SBwaYRy2yWwM9AW4TpIQezHr7AvIJ8fwd2GpbDGM9cvkYBABC/qdBJNfsWXGXMRBwpfzofnbPBN5p7+xT+a6sOwsmAIf/u4aeUwVtcZcB9nVgeCyXuFMI2Vcw93L9CKbDfDRH/YflJBdNW/275VFQ5UCo4rGBVbAVnx3/2oUMc1Xd7TiOw9hSKZUya9Do4FbyiOAn7a0n5Dejhi+wQfuHkSSHiJR7f+dunuhde5TjiRVYaVzBRJ8w0hgZ+BICqMystE1hCYt3E5Pgy5npvv/U27QoBWesIaBUHcjB3BzoHi1AGeGatcKcs4Uq3eMUcNOUijO0xPJKm8CR58KQVVDlzM731IV1FlLsRC0rGRIUWS5A/YZBnWAvPxm6rGUUYs4BVi4/6Q+4ML+SDth/P6vt3rev++Fv3dZ9/WMUJkOyv1JcmqCo06hIbLgUTCmW50qeNL87gmD2xCjn88yg9BaqEgbdWDWSbPF584zjOYX/VvX/8Qk7VWWTl2rYyEBoEY5b8DS9g2qH5atxyDruaCYGKE77Sj3Qc5uWApjJVPNOWvEOftFfLRj4LccBD2aSteDWA3KV5Iy/BQSbm3gOsEqnyD5bCI5rlJQE55j1HbeNNXniaAjMSVxVmYLevmTk9g3RaFNhvFhd0Dnuhu/eeGzlV515W610Jke8Z17qwL9XdT8KyMoJzFadiGiY+pDBKWVJZduDY/hZVVtWzE6ATKE6uRHEiXrbGPq3pcwmHP8KSIWmEHvh+4fV6/M90y/J/iR4E/THvdh++8bG/DLPcfZwePkBpd9Cnr3+FrtUINqvZZmD9K0QFfvofMKRRz/+hdrJrdWvO7WrYDDFZBn5f16Zp5a/ytPo3DX+D8gsfmgFfQbfGhT5xym9OXGRmmJS2fPykXS0i5E2yCrcbHyIpF1lX/Sc76GC8QOTHdtJfgmvfU7D8tNDwhKBpgnQE6gtC2QpUKl9FKPy//5/vda+/fapjfYKJEYLoDPljUbEBAXQDZslbtm/KRPwY23icFE+sRbEC6nNMoibgj1cuHu/OnHq1u/WWW7qbbr2/m8Gp6ctvvoZlh8prLFfouwrHXgqw8tktHPm4fcf2LCYcP3aclT32pTtOQBMnOG45cVVP4bu28myP2fVxhPzTKA88CtwtOG79cUHCLT1rWZH0CN49CP5XUdqcPn0SIbztlbctFQ2lxwJ9d4otdNu27kTI3xAB2WO7Vcboh0WLE13Q79uzpduKq6UHcQ776Ttu6LZzFLNHs1fzth7A2bZi5Zj/EvkjlQOlq+a8W+N9PXK3BZRz2LP4OHmJLPrjiJ97lWaE4oTV0cs4lVZB5FalC1icrEKptQnrkbNnsTiB/lrXzKg4QbaYnOYEJLYRqVjSWWIdR6ziZD6KaBVItVXHyciaWI+IjcoxnfdebMcR8+0SE0tN3aexONFRrCvF7u8/ix+BGZxNu2KsQvsKiqhJFSds/ZHul/Ag6hHJRefGR9IAU/Zh+duTYQ5e3JQl8uBvejF0oLXat6Wz/Y1YWfxIx+QtHYO7xKTNKH8aK/IhcYqPmcq4ffrk0qclJPDtm3kGkXQq8vCb4xur5Svx4Xb7fn2GYXFy7x3px5ZffwelbBAmSVVMWwbxMp/k6h2ZiWdhiJX/6qqwpGlxLavP/if4I/NfE5rtGHwhJ8sB5PpQ8BNVmkMjJ+xMj60OcuKfFUD/vITFyTMvH+n+63/zOOItdGgWrEgeJTnjbixzR+T/Jv8Oy7YM/GBj5aWgkRmzwEM/S9nzuXCSKnPQxsXHUEihB5onfuYE5M/hEYW1vETFOzGpmFLE8q6cR73Z1gVQzmFpAfQRq8xTeEIpxvbYe7gASj8DKn/wEflviMV32qOJTDfs/9Q/NJQuo+W3lXhZkkTPr0/+Ibd4k+D9lfo3jMxb+1fuUIbN+EeYi685ACFprSLkXpQk8zNnu1tv2t19HcXJ7dy1jl2FFTakSD9RZpnjSHWJ5/Hhr7z6Svf8i69mK97WrbvgIxzhi1LB4sUynrtKYk8E1YpkFsWpioVzKGV11qoi4idP/ajbtXtH9+BDn43FiM5cdTJueh1+a7XmVhuPop+OIgT8I7M5LtRzLFPg56dQjltotxquBQ9JJ69L+3K85936ik8lrfKYW3mCm9sWPRFI2hlfujneJT0vypazjHlb2EqojyfbReR9+Q/tthcpKLd5s50Hnurxx0fwcfLvv/5V99jnH+nuffDTWCmyBdTIZJw5hwhBLOVv3Rd80EVMa3XJ1b62+3hwZLs0MntjtZ/EEKxNxJckXT79aG7Lx2hf2300hUWzbEL53cOHp5v9CFEGL4OHcWyWeWsxzUg8230QtUUYfBh/GAsevAwexiMv89ZiNrjtPojaIvQfQlCe/ew1FsxLyjDytUhPxAQ4MBqJShFQOCCvNngHPzspTeL4qQucqMOpOmzVmcI57OwKGKMMPeOTjEbHh17FmgaNiC8fDL9SCVqGYoPMoMG7DC+9k2fz4Jc/A+p7+km+D0KMFDwSE4Zgx2rwhaHgoXCTrTqcOCDzThrCftfwe4IEa2k78L3R4+hkV9pbHMsd+DCRlN8xrO+NThgbbjJsT2JI1wF1JzSrM4hAfyaEHn/70IO3dZ97CMWJzmGZvBMQOixqp40o6sRXpKxbIber3sbpLxMarRvjO2BEaUEmiuKeqvPiK8dwsIfiZBXbdWR+MEotTmJ5wgArvjW5JAclftIq8KqBuMxq2ZuH34kZ4AoHVP6RlIt8jKLCh/RZSQmS/KTyzcuIfjSGTyPtr//Y6r/F66nOK+ksHJe/xYLBi28eM1dtEfjEUw6pQdW4vAB/gUHDvqen8jWsoJhJTA0p3hrOXl1YYKXywsnuhn2bUJw81D3y8G20AfNWs+7RawxG9DPxMk8p0mDGlwv5NRRlmBl8MmHoJwppOEbyj0GMrBRUn+dUnX/UOazHEa/Bxwk0daAtxUmiiioPI+VPFlVm6X+WQdN9sRvw4m5kV6lrgHUlwE5vO+C3SC/4yrO/53OA9GAS2kcmjoq3Ijgltz1LFz/4TMYaElX/N8Pqw1IqA6RAUwXzDMBO9sDJhsKfE9EFJnTS01UH24PtfRbz0wnirUJxshI/QPffe4gB+H4mpgyxJFXpJYXNOGbS4sFrtX/7n7QBaNr1SJykqUFUuvXclPjQ3DIBW8XJcSaElzGbNX9X5UMaydG3P2GnOgJf2irEkC99RTpU/yc+l2EDHmklgHmySihZ5otf66p829f6PoQ/TFTxRvIGjkxewdfjIRW+NK+1ra/RR4WZSBIgSePkyVuD1Betx719/WjwR8tv/Vfu9kXbDW99+UXDGjR34ebd/hB6G9bDJ9B0CnSugsWijVpz8m/qaZS3m7fiZ4OJmFdoYn5OxhNDvuYH+8DyVw+pAn2BRYlnCXekT516t09h1A0f+T//7K/jl2j9toPdHBOhmVmcsTL53rF9G4qFbWyLOc6fgqyOAamTVW7TgfejuNiAZQeH36LYPNkduvW27sSZxe61w0fxN0WPwkJCEtm/5J+uEqpA2YTSZN+BA3k+/NZbTNjPxkLEo33nHS8oq/U+N8/kHKuUAzccSHnePvw2/JqVU9qDArFxbY+2cx0ZrsfnxoGDByPgH8FHidt5FMS1rpBsKg180EJ1YmIBpQin+EytZyvb+ZRvDnqsZnXTsWMWvyLd4qXu7jsOdrfdvLvTX8DBvRu6jVPWDT5UqL8SJBXSbX/V4pfSv9r18KuxfGuVa7tScaJz2KdRnLDZqDv8jscRY3FCW3d779UcR8zWGnipx2x6NLSKkzPn6lSdzZyqc4XVWU/VWY0yRCWLThI9OcItSKdjcaLiZOjjREeIxtN6RHx2YXFyBsWJW3y2bPQ4Yk5cYvuTZu6DrTrQeBdHS5/jeE4VJ+s2YOFJ3XpCko58p8nTenBLj1uqpHmNHSmwP6FSlZ+XPNSt+pDtVP7DN/u/bdUUtJ2W1smp/KjkGUPH+U/1U/OsELNp/TS9yL4g4HRgc+UZOMoMc471ZG67nmdSqTLK/rOSyV44OjLOHWzV+fZXcQ573+3wJI+lnsuEbYbJ1QTyXpRowV3AZJ0bD5aDq8rpqwF+r8tQcR4t//Xkn8RJ/NG09TyEVBGWwi84hI3CF08+KWOUfAQVsUbLtlPKb+suOqJMgyTPvvw+ipPvMb5vYnwvqwHlA+ssPK9QWbb8y8EX56JLX37quNV/qGKiXMQKbbmRyMeE57fyqPqvBMV74BHUqWO0/MeE8XECD9aaQBlKR8s6As1WDhLJf0o0Im54rbn2mAc+NcW94A9CRMYqT8z2Ypy0t3zt26TPVTGJTXErd+PSFqv98xFFhxka1T8FU+dGtlXbo20NHXNoZR4epb6SRZqV+FKZuXQGxcne7htPPNwduhGrPgQ09N2kUg7gQkaYwfH3Ktq6hyIcZvHw6ef+vVvLtpxPHbwRmmxBaX05eUsPy+E4q5x0FXnmPDLaGazz9uzejeXIVPf6G290P/zhD7Gk34ofEKwy2NqpVZr89xQWxFFuo4zfCF9QaTInj+0df2vVpvJRy5HTbJW8Cv9Q4bJ5k1t5kCGBrcWc9bcK+WUCq0JpdOrs6Sh7Vejo+kCn33MMDMK0vKbRylDe5JYbt5l6gtomTjTdzJ/KMBdjlClV3Nh2VbzPMk+7wPZOrWjWwdNUyv/4xz/qfvKTH3VfePzx7k//t2/Hyi/1an1Bznqums5igTQGxwQOH/z6gVdLYvu23O0+SNQiDD6MP4wFD14GD+ORl3lrMRvcdh9EbREGH8YfxoIHL4OH8cj9GzKNYIZXRR9PNPbtGqyGaZd/GuaVJ35s0FbcctcYrD7C2LffAfx+aOsbzwgW1Lrk4BcEZXiE0dqLudrYDCe1rSONjwgUxn9w7TyXc1hP1UFx8gyKk59zHPHaXXT9yTC2DMqatjGFNlUsXRoxRuHLZATfwzf/MIOU3zDwTAZE8uI5zKsek7YH0H8ZFGdA/4I/CO7hpZR8hNUBawZhzA6s0LyNIw/DyKHNJwHfAoX+1oHKBEhq+eNMUfoSbt1J/1hAhG4OX3lImtCMd2vR/7MIsi7ma2qXfY4I3ZrTaRo4xZF8k6tnuntZmXuIieGubWhkqSnEZf5kZGbNXVyo7+AnlWwb4CVUqyOX4Hgu+HUfRkiLIlc133WqzouvYHESHycoT4BmybIC4Wo179JfvF0N1MDPHFYxCXA7hhPKN97BOeyUPk7ETxqZi/GNw4SUbwrrhaCY9riJY3CvVhv8K5jEFb2wLRo2QcjUw/ZHmLglUyfXPts/FEX4Slth+CwrEwBoijqBptxEDgJq3RUiPR1CLBZQUM0jaK+ZmmMAOtkd3LeFrTqf6x7BEkgHrVqcrGS1Nau7PKudF75Qg7/0MZ4zk6JChUVJJCZO2mizJFIRE4dypI+gAa4voDh5sj9VZ5HjiMu/TdXCB/U/aZL/IOEReAoy65kciJqKRuvEuoiSgm9pG6F/0c544/S3PJQjMcf7v23AiZzhCiBkXCUN7cmIq9qfFDUtdAWAMMOvvFM7CiKLmAvbj7H0pD1SqzgYdrVeelTd2P4xvkGgWeRUhg4fJw8+cGv3xUdRnKxVEAFG2lAy5YcLWAW/yiZcv40WcGn55WkqqBKVOu3nABRtEYHhAu0c65iUQYiWiDIZX1j9X+jXf3CCXrEsK9d14VesJcGJb35Ff0MFQp79N+FL/wA3lGBfcyWq/ZBvELLVv3vSncgNTtUxD/Isq7cx8gRcg5WXjwF/tPzVR+SYRQ/v1pE/9uUI0hUpRRBclG6pE1sAUS0jPIflwPQixy+FetuSY4JbVDbiG2QNk3NTJHvvAkp+1i9YRTlX8AVp0HKX6ZEkw0PyzCCQWoUZC0FrqMszK7r/+hff65762fPdxVmsPKY2d+u3bM8+cZWX8wi3x48cT12I0AqEyKlpnERePItLq5nu5htv6O5DubB65cUoNZ5/9RjbfvCpA2JXWAmcpL6mUaDGMSN8f+fO3Uzm2VqCkHyC7TmemCBWHrVrPWqRohm9jkgVcA+iCDHOO++8TR/zSM81CNgzCLX47ECodVLrXnT3xG9FyTO9drJ79913QRTaQpjLKAJcYZwgXVvVnMYJ7+7tW1DW4MyUycMx8D1z+gL54wwW3rqG041WLnKSwvG34Zt3dp++62B3M4qT3Vunu3VT8loc1SKcSw+3dNqW/Rcat4oA+ID/OMYZSD2Oyz/UPGFW0VFP1fllf6rOe2e6Z5B35nUOq9NXTtVxUqBzRhUdbtXJqTocRywvU3Eywzh8Cd8lKhXjHBaaSctNTITO4LtEGmxEET2Dn4LLHAOqdY7xzmNVYvuK4oTV2Jyqw9YlK1ynjHGciw+AHAfKJGOnxxGDg86mdaio8mqGCdIkk5Mp8sxWHRUnnKpTsoYEkTbWR/V/ibH8+NcTzzZN5zWN/zMOktZv9rNQGpop00jTon6RuNVB8euqE9tBuqbw+Vfw+bVv2v+SB+8MbyoU9XchA48FAtGVdTL+zJyJxcmffvMhTtXh5BFoNEdbte2qEJRPeWXBTCWK+ArvA+o/YQ1+Qz7lF0uS8pfPeWhf+jBeW/8fxBMBrw/Z/gIfCKZv3M3Jezl55TskCp/ulbdapLlVx1N1PI64Y4yvmlJOcrQsGiyFH14s/VMRjrdQJh+hU1/+Kqi489+y1WN9/pjlX2Bg1jKtFMXlFyMTazQOTWmoRVsUJ/AUL3mydSt8J+XKf1URDaOPRv/I2GYhDV3tpd1JJ7N1saXqEJhm7yU9iKM/E6WPWJhQEVkIJFGNDdIay2T4UMqGDCIPmIVXbGEM6Raudgf37+ie+MKnUaDsRN4lqjIcC2orUPKtxGpNH1Qr9dWEbKvi5Je/ep6T0HZ2+w8cBJZw3SINXNqSW2Qdi7ONBt5i3anI1Zm2jqF/9ctfdn/9N38Tn1Xf/va3OZnn9ihCZrAEOX36HHwZ60Us1LSG05pDS/RYkKA0sUD6KDmHslcLQxUdG4iv30GrXStAxwS3BE3CY0x3AiWIJ72pKJYP6bNJGZ1c4y/FZ51fWw7nXBfhZfKfDfjhUoZwXHPLFh+RjYrPadGnsug8yptZ8LCdbEHRI+x//uE/oTh5qvvyl7/U/emf/nFZKYEbIp8tmT96j42WixEid9/S//lt19g3cBP+h78qtfHzxE/r/8vlUbGHaQbpKjWJCfs9gE8diUm7RhAePA4eEmn0rZUh30YDWnaD+wcGDmL1pK33QZLBQ76Pvn1c+Clx32BSHz0GNhib0yig9s2vMhMFRivPBu3KRawR6EVxugjzkr244n7sRK84+QVbddZuZzpeK0i1mkEnxPN9wemZNi8NVn3PB376Euc2jCE+gE8bEmc7m5050bkN7mlkGa0FQDzR58HVi9F4eVbgh0n236WGTMQ91wqIW7aV4kTYnwT8QeksFygrcMiEUyfiTflqy4j0d9InIqaCEaf8REixCLSMXPpRwEgvE0XLPclkfQrtL7Y0TNLRKl883T2K08vPfQbHl1icyMZMGQHHvAAakcksm8DCd+HDC69Pf5LWZW4O705ke8XJdzlV51WcFWpxspI/J69pOQyCaZwOVKSyPm1n/JXwVUoc92e/+fa7MGD2RcIoHSRl6owZ1BPvIGa3LrO8gk9G9Z98r1f/A/qLuGU3aXvo39vn5MpLKXtsThDDkU66U84INBKJTOwz80jars668pgJOytfmpf6btJZtkytWYkDvyunuxv2bu7+gBMhPv+ZuziVgXIxCM1eZWWSQUJrkHJk5cqFgm1hqDmolZY92OkIwlWqFGMv2zXx02lIo0Iq9TdUnBxhq84iR5qCaPo2Cn2uvsS5DSlkkBYSwjfLs/o4WcP+fMwhvewvqQefrcf/IP0V8GL+DT6SORY2AE7WrVy+pN0XUSKI0x6yIieifkYRqiIH0gdvV/A9mclVHLcWOUnRtHSKdrVmgtXJy6fY3nCuu/fTN3dffOwhJnq0OduW5YZ+3m3b1/Afi5yLCNT/WPlDC/t1CSJGUxkWfkqG5nuSSZkr116Z6FFmi1Y9qYfPt6p/4RMoT0v5TZSkFWEp/BY3mffxuA1rlxfTk2X7Fr7Je//ZoNCvtb80M4VMcLfeW/1fzFadcgZqohK+wbdnswM8f0fwR8sfEtAmpHPqyPLw0f6f7TnVQIBcl2UVQ3lQ6Gl8vqk4XmGnNTEFtq61KHKs0I+FfiMm1/RlClCyCUH6jBn9rN/ffpkYiDUXkPWSjrFSWBKWPqX14Azbbf7HX/1T92vM798/ealbNb2127Z9d0460LHrKRylXj6PYMkk236/Amu2SRQ75y/iwJn2/hnMlu+7+0B35uT7KduPfvpK99b7J5hPYY2A8OqWGo9atoW6fWPnrt2ZjLyDcuPYsWOZdMZPCf1PyzKFZPnvlm1bmKTvylhz5P336/QXihSfItKQ8riPXRqrLNizZw84TnbHjh4l36Pk65YgnH1jzZNtfvYF6K4wfmDfPqwv1qK4OUF8jl9mG9uqiUlMyLGSxBHu1UvnKd+xbt/OTd2B/dvY7rixO4SfgFs+tYvtdZilW3UOatSd42UasD+tvoryvPLNq6rC2sg3v5b8Y3tw7MXHybE6VWd+cQMWJ/g4+SWn6rDl6TKKE48jjnNYeMv58yiVdQ7LZO+cihPQ2MQJdp5sc1nFCRMJ6+oiW3Usf3yhsDpcxxGvzyrvZWSQnJJBPFfdxWfbdnycnDuLsuQSEy9WY2knF/EF4FadoY+Tq9TJNnBgbEEpFsUJ9eYJGR6BrPNGJz6eqrOZulZ+C3H68g8I0b+3z8KX7y/t//Y//vd0lJr2F2WKomzkCrFv/YM8zNM/f5uVaRgJkVVyV1A6Q+VCQzJfxwQVJSTCSTMndzAZzOSO9sMDdOOu34gDk90ff+2h7uGH7k77m6MCXKWO6J8ygF36tBhYMi4REk0eqhR8orB5Sxr5XCu/ZeSjpDPd2GV+RKzMUuzAGCl/i95g5f23wTdXs5XnJgELTjBV8RgoToDhlpCLWJzFOex/7xUnWJzU9mfoC+2a4mQ5+ANCCIT+KJ0skfX+SZXfRQ3HDyfSzWeSllguMqg4sQNqlWV5Ew5GIb005b/Kx9Rt6PLx6G/6TKolpiVWGE354R20u1LqeBek9U+8NG7ao0hk0cp2X2EuRIYB0mlKbjdf+WHELeTwc9mmffstN3Rff+IzHEe8O1atSIuUhXZsInhDYFFaZYMX2Kbz8ssvdzffdicWJzcxT9ESzbIDChydm9nP3SqYo9vZbuN2F+XOy1eZn/30p91f/s+/DB/+kz/5k+6Wm29GKYsPFKxElaO3Mt/R74l9LNs2yU/LH68ZtkWq3PXuOLgZy7i1WA9KK63ym/LG+PJ9/TcdP3ki7X8zC9BuJwVNcK1xRkezU8TVibx+VU6dPCUYfDdx0pvKGNppakLlJvnPwzunp9en/h2XFnlfy6rYurWctgMu+md59plfdC+89EL3wAP3do888ghjGWMFZbAfp5MkR7EQa8dnn20vXIPHwcP4Z6MQZFaJMR4tcYc/Hxg4jDYE+nsPHz5j8YfXaBFHn4cx2lOFfnCcFrfRYRg7T8PXQcTRT6PPgwiDhwr94DiDyH2V9GlkCgQN5Gzbih/6m43BdwdAeVGUHYbzImPOv3BnU8i0TECzJoGrMVGcaHGSrTovMODvxr4BIQ6moYbfwdDJfMGX5fVXg9ng89nH9tuQDH/0h9DWeI2Y5/qcVO1H5mX3dGDw2Sj+DcrfR3Sl160srfyxOMEEzJUbmVA5h3VyalEL0O8SvkiF5uIDo4ilR49bbtA82FsHA/iUyLLD6BR66rNU8096MwjB0BeUximfp+zI5FEJIVTgpZqtCI99HsdpD94B80MDDKNWtE0ZE7c0vMIXprl6VQsqWlqD4i3+y13SHPafYO1JvsNWnefZqjM/zzYdTnFpihPNrW1HsYwhr1ickG8pi8hFIGj/r7Bq+vo7WpzgWT2COAKXAxd56xA2K03g3vYzCr9Cxf/69S+CoX9fFovjY6NlHgdvFSrjj/JE1Pgk3WwrjT4SRbwtI00rgooTDYU3VyPVvpOCQrLP/Px7CNtz3R2H9ndPfPHh7t67b+6m0RJkjyyrj2WaaO2o3vKyL+WXu/2v2iYjbF9OFSe2GduG9Cv8fKB5KVsyaqM4GXEOm+OIGcAyiSd4WH7hWebKw7Zgef3zuxYnKh/0jm4iad8UWFoFfZj+Z/bXo781Z5hxIp9YLuNTbNu/f0FPfHj25t1ypv6JbJSZK1pZ4RsHIXsBmpbFBxQCd7dSzTKAr+OUiwVNvhcw/ec4y7XTi92D993WPfzwXd3m9UzCpH6QMUfp378GaD6N/VT7SwmCi9GCJsgjYgRPB24nVCqZpKmKk0us0gf7WCu0zLkbQeZFbnm0vUkAP9VnHtol3P9Y+zd9teegGhhD/gMqgpYUwJaH2v9a/bt1QAeZq92bTAQFYlG0wpby36Qnn1aXrSgfBX4yFxeuwPGXDJU7ESlDC7/7ZDsK3XhPEgLS/gjVn5Lx6ruxTQQXy1ZGQqinNZgmT2NVsGEDbYLJew/WmCQ0tZdpoR782XufYz0NXxOzYvNrv+RSVrfmatXfOx/BawYl7Hef/EV38twK+ODpboYTdTBpi4XSuVM4eEXwnaKN63dEqw23L7D7G3xXYLWxq7vnzlu7vTunu/MoTjYhxP7DD3/ZvfLmu936zdsZFQBOOhUnmzCl1lpBk20n6zpsdVLjan2sbsBfqwj5nac4qGDx+W1OPHOC44pitraqSFMpEkUIq55837Z9J1Ytk92pM6cigNte3C6iNYATelcqZ4GlgmYnK6obcR579Mh73cnjp6JUUTng5OUqilD9A0xO0oNQdO7bwckwixe6jetWdPffc2f34KcPdTtx6m67dPuUlSTZ/VNuGXRjP3B5KzrzTFyrcVz+oV6IA3mwODnPVp0XgYdzWCxOfoHFyaKKE8zih4qTiUwaPP1NS5Jz+DjRKfdGLE485vMSCvFYnKAUHxxHzKRQZ7qxOHGrDjTW14AWLDpHzHYFMN0BXc6wZUqHsJuxOHEie4mJj6u/Q8UJW3VwDnuW+phBlnEF2S10Wp+obFAZ05zD1nHEtGMJQLlDC4ki0fKWl/6tQuU/oZdJfOaPJkrsSpM+DyHhCvQpv1YvkJ8oi8if20UVkZYfGQrxsQ3hkUlj8rI/ohhJrVVPtkdne6JwCXFLxAS8ZxWWtMJagJcvzp7t7ruTI16//EB31+03osRzSyS4MLlu/Kfql0y4/P1t9V/xEjPxKW61o8rC4LHLMn/g+FekSpoPDV9imy8ElzLlQ6N4TEQBwwnT2hVxqXvmJbbq/L/fxdoLHyfIXMpjLvYoF8iTUxCSpM33+BQEwfAhL0YQqrcKrd+gMqj/ROBTuz50+c2ezPWRJU6xWoQHCKPqbS59REDyG8ds+4j4yI+UWWx/5SzUVHV9VPhV/7ZY2on8UAARJEppogxa7buNKIQDuKhpfGHTGasiyENlAy00DdyWQDhxpP9KzJX17XcRPugi2QP339l9+QsPsMWQeiKLyZgz80ASfdap9NDqxIXDn/zkZ92rr73Wff7Rx7t77r0/W/08OWw1PFNLE5WtWqORMMrZDShO5Kc6aj956kT3zM9/0f3t977LccR7uyeeeAIFyt7wBRcjN2/agnzN6TmUOwpISqByVz6t+4Kz+Lcy/wnk141Yj0yTt2Wcw3dS+jb9K9ur6IPnsDLRMsVuvR5F9zriRwHDODPLFiUIFJzdgn1J6xGULCpe3K65Ye16ykLOpNWCRgWQMrdjr/zz9OmzKEnwqbVhOlshp+CjWpucPn2ie/2N11AIs836vge7/TfsZYxg0dF6oF6soSy+yQSC+ajFST4ljlW//GW64mXXjzNMuTR23uvjMBJPo59Gn8ci5aVCPzjOMNXS2HlfJvHop9HnYU71NFCcXD+SzaCx+55Eg8iDhyX5Xu+70VpYu1fS8bfR7D4J+GARxhuWDzAbQJVS/Cyl+NRliM2N1msIHT4NmQan+VYxXRmv0p6MnMGMOMdOXcS/yav8Pd9v1cHiRMZBVFfRMM4ivvCF9iHhh6sTnXvxcoVJ0vo9AwVhya/vaX4flKQ9t9q0hOPlN6rm14OY5KkDo4toQJ30bkHIdDL4ScEfQAYPaa5gIPfMCikMPE4TnTDCzMqHguVXUQDdNSeEqaX2oH85wyI1e3xXoaiawjxPpdeVS6xkw/Q0fb5h/97u4A07ultuOsi+x80Isq49AAeo0mABJqXJpMvDrYYSCPxr6D9G5yX0JzPLYAuC7XV//b0fdc+/rMXJOlY4Ea7FPXApD+3Iib6141hle1OBkpL1R9t56sOb77wbAVxmapXEISVIqUhxgEntgjchZtL/+RUKpc0kRt4Nr/bPK5exDa1rpP3bniROimca6e+vsYUjuhUh/SL4+l26UiY6ThQtPbPmBQWKR3giHLA3f/f26e7mG7Z39911c3fHbTd0O7BuUOVg2a0XSgcA3syXR+EOypIVDt7FERxsF9VRHch9Jg1o1sqIj1qcKJx0bNU5yeTp2e7I8d7HicIluEm9itn3f/MOcar+F6wraCn9PQXCuvAUCDFTMI0TNMrsYNsnTH62c+EXMb03+ifzAsHXuor+QV7aQm+yzyBcLYq0KXMKR5Kql1QDb1HwJCPzJmwWZRWrv1ocqLySb8WBGsKCgswce2Q9pvsqExr8mXKUKavd9JG7bjtAf9lJGcHH7QdaCvHPeg78lEdAwiHfdJAqTxWofyZeqzNLLg2TBuWIdCzlWodZK/3UrTqELmKtkLJSzigsqBv5VLU/7hbWiIK4Bj6Z5tsIfN6Xp7/pK7bZ1VX0b/CTFYEpA+WsspA3OBnmLQJeX/+e+OEkdxLv+yIpWcLnRMdG3CMulHzii/ePC79SD+lfvbPaX/Knju1HY+W3/QR+tczCo36z8kioystqb/ZFLCaYnHoqSU5OmaTtCGJwgX2KZh7iIsxaqWvlHd4rkeUtiDyg406T9gP92ucVKL6DN+1tdnYF22B/w9aadd1Lr2N1cuISE+Mr+Lq4BH+/Qr/jKEiUyjNXFbSJT1qqgBU3nMZy3O8uFAkb18I98PfjfvKfPftGd4wJ/QQrewqu69gCuZnjbXXy57aOo0c9PcfjxlGAUe74ZUp9V1veQJ6uTqpUOXkCc2z8abjoIL4Kq7ECoK/ZZvUJs5U99W4RMV8tTfTdpLCvQK6SQEWr/GOaSb3OaPVpcgYz75PHsaRRWIc/uWVnka05Hju5bnpNrEx279zYnT9zBLZ3oduORccD99zd3XLjPvoqYyTbe7QyWInyaAXK4sg/Mr+0tOoPaZdpB/WVwIRb443XWElO1lXEHTlxtnv21/g4WdjQvfWup+pgcbJyCosTTrOZ4xQcVk/XwFdUIg2OI8YiyLp2C85VxtfLODN0IpM9/WzBcdIi3U9zwpG08JSyGbZ3XmGrjjSbhCmdJw/70Xacw3q88wWUJU50rBLN2z2ZR2eOlzjqUzP2OlXH44ivxqeMZdc3gdY+1ofPLg5t1scJBaxekMKPl9+GKMnSvaTZtf3fVuz3MOj0K+Jw96t/Ilk8gzh+gZ9Lj/ySvZMaw/3ihBL7Kr5V51pc4GQuF0oY/9z6VvNQ4jGOraFOJ5BzZqDTzOUztH8cDu/a1e3fvbG796593d234ySYUwMDCRzsF3Pwcd/Fw4l3+cYQbn3llqel9a9gEvY+Wv7kY3xTh0Bk0ueUW8u1lV+4H6/9BT6pi87cJSt8wuNvhRJ/g9y9VAxfnOVUnRc9jvi7OIfF4sBTdcTRcNJGNgm9C8eGaSLw/T+7/CpbVZpYR1qUKN/qQ0pF8AWcw3pogUq+ldS5PCfyMUjXnMZFkLQ0C28Jq0ras+WEJhX3A+gvYZS5XORVngoxiG9b5GTHyN3mBfUy/tJuB7BIFzklqfo4asHFhTYdBW6sSFi4wTJtgW3ae3dv6268YTfKvYPdnSycTck3GQfWoEwxBwSXgg1NZq5ylPDrr3Mc8T9wpPBr3Zee+Aan4jxaJ6chK9gu7OsXzuOIGvlSy7d1KCDkRY7NtvNzWMG98MJL3fe//70oYB997FGcnGPlAm/1xK9pLDec68zBe6S/BxO4TUYrOHnOJbfFwGc8uW1a33uWDCVJlFf0aY9EV/47z1gQRQidVSshDw+QG5hP5Cb7O2VywXoORYi+s1zAEr7KXxXotj8Xq6x/5fxV9H39phw9/n4UKdujhF+dcUen5efPa/08EV9bHsV+AKXJlAs39HH7Zi1EWBe8p9787mLjcpftv/hX4hulTzrysCThIMKS76OJx+OMv40m+/2Ez1xS9jtCskEJBg+jpRh7Ho/RCthHGQvkxY7Lf/tjxoVBTmMRR+piyfdB/OHDeIwPD79yEBGewiAKtzSI4GkYf2E2Fc00wvPXf4bLfGQp/nMC4gDndMsJ73CrzousauzI0bM2wDRCmNHiChUnXiLhX38tA78FeReHFnv0ufDqcTHWSKBQ00mdbDimmZHlyzWMmPI42TbM8hFXBhTTNTqwzoxcrfik4Bc65C5sGkkpFGDNIsIVcR+au1VhSRFDe0X7xKoC0s5gerMcpYyPBhUtyKLEWICJrun279veHTp0E4LFZjS16xCiGD7JV1D6M5G5uA/dFUvpV/mKG4/95dfha+GcifsS5Ewf5Q70d6VOxckLrxzhWdPwUpyIl3YptZXEAUr2X/CymsKzSgn3m7pV5w007msQ8MsMnLgwVDuWvlx8zgqjs9DguKT+g7Q09aGVwruvBVP4ea6vg98WO1H5Ee8czw3Nohf0+Bf7lEqJ5O/QygU4BXa/rmIVYR7B+soM+04ZyBQCdmzfwArpText3dbt2bkZ6x98F4BDVtFk7uQp7JU6WXbgF7LoD8pCmQ0HfyCQznJIE6lIfBUnMh4/qfS0fkngJxUnT/6LzmHxWL+aQdABln+r+v5PtoNLGAHJvSZHoobFCYoTVxE8EYIv4KlCD9rYXu135vVb+l+AhP4ixt8YfKFa7uIiFra+JFVgiluF1lP5cumfAx9hZAZtSDzT10rVBHXhFgEVU/OYtepg0m8bNjAZ27eLiRd+EugjWzfjVV4nkyo6KNPKCChVn4WBNLWNihUwBetjboQFPvjbDvLd+CUIGHEFK9LIgNS5dMXixOOIcappeWKBRT59SfJkHkMQBajebW08DQPJYQl8YtguCsEW0TuXeJrdCP0r98oycfqflrJegeED6e1/rf6dzOtbYo2KE+rf8GQf5aFPLZffDfzqbeRV/8lf+o20Pxp9ypOfIDzAybdSTpM4fc505sBPyqRAyJGPTPKnpjj6HMHRrSoMDZSZfNO/zcUrCXOXRyiYCTj5DWD38QojXwpv2gGIkAMR7b/BpepLIX0GTfJzz7/dnbuyBsu9d7rXDx/vTrNFw4lgREwnE0yW7J/6PPHUlq34r1rNZPLSOawnEaY5JJeTdY7SZ3m6jD8T+Mol0/Ou88BNCM+exnPq1KmsZDqJz4kGFNaVPi+VHZvhXZ5wYZ072T+DxYuTf4VzheMo+DTnhh9oZbGRCb6+CRS+T8Iz5lmhFAd5h7RR2aMgvQEng9u27YjywVMVTuEs2X6b4y5VYkETlZjyyYMHUP6zLWfXzvVsUTrVrZ+eYHvOWt45khkniIsI4vPQw62EVgJdlwtg6SrcQ+f6VO3fuuPPvlChuRvTy5rQOewRrMJyHLEWJyhOfuGpOpwSdxkecgXlSXycUDZN5SewRNHJ4pnT8Hz+bdq0Dgeys/TxUpyoiLug4oT6kk5ncKLopEWzc1d8dYioBY9KFi1OvLbnOOLexwl0Fa9LF3TSqGKp93HCBMdTdc5xbPNV8tGUXeW9cs0aFlOmyVPFiVt13JZVk0Db3bD8Adb/SIPWfEefpWe99/xnJFH6D5UbKyWQLP5X45WpisztTkKzsN1bz84ek7Ptgz9abqROxjNlCmZZtAsmU/gAU35YzQRzLdvm9u/Z1t19163dbbdsZ/vWeiZRnt4El7eNw4f807rQRmfbdFtIFjvM07KPlH+0vCPFClYtLOWnYB+F/4eQNvqP2P6CA8lMGTx5Kqf3buPlE7yoFin9rsXJQvesipO/4DjiVZwAho8Tx3dbsuBVnBQufFqm/IJplzBbmUeff5fl14GvShG3i5RihL4Lf/D5PJYL/x91b/6mx3HceRaORje6G43G2bjvg+ApyjqskUbWPI+9O/usPf/gPs/uM7szj+0f1pYsSxr5kER5x7pIUTTFmwQhgiDumwAaQDewn883qt63utEAAQqU5AL6raqszIzIyMzIyMjISJF2O5vKVY+d9bJNWSLpL0/uT+tsf0uNvw+kv7nRT+TdKa904amb8Vh2ua3NJBThoVuslBb+N6F4af0hXeWLHrd7G7lvnj+NjVXwbkCh9+xTB5q98LENbCdZM8Ekn7ZP1MjlQnBriv3CMs9Rv0ePfdD80w9+2Lz91tvNV7/+Daxhv5ytf6OjE7Ey0fJM+ctTuCZRjspXaApp/yqjICc+qE403/7238GLpps//sqXsezbgiy0KgprKGbxKYJjnIqisjxxLqT/Ksmhw3CVMsEJ/GLJRZ3Zt9xqeY1tiNaXCpsxZHVP/Il8Tj1qAV/9LrWDApktQih97Z8qTHQablzrvPqo47gkYPxj/LkafnYTpfk48cebjz78oHnzrdebUydOJv2zzz7TPPPs08zZcGyLBfEttqut0EGt/xxPraBUns+GlsVJV2/5TuiDrraK2yhSzHwGr2Y7fJFg/Le/WY7htTCXfy/wBxYniwk1JMPwaVjY9mlRmbvvi4O797r3f7sU3rtYFTaEOnzqx87zwiSDz4uDu/e6m58VaCX71F1dLN/9IsNwBIO5UtODAYHG5mTIRleTXPMhnhMKnmQenoBxdrBV541m1fhmctGEygGLqE7os5ogTGF4PQA++fovcdqbKWz3tkfh+q371M/Jb5a4SiK0omc1cmN2l+XiuwgmTTE7j+9yn74aznXrpsOMyCLX44Zf5TBr2bV0dxLuW/1bLgPmypzDcPGg0Cl57jV4dO+IuDguvUnVIJAtuxmGuG1mfbNn385ouDesm4pm2+PKViioO9mGKdfkDcYNgx1B2PByNcN2UJOfBD0U/cWckohmalxrpL/Fx8lrb+kcFqUJp+qURl+1j8KSZVRAsvzAhsnNk0aGWW1wBQJ9KU7G8HGiCbj1VpM1oFF/mupJt6rLPKbeq0Xfv/5DzCoaycW5Wo1P/msrozLkrepfPK0McTexyh3iQq+yPjEdAyA0NoIC3jzOvubZJrWSbVSbN003Tz31BMeCruc0nRkmIqw+AtaVAsV8ZGAYLRlTJFcAVmKST0iwM987mWCJrYoAME2/5kSBdAwnGF1PLsskNfcWRVpksgde5Rz2lxxHfB1EVyNcMYmnP2vYaMz7ld+2kqKCnooTTfidREkGVwpqqw7ggFE5hTrJr/IMwSwGl89Fqypf1dZ94RPdsqaYSWeOxq7ymV/VS+VQ8PGdzhHYcziUVCpYCe087g8pBvlbk3+2CUD8nds2Ndu2bmz2791C/eDEDEWjPpzu0pc0rdUHRJRTXZ/owbcUC0tS7X9QfupDTNNGqFSFGy9XUOWPyHzQTYsTjyPWcspy1IA/KBMN3q5os6jym7tUK9gFwbcKE4J877Np/x2Ulv7gVk7hXN2qU3UUelcjwAlf3pKY9G3b6/DqUy0l5XtRzfj3738L4ZtLV/4a49Lil25/Mg6u8J3gUnAMLmp1FIUjQfARtuOMYd3gEcSrmISnLJaH+Mp3sKIQ3FytH3PxZpt0EpDwBb/G6a766ptzxWq7pGas9FlrwcqLLS3sIvvJS2817x272Bw7dbk5cfpqcx2NtFZQtUqLrx54pNZ3k5Nrmwkm6dPTo4xj15rzJ7H0w4pyDv9W5069jUJlY7N2w0EUJyPNLO1vPRYe6xGkFWLPnT3PogGnUFF/lXetFCoUa0W0BrPrdRs3hI+cxXJEJ4Qqa2zTOvMTl1mEZ+tjLf5JpjfgCBbfJppXnzl9HsULlhGsVroiqfDtxNWTeXT6vWET++Dxl3Tx4oXm5CkUPPxz/HMlFo0CE338gcBLt2LS/vST++Gf25uN05h2s5VuzYTObfF5QWdSRnGiIS9zHMkJFziSTduwAw2uIf2txK63hHtQmUP5hxCi5lSds5yq8286hy2Lk5ewONE57HUUsDcHPk5GWN29gm8xJjBM9rQWUWGtv5NbbOlxwlA+TlrFCbR2UuipOk4i9AVwi8mhlkROUmJxgum712Z8nFzCx4mOFLU4sWnp78TV3gkmH27vUSmyCcXJNZQt+kWYoh70TeVR56NMJDxVp3yc3MhEqFqZ5ffp0/Y/erqNVZK2t+I/juKVs7nb7/hJSH5N1NaJ9FfOURli/4ri0ci0EZ01m499QusC28NNrExWo0jbAe9++sg+TlXaiMXJZiamq2lnlIQ2WYte4Ea/UBl9VzmBa2DhoryR68H135VLVBePP6LqH1jlqlL25E8SWfbiTSlYG3Nhqge1vwAguvAd/72UqTvnsJI1RDMC5byK4uQVtur833/1PSaPLGys0OJEHKWttWy5zWeI9SfCNzZJPqn85mmJqyUxDj1E+Z0sKzPUxBkrBmDJf5RHY3FC+7WP6C/pNmHCqPYmJGDAS9pGk2+PCj84IwuULAWNwCULUeYuoGqQ+a7IDEH5q/6Sug1Uy130vMN2FJbzKBN0wEpurl2c2cox9k8e3tfs3b2RdruuWYuVmOO/DMayKj8qfztuav2d9LwL/sy5i81LL77cvPHWW80zn3u+eeKJp+CJKEug9HW3POa48VEUx1htwFNgJmQLH6Qsc/BRlQ3nz51FcfLtLNx96UtfxBJwI8WrPuB2cukvH7e/3Ca/62y7dVufcuMoyhhPCQPF4BN8LTc8X4V5rNhQsmhN7baqKbfykI/b50iePLN1kz7oiT8eM62iWB7ndkItRpxSaiHjgoBb8MdQgFh2T167cvkqi1koRdj2o+z5i5/9tDl29D36+srm8BNPNH/85S/jpHwnZXDeAo4qWMHNMdr2UP1P/kOG4k0N+dtdxrHvVux66r4N7iYdJrlvcBet7v3fQRIeulgV9ocOP4oTKyP9YUCsheXoFzXPC8rIS96tHdItIKTfCFgQVoTp//6u4Ufx0UOghAjRt4vyz6LYprwTr1iC7MtObJn4ywbxYg0DoYKy3nHCRCM/c/4qPk7e4lQdnMNOeKqO++7MAoaCxQm7hPnzkhG0MFv4CQbEYviyX/FJW3fA8CX4GNgiG1r740fj8Oy3isy9nhItb7y38DWbL3PNNhJpPNfcIzVlOOvWbQjT+azgd/QPjqIurcSNckf8R2DwXcuQTKYtYspoQPvM3WA7HrrZZjkTwvFRJugbcZq3c0uzC632Zs5tH8e8XLJU7uQJQ8kxvowE2Y6AdHeH1RtXYmwvoEJkzO9MIYBHoX/KUazAWtfHiccRxzmsxxErCCVT/d6YebhbMWXgKvBm2uGKE5x6lq0675/AxwmKk251oawfoBT4ytizjYy2mPp3LDDb9vLR4nRX9T8Humr/+UgeWakOJa2Bwj9pbVOD8sPdbdgSyHAHHkcTMrGuDNM0c0RiM/Geu8Ue9OW3mw1rVzf7cFx45Mgu7rswp/YINUtuGtSKDFQeh+sJCSlbhDwiZBJdIMQl+zWJ7x2ApKYu6WshJ72OQHCTntabefPORx/dqaKA8tq7HEf8wq+aj87idLDnHJZ5oqn5s/z2/IX9L6s6hHp3r70m9A6Slt1Vh7L8ETzwkxO39omsB1fX/wb094to3g9++AgRQmfvxCdymo50MKH48pywgDeSji9R1lFlK21T0PcuZrI63h3DhH8GS59d27ZQH5twSjbFKVqs/JgVf8uYpHkkqvU+gnWIAmm2P/G5g193A0wkPHHjOW2jeKWh3WV5yypHerF/FzrKZ8T5HJMy/RBYVBUnLeUVdVO3y2IeZk4tjLQ7XgWdH1NyCT9hQ6zqw/A3TUL6Ac32Y3T7/OL2bw7mmnY0aP/CN5BUeSQXxbaNtQAAQABJREFUymHXi+LErQMIM/pmCBohqNFbpAKr3h8H/BYIuVr34gQi0gbkhCiqg4sXS+3/VFFb/kSiT1WRqt+vwWpjNULdCqxMNBunupIuY2myKLoF1KBoUCv1L5xB4AB891D0NztxJJ4shTypdJLbjqGnK3T07DnKc/MmyufvvdD8y09eb+aWT2DhQNvRAopvlkeBVysL89i6dSfHVa5GML2I1cGF5toVxjLizM+eby6cPdrMbNnajE3ubq5i/j3JdpuN+jRhkn4OB6wqQqK0gX+43UWhWr8jLiJ4rK0rlWLtUeT2f/mAE3vhepy3dexe9SksGdayNUVrHS0bzmPFYt46AvRUBE94Sckp72qUJh6L60kMszc8Jed0tqJMqjBBIF+PJYpLsZcunoQ8t5vDB3dyis4zWJxsZYsRR/lOYT3J5MJJtbgpL2fhhzdrYA7/MCOcspOrvS3mP6l3qq7q31yK/1khtn+zlrOepo++9MrrtHN9nFxuXsTHyR1OvrjBiugsivFs48LnwFW2xo5QP27BuXS5jiPOKUWs0uY4YvqGK8OatTtRyXHEWO44GdEU3q06N7NVZzw+sT5mtVU+tAnFiX5nrmGpMj3N8aUgp6yiE/FxTiqK4gQaOt57JLK+aiZZ9bVf+hznsChOrBMnOlPT8m77C4ShLhf3f+kgVR/U/8OcjKO1o1fL/7IokxqQnnwfPNvqCxaqjYAWBQktHKvI+qnxt9Iqeuq7LdvXWITw5JHpqbFmzw7893Ca0hMH9sK7UZyR2MmZYxyFJi8hiZu5Du+x1Eo8wvzHZ2l5v/rP2Hof/heE8yMk/try22iEauji61HbX+D7k+xBlLy1rlyWLUzeKWkKAO+AZ1yjsb78xkm26nwXEY4+i4+TokTxyGzxscwd/xPRJcq/ePwXfiHBrSOWaQclDYI0KTNPZD8mWaLlzU8tzVv4Kt6VeDvH08o/Wi149PRVHCiLm201Vm2xGgIiA05nZeJ2k5RfSD36t+A+EX7Qb5tv8AXZhfIPCCyo/67PtHdiFw240e7ix2Reawp4Hoq+jRvWoCyZYWFmcxbL1qPYHsNHkwUzW2mlLF6+8oBMGVbQxlVgsJxJODIw5X73nfebt95+u9m9d3+zbfuubHO5yPHkWomPwU+0JlMmk+L2+cjWlgV5zJH81KlTzfe///1s1fnKV74S2c2eJtzyj8WCDbLHLS3eUGxch5dI5wlPz8EhelkVarVVeXpXaXIDXuIWQT+Mw88dE1S0a3mjQsbLxRQV6ypM9L+kklsFssceu9VQfO2LqQvyVf6PxRGKYS3ulC/d9uP2oIsXzjcfHj+O/7qbzRYcjh8+fLjZsYMtmiqZ6Av6HxOXcC/bqQ0oFxnTj8JfKKd9IjD9xqfAHjxWQH573xIx7+S7ILzNxPYnyAdcdo3qPv9+4EdxsmSZliLCgAItie+JY0498g++Dx5a4vbelwSebBYRvJ/mt4RPw7GKFlzWLyCEImAFj1gYtCHDViFshTk6C4OrTDGNTUZNoGnm6NjnsDj5yS84jvjFN9BObiJMHydmTWwmjlqc1CWTWXyRedeagFYlb8ucqO2zH/rJebdz2OUMNgse27wSwrMoVHq+LLiiIFDiN5Hl56l8nHByAeXUG3TM7viSJm6Exwi/j4yDgNYZBcnVOL+yt08aS0dgZ8sOP6Xt1Ska5ebjHBpuBQ2VEdNrRpodTAj3cT78nj1bEcq0WEBEcdmL+0rKG5M56kUNr3nXPFc6KsiAgYHWG8ylJWjKn7HjYcqfUjAAUAaUx2zV+TEm5mcHipOBc1jhpXDAgf53LDTw55k8BEysOLA4gYEfPY4jVRUnMFnrrWO01pMTUNtvtOfSjWfxz5XHomqImEDhGK+i5CVlhorcHbhDf9p1BjPCcqIM8ZcziTJrkyZJssmwT1oICv4ruM9huq0iaw3C/U5WF55kz/WzT+7CzHwjdUAZpTuFNK8Idtpp8ixcyyaAHEMnijy3w0prcWJ/dkCy3DZJTomxnQhfKTOFs+54st5VGACoU5zE4iRbdcrHiRYnwtQDuXmSghyKZoZUmIgFGHDcqoOJfvbHTgY325WDr6mkGZ0+6czsfv0vWeeH+IM6a+FD3PS5tqLALmXNDL1NaJ+xzeucWkFHxU0aDvR1cLSfjK7U4Rm8hz/WRpgQMjnBimA7pt1P7N+Fd3pO7+DIYdu2aLtFQXyX6XAQurWqA/LDUZoWJ2kYHQJFp3v4j4UWESuH6ujKbw8bWJygjAm6RLO661QdJ5RcKmnyVLXQZhIcozsKTaTZfeAPaFnZdPArzwDwp6VsPcWBdFdn+SomwGifrYul+r+doBRAVf+a67oVahSByCujhXRYkHfl62/Kayzre0Gch4MvhTr664silZikENYOKmjgV1sSn+4JeBC+6kMVBW2Jif5K+rc8ZQ1WR26TcHgYpDe7vKdog7w7+FGqpGiEpA7bAlp+IN/vUj/nV/TJPNhgBARvpA2BFSt0883/8V//rvk+DmKnZnY3K0anackoT1AKqFxYTv8dx3JsEosNjwfWed4pjum9hn8rvJSgtMUcfIyyzl1kgj3RnGEXzHW2wWzatiPt+fyZC2zpuUI+CLcwpQi70M7u7tZIJ/YTrDa6CngRp4Zuu7GsWqJ4zOUIdX0VRYr9fyPWDptmNoauVxDqdYCqUsW85TGz1/W7MZEFiukpfKVgweL2oqvkeYn4OeqSOhjHOmUGs/bpSRUtV1EWXGI7zurmc88dbp595kB8EWld4vZGaWS9+t+6UJFnRak/hjwo8oqXLKC/TaWSSHSeyaflfxWPj4R7mfo2lXNGi5NXW+ewJy5FcXIXy5JrKE5uYXGSY5KZMFzFKmSELalTKJUvc6qOLVSLE1dac6wmEwYnDdegTbY7QodLnFIm3cvHye1Y7qhMcq/+VSxYbLUzWpywDSjHEeOTxuato9gVKE5cadbfiSvEm7Ei0mQ+ihNWaB0XZlFWrRpYnIgHW3XYdmW+4VVm1va/Kvmwz0jY9Bo/2Ea7i3f5o/Sxpcp2FvQ/84YPS1Yn9FGsOUYZTT5qG+OLX42j+byfswWA8S3yDfe7WgfS4lehPHHCuQX55hn8gj375D4U354aQu3bF0BgJZOm4tlV/wAgb8dbYQCESx5vO9eCQehCflD9i5/lX4r/9flPld/8zFVGwb2Fb/oFl+UV5QQ+BHzoKPw7EQLFh5bOeJQJPnwv+BPqSXTXsDjJqToqTjiOeLmOowNH5aqyirgRUMB54EXk2/ZvsOUS58K6ffbDPfU/5L+PWv4OvnxDmcEJvmO28pdjiD7JrtGOlcnW0EdGqK/bjuXiR3zHf/FcBm9J+QtxvraIfgL9O/hd+cM2zBMaun3d92H5eTFA0tFP5H/y5xxOwF1elM+kW4H/OpW841i6zcysa44c2on8t492i/LBDJHJHB5QFxTdzZRLqbf8xPEN8U04WiLbV7QHPoEM/O77KL+3bYePr8m2Sq3PJtguvQ7+vBI+oMWGTvCty8hz4gkdtIJ7G6XLD37wg2bXrl3NN77xDfjEZqkHX3CRg15D3FsoWM0zihDguwVwwm00xJnjG15a6Jco1tm26WKfvOk6Cgzlv3G30WhpQj72YcdWle58Sv+zHsXD7T8qy6ewYLRsOSkrfEHZGk5hO4BXnuOknQvnzqHAgU8yRuT4Y3DR39VEyxsn4W8TKFQ8djl9HFguXCr/2cFsAm7VsZzVLgjguufUO4OrsvO94nYBvfbfBbWxjGfeCR7kMXhos+m9D9ItergnSj/g9w+f/kWt5irEFqLX0q4f2Is7fBymNWxIy2HC7qm7J+2Cn2EeXfpB3MFDl6AXkMd699erS9+v7C5F4qTIxLIVOZBR1anwlgHYbW1nhsbk3/i+0/kUaP3ne9LJpHm2c9hgtDYx3hkUJz/D4uQnHEc8xladWw0TdvkCHUOHS24tKPg1KTAv05pPBqTAK/ipImEEPnCCj3gnevLxjexzGe5lnl0j9lvit/e8teUnl8SNaAqMDn4EJBw0qhWVia/DqZ1HYX1W8Ifld/ABdzq4DDgaeQaJ5Uyu3I+qc6yazIM333WQ54RxhFXzeTXwaLZXs7d7M5rtJw8faA7t3cFRYcvRQBdFFFBIBTkVcBxspbv/6x76yWz4V4oM65+0ML2QzMjWBbQkK56Lygvoz8cavK1vGKVxaVRanHwLi5PXVZzcZfV/OX5OCGeaRf6Wp0xObX+6olAEVsvuwOh0Rkdw1xG23//QE2jYqhNhh/JIJ+KoxdYE1ybihKcwk54tvqLLX35TGN+MZfvjmXSAJa88ZlDrFEd6a3dgVNDSRNTkavfdLnQHvwNzEMMsViGEReGC8CyBRnHwNTd3DaeMU/Hs/xxC3r7dm1klAwsIE123zFzcA9+yFDWH/c9+Jf35RhzLyo1n3t1uQgmTAsRZa0tpXZELTcHZQTjmpsS33k2vA1gH4zfeO9v8E85hT+BkkjNJoSGDDrlDvvT/+/U/EeG/lGNCcBk66Bx2Tejb1YPCQ7btgCHZBWfvS9Pf9rd0/09NtvVj269TSqQXaWw3/NMEVbjWu6sTbq1xINYXhZMQTWRdBb6L81f/7A/bZjY0h/bvaPajVNxI/ah4jHWQsKSBqMbShPpx1YeyCo1a5pe+mMl5wX9Q+/dbmpt3/vzNlkUIVGVj4ms/8CORdQ57HcFksN3L+g5K0px2RsT2NRnag4bwbRl+J257f1D7N55lTZviseN/wqv2R84t/JTBSud9CH9Y/rRH+7D9kvtVtguMaqWBjxPjd+lKqUk+AHmc8AMi5QGa7USewb/QBsFbJAKPdh9FZYhqW8EkWOGOP1No6TU2MZb94aP4fyJ6pRVl8+/Kb8Zt+wuM5N/FrbZCVFK09KpP/C6+yLMwazTItFnbtGjhhFf/DhIE3kBx8l//8nvND378cjO6Zit8ku2O+CjRMTsFiNC6ZmJtc/DAgebDE6eac+dONzfvcJoLwqer+Pt3IrgfRNhezZ4feMXLrx5v7tLv5zmp4MOPPsIwjoWNjDFYvMHfxN/+o+CrefbM1i0p/5mTpyJUW36VJvKTnHYDzlo7TCI0b0coH6XPneB4YnmEVifG9dLHxtVLV2Nt4Qx55949CMMTbCE625y9cJbiQgTw0DLsANtLnzm8u5kcwQqrcwi7fVOznf47Oko/pdzWX02urQT+uOzBoaX0T911NOU9AdaLNLb9y0+qz9ssDB3yX96tc77LG7X8ieLkV2zVaSbwcXK5+UWcw47i36RO1XFbkj4DPmZysAKFSmdxIgyVTzfhQfoukQ6jbC264jYb5Ayt3S5iSeKJIWtRfrmnXwXLGJOWMSYoKqos32Z8nOj/RQfM69ZukFSpD9uufgJuMCFx9VcfJ5dRyjgJ0s+ACsKbKk6w4tHCx0mRFkBr2IrsGO2Y2y9/N/4RXDQEf+FLYR8N9C3tH9o4Yodv8pxxJ9/IFZr57qTY/q9MNafTM+K5Vdj2rs+KxGJM1boSgxLCblKvnDCCT4i7WGzO3b6KYmglpyVNNE+xLUelybYtbHVYqzIbhMjS/qpkaZePNav8kk9Bl3BiBO+F/Ofh6v+B/b9ybrM3pmUNSo8XvvkCy7GQgT1l0Y+Zbdm1koIMP4MnXMe/z0tvnOJUne8yvqM48STDkEC+p/RnXtIfopGfdbew/oVhwaShOfvuc6IHtjkkT6OZN5d5dnyvHSZN2X4nVtv/xLYPXxjp+w6G8iH7Ne3BybdbzrRImULJl606bGExn2zzkxZcnd+7Twu/K7/WdcuQ67SOcGAO33K8jraD8saqD9gqryi87dntyiom7bPKziMeBb/8Fv6axpsjB7CGQmGyY+s0vjmK5pLABRwKSjnA2LITIq9J36FOe0QDjlt2cJlM9OPHTzTvvft+swl+bN3LM6bXbYwiwVNntF5V/jKN47/bTa0RLTbkOy/98hfND/75h1ii72j+y1/8ebP3wMFY/3na40r4jI7Cr+E0PMpr+FMcwaIUN8vb1EEOBwBXLUoNvAhfkjfZv3U0PYllSk6PhB/pX8m+LQ+T/1zFSk5rO9Mpk3nijycdqjRxTqNMLbb6wJrHR9Zl8j19BkfoJz5iTPsgSuBnn36m+epXcWrLtmrHqbRLcEv7oxKrrUJnAmybXdvNd2O17S8t17EzEP2ty3LmefDQfhDpLlYe691fry59v+K6FN29YvZ/h3l06QdxBw9d/F5AHuvdX68u/WcJH1raVO9DIL9Ida77xsnXfoQ2YJCg/94FtveFtx6QLg0RPgP4GQglr9nTwOxcdmOZZo/q7fvi8tM8bXBpdHby+leZ0ATBVwZ8hqP6fvqSp+q8zqk6mxE06HAwEiRq0rqSqnRIOXkvsA8LnzSFUkukjlbVQarzdHkZUSDeeM5ktAszDuGg08G3Q7vy0F0RMGCAakRl5OtxDqvp8WcFf0B/ENJMcRRTM4Wn27fmmpsMICE7gq6rdU4O4UCggnBLny9nUtfZdnA1+8mfOnKo+dzn9jfrMZG22B5FmZU2ynyHfZa2egcbtbkqLlytkB7WZq7Ur2Wt8tbES2LxvY3S0Sl38xClENUHMzOiRPbVdmGtL2u++Z0fl+KE44jvKLBn8LZ8fgVeRB6ZHWll/m29eUQoSyww3PJx4l548bcsMW0kbtozDTrWB0pbLfxPrv82ar9s1gMDoTAycchkPEQicl1+u8mKzvxc7UddOUIa8pi7eS1OwEZo6+Osin3+2SPNU0/uafbvnmGlEJNr6mMZbckTMFhQTfsqarUIUOZwpjB0SxH1VluciiNl7Yu1JaXqyWoDk7apG+YkxXorWvjdj1Gc8Ekzxl+7VYcjSeMcdlQfJwisln1QhwU2UJO+yl6ck2/Eu4TfAp0SOrgCIINnFBy0UZUZn0x/cg+8yju/Pfh+IxduDKdUspynaycVF/pY/6Eb8aBrlINIkfYV/zQfv3TxdHwgbGUy4Sk5B/Ztx5v8+mYSx5nWh9sGxpnIEJtMiqahlxBJT0sDnJhowSNNQxXimj83/iVsgLvvXHxvK8UX/hSSrAfaAK86m6V50dZo4oA476k6WFaVnx/aeFKQMvmKCwGDS8CIv4EvfnngZqT7w/dT+ot38+qA+Ow1KEPv0+I4FZMIVX6FqWzVIWPr3+OIFSZzqg5ljcIEQLl/BvCrfoblVxALPtSVK0/F36v1WJXLaB8KufL6KEswK3aiP4qw5yRci5OO1jShHiG6grd3yyLYELIfT4JJXz5U5xt+bGnZ3gZ5q6erKuSLWqy06YjR4KmZdtO88K+vNi/85LXm1AW2WcxhrYTzUScWttkNGzY3M/ydPcdRv1gu1Hh9GysItomwEv/UoQOcOLOf9n6ZFcFJ/B+cbk6xtfYyp7Cc4xjHlSjoPSrY+rc12eZuQiMn/rt27c4k/iQKFp2W2re9O0a5P10h+hbC7mpWANfj00Rh+AqC+jlwuT1/kzhannIMJfl5ye/1i7Rj29ZMeM6dO5M97G4NUpnwMab5mm7/8Zefb/ZtW9Nsg3eu53Sgtaw06vSPKFnMUMBexTtv/JffeJNPcLdeQn/pk2lJIFsdw/ZvJAP41F15rwDzyRMPTgL1cXIa57AvcRzxHZ3Dfni5eellfZyMsFUHCxEm+B6n6cqnq7AjOofF0ucyE5EooFScQAPjOUnR4bDWOCqW9N+gBd8cdF3Ds/T11BtN793ylu0K4LMJxYmnUMTiZGpDsNZxo8qrcg7rFpz2VB2dKaI4nsR/gJZ4UZwQbzVOU7NVB16jMsdxYUCE+5R/AY0ClZ+2/8tzpXYd9elH+Zz10fFs+aakpm3RFnXYmLqy7ToppFK07BM0aKbe9MWyUsvaW/iHmb3crJsawfHr3uYLnz/S7N7OUdWTKDlVblbO4aPpq/w42XX40cdQOYU0Vlub3D5N/T+4/F3eZt7BEiTP9GPp8VjgC8acwlMcdRwfHY/gsbRN+1iGfXC4dqs9Veev/56tOuvQR7FglcQlB6jQqvFOnP3QXp+q/h+t/MExYAEceN5sA1WfJUPwzsDoaTv2EfuPfSRyGfKxl71auvovW3Us/0PQfyn45mc+/ncxSB6WRT8Uj/I5nVmrXHAM8wAFrTpU4sCmmOd4It81+i7HfdP/102NNs89ubc5sn8nW3TGUXCyzUVFr2SinI4/bgGmxvIu/SkBf86LCM2eUMP4AKwoHWnXl+jP//bqa1iNvIvC4wBzkxn4q/xiLTyWxS/+uXikM1oVSSvJ5zYnBn6MxY6+UJSVXvv1a5zM8w9sE59p/uLP/7w5dPBQFB1FV2nt9j4dyeJsGksO+Y9zhjnKG6USdeTJZ9bXefi7/md0gKv/FLfcpM7oxOZn/XTOuS/jk+Syfpoo8hRbK807i2vgq4Wcl9Zwq9pFUceAjz460byDhcw7lFffUIcPHUJp8hW25RyKnybLG8sSMw0Rud9T//ImG3gXhzeajW1sudb0bTAR6iK88uh9WhxnEJcHs+6uxfHy3gW294W3HpA2kz9g+FGcBM0ekl3ZvS8qav/T4LmLM4w9+DR4SJxhxDa8F/C7hA8sIRe+xXDsxDaSrllVK6h4tr8OPbV2YboKwZUJ+ZDOSQbXHRq7It5ZBLGfvMhWnV+wVWdiM9YFCGIOhAijZXGi7YEZOJBy7+D7KDzz5NcYgd/yQfGrSfwwRuIQnnxSgmHaBHf4t3mbR12mHJbfNy0KOvgKofo4kdHY8ddv4Mg/J7stbh2UxwU/9BdH/qmguYOg6cRdk98oHhImfRDoESTi+Aht7Mcfe+IAe5k3TmZ/7+EDarXXwEC1JqGEMMkqc3KmWB5z6oSWcCZs+ZYlgcQu0uQXps4/mXuUK9DGFcsw8fxWDS0ofyqro0ybFQCcHFoG/XJ+MxYn+jjROSx7rglX8aNg48qh1gTCKKUA6YCvtCpzVMFzndNG3j/+IeXD4sQBlXqKmSZ5O+CWabZ4Wx4uwtO8xIFn/rfXEPN721/FU3FFqhK8yM42b0s3Dwcx62yWFWCdtub4NJaLr18+z0rqRzHNfA5v6Z979kDzDCtjWza4/942VuBdmXPLVEfROK2TvuSZsgNXzm70Yf/zpatPceGfSEE7/2UAuaNlkljT4RTWbNMKR9BQWBZCOpQp4wqcw2Jx8kOcw2JxsgxF1l3aHGjRM619ohPX7p3i8t61fz+mRYHcZbfqMGjWccSlqHAiFyGISdD96V+YdgW0KOImVcS1gy9CNFexoVQKiD6VQsFY2XcPkgoLYjiCZOIqVU6Wwvmrzs2aO7PgNxILkyePHGi2b5tGYSKtyJu05i9/C1CR4Cr4bT9IUKDxpYQBvyeOkX0KsYa4J5h0XflB2v8FgpLYL3yjR1NEfiEH3RyLE049YbU4tI71RmVt3GCwIB/zbOnYAuxgtK9Lwi8sjMkVBMEbfNqiVzC/i+s/KBtusvxVbfmWCRGFUhDKCmF8LiAEaf5L5M5SJpkm9eOFDzpc0r9w03+BK5ZZ4ZZKNB1bj/wvPngsr4IvnVIhcAohbgx/JnSZXJJF2niXnVS+3W96U68UFpBv3gKj8sgLj5a/e64vvYCW/oYPFCdm5TjLn0q28Bv6sdtEfv7y0eaNd880r3KqzvkrbA3BPE8BcB1bczZvmsE57M3mJBYht+FNKp5XrtJ/jls5rnPk+c7mS88fwq8P1g74NHrtnbPNq28e5WQerFLgw65WjuqEGviOf8JeN72+2bZjGxOFOfbGn2ZlUZ8mQ6YQJ4LEd4+7VhZ79uxBATXWnOCUg0soDnRw6uRCKwcnQDeZ0Es9Bef1nM6wjkn7OY6XPPnRh6mXCSYdxvF0ma0zm5ovf/HZ5utfPtxsZ4wbZ9uUFgpyRhUvWZChzdm2gjQ08l58VBrKEeQR9TnbGCsGv8blQ0t/czBaVZQ9jScCu+oxD+3O3HLq4tCL+DiZR3HygVt1fvlW+ThB+XoTHyfu7V/FJEvfJSuxCNJa5/IVtuqQmSfn1Fad2VicqLSL0gPeq7JESxL376tU8q6zXq1Ix+hHWo9Iu1icoDgx3Tq36pBvHUdczmGvxTksW3Wg3xUmKrdYcJjA2SwEpl71caJzWC1OZnPM81osTiRW6DagQFt+q9r2T/l5BH79SZ/wo4TUmyFy5oxLxV3BTd5CqAsBPM/Rjm4zufQoa/1X3GL1eY4tOK5bKZeYh1s/HdM8XvT27FUWHNZkEvrcM4eaA/gH27xOfwjCtBbto/SVjPk1ZoqnV2JAm0xWybpy7748ev3L7czV1mfZzd+QQEpj6b4kMB/TftqIjwU+Gab1yxvEQD7hGE/mWkk4YJf8xFYdLE5+qcXJX2JxsqqzOBEZFgKok1ArCJpj0WtYOvKr7Hr1X+XrSvko5bcvCqMuU95Lf/mYspDNIMpgyqfiS6W829m0LFOJa/vwZKTkApLhA9z18ZTyd2DEn0i23YeBb362nXm2vKskcBuddrwC8lRgff15MpjbTlQAa8k3qgNiTgq9MXsFZcPHOYXvCLLfM08eag7u3dCsxUIK45MUPlIL9VOM3jAnRnxMuxeM/aTmRco6RWdwQnEyQvtWXv7NBx82L/z4xziHfad5/gtfaJ488kz4g5adkU3IP3Iwb3YJ+9t1aHcFXqSPEhcFfv3rV5tvfevvULRviOLkqaefBgXgQq8zOAa/jeWcVu5uMdQyzX6p1YiWheIq/ZcxqKgYvnABZ9bw/ml4yBRWctaxdac8ngMc5M8ku0meKtLlVWuwflOR7nhg+bXkcf6l8lSl8Sw84QaKlDOnT8VaUcuT6yh+1jBWfP7zzzV79+wNHMlaC2M8pbF2LZMPXr3691ncTFO/1f4iP5l20WW8Lv69XytyF6fyWzpW4gwjtlB6AdL03xH8oeKkT7BeeRK8+H0Q994P94YMIg8fiCRTs3qXvBZnsvh9kOjeD/eGDCIPHrKtQcEDRltsn09d6/AhHYMwuUy4RdVpvbZDRT452aWhK6jA1GSCnqjjwOtWnZ/iLO2nP0dxosUJW3XkFcUcMB2DyTjJkwq5WlBB5EHwzSNJ+DFDmU2Lp/SMgFQ59srUBbR3iCT9C74U40p2hClZt/DNXh8nCptagEyvx+JEiekzgl+9mcxhOA4CMimZiOVyUroMZh3hWXyh9xxOLe+gBNFhnj4a9u/e1uzfz1GS4zL2ZdEWu69RAck+WTWHIgBGpTLC8jsaVjs0gkzEuvQmkQQkwXlo67VobbhEkBak7toJYfejv1Fk+FGcfPtfENZPw1gnmvkcRyyzdOCnPUl78eA/IeSfKU5bbgLnO4uTjyKYq1iyPTs5Cm1g1Ar0MuasLqcM4tu7AuJh2r9KfgZm8td5nwOAA5ADRBAUQ+A5qDioX2e/+RyD5prx5ayEbWDQ3NuoONmxdQKTfybo4KVwOOeqBftdx1BSWEZxvU2+q9i3WadvUFNphlV+yU/Ju2Yeekt9KRPSWzQG39QUAStInDp1QE4oqZV6O8UJ8Nwbfbf2QtWpOgMfJ633c+hexxFLrFRzwRKIwMXbJUH6n/2kO1VnordVp1NoOXguuMxyif4nnGRuG+gXOK/8ANetMWVxQgQLn/5PMuPwqrDjyqV7kjWBVajx5Jxxtl2sw6HlV7/2fLOJ06TWTDFpoD7NU1Wv6hgeQ1sqkwfe0h94Dj6G8RdQxiRtCCP/891v92//ybyipErizI93J6peg+OIrSaCzl/U4kQFZ8EKOqIU3MDX+mz5X8E3XrJOirx0aOWl9wPch6N/m4b4Hfw0p9ChD3+Yt0KV/S8KTPBzf3QsTlD+esWvgHQg+WL+GxKK/aD+k4T3h4c/aJ4BJoksrFmCl/la197SPmr8m8CR5mr+FAT10eHkTfLmMi15WP4hYf3SEpcPbbNI9MXwq0zENnp3mSeJiu92gcN7hHP7v0HgGZwlPJmoANXa4ecvH0PR0TT/34uvY+1wAX4xyrGWm3LkrdYXH7x/PCcfuKo/i6XBqKbh8PzrTNx3ctzws0/tYkvQ5fDkd39zuXnrGAoL8lit81D4UynuoRWdahKrlB07d4T/HXvvfSbqV6O0Fi8Vy9JNKweFXceoLeTvaS2eiHP+/KVYVhjuGKqQHGswxpe1KKo2bZyGf46gYPmASdHlbCvZien5bZwpnmRyMMWxwv/xa19jS8YMPk3Y0sKpVq7wOk7EgshJEvQoalGfhJcSEVrJm9ov0rulKHyYeMBfAS1pFYRzdVn4YDux/rt+zav1171aIzmO+Ozl5pevvgk+3ak67+Csd9Vgq06OI9ZfixYnhLtKflnnsLSnKfy1uHp9fdbjiMvHSSxOULRMwUMvXdHipHycqPTVcsSTLGJxwsRDhDdt2hAT9msovLpTda7S37qtOlqpzOL4dfNmnMNm9bh1Dotmdha+OIaCzFXxKE4GPk4ouCRJ+bl79cufBt5FqDbZJ2GN//JeKBsZwjoxMwksz2RSaH3xbRWOc61Dj1q9jeXNXXi3iiatjRjWsf7DcTfhkxzPOrN5LavLe7DcfKLZuXWMRQkiMA7TrMiKNuEWZaAuU0Hbdl6x9LI6I0cZIBqgLW6fTv41R682o/R/nyvz+8k/if444acxAg3S2rSzjTTHNUN7y5c2bOvWxwkWJ2+eGjiHXaZD/vA/SEHakgPbIojoJ7R/S98vf9cx5Gf3LX8lql/xfQD9M34wIbcfR7kNksq/cwyMOkZW1pqams5k2TDhh7+Sr/C7rToDmvdh+/wJ8C2/+LklSMWdC5RuX5az3MWy2PmOPMSBWmsoHXdfuswpkbeugteq5sDBHSgy9jf7D2zLSV9jtGfEceKRg+V2ZSTOjSU1yABDfWLRVB4v7zJAelofwPTmuEqoyvA33nq3+R/f/37z7tGjzdf/5D81X/riH+dUrJvMT26h7LZNyC/0NaLfo4tYHnqa2ios2zZu2AgtbzWv//r15lt/9y38rsw0//l/+8/N7l17kG2RRVEGncPhqlaikxNrsF4rpYnb02PBZf+kn2ktfRGe7Qk38nZPw1mNUqajv63PscFTvszzAjh4DLphOhefXF0+TVTQuBXRPCYYa+zXJ04eb45/8AEKmfNY06zH+fcejl/fAG/A2gfZeS31r1hZTRXiWDv0+/u2v4oyaLaL29/ABxk52T7Sdnz2WvxeoUt+uG/UQZpK9qDx/x5498303g/3hvQBt89E+m3hL1ScJMMezayV9MglgBvUw7IeewH97wn2xyR08jbaotjJT35oHef6jODTugBCsxZWoSVW6cR9+AonMRVPQVucTEe4A1EJlKZ0cKyJhysKKk9O6+ME57A/wznsqtUoTlhxsTjm5yTlLoqTgi8rsMwPAb9HsPSD7p2MIhgFryKgsCxbGE7i+QJ84maiIf6Lyp+VSTqveeeCUc0itHgMlxOBdW7VIczLOMKoh8cDX7p2ODnBYG4bwdRz7S2HW3dWsLXCyeAc58C7Srp96+bmiUN7c6rA2kkmg5g0qECw7Cp5KA7M2h/Lb93BioN3wQr1fZclO+BKk7BnC0hAhBDiGkUkUvD2lnzyge+27LTu5JEEfDeJdLJ+3MZ1GwDf+nt8nLxzprnF0bB33Z9PWOGhQsKhIqNI8tTc+g4KANuH/k0UvJxQHjt+AgaK5pwBVbqp3RaWg5wTdoXT0l4X/Cq/z0UbsmtpnZLxahsumlhEL9u/ji6JCVNHGdGWP+2+YhCm0EZ9YAa/AqfHM0zKnzi8u/kjnBYe2r2WQYeTOMhnJX5OFBKzCmCe1gVlq7JSXoOQAWrVtHAzMA5yuVtnwYsf51J5ppzWl4pKB3gj1ZGEltEP0A0pKoKruJuqTevAwZwqbUWLk3/+0SvNR2dwvMZWHZ2rueqzMvSQ5kT08mY27S1lEQx5XmRlWAdusTghzEFM6ywTRZBtC/Cg/rcU/Qf8B1wKvgiIUyuQ+5pPHFXKqo9KkzlMUt0XvxxBXOeBG7EUO3xoT3Po0AzKExyHZRJHOi4t5eRH2fpGvaSI/ggjL7ZFH/hLGdo4gU9wS1dDH9T+Ky25pP1JIPsDddPyE7fqYFUeiwjp6VYdTxWJ34r0wUIhfA5Y0RcFP8PlP9a5tPa54gZlniXdQvi+J1nS8Mbr0u1/QH8jdfASv8u3woUpfK95CqBJr5PvOIelPbn9RZjSNWR0UttL87jgL1V+8fZISIVw26y9zqNwR5h82RZW00c9Mcd+HZyiRLPPkRB+kioWdS/LQCTruvpVBd5b/0QkuWnNRoHZy1Zi+c2nvSU8PwQI312sgSMMm7qMIe1UmqE4YUX557882rBTpPnhv7LF7tzHHCm8JQKxVkpnTuPThJPHPLHGbRirKF+zHL8bmJDLQ/fv3I1Av5xTGV5COF7dfDy7url8HUebrP6vxKrhJpPv21gnyD/X4tdrM8e66tfpNKcwuGJovYpmHP7yEOUy79JwE2bf+tE4/sHx5sypMwjCKpWhuOYuMB+FevfaT2GyvmE9x0kyps1ex2qC/fkeJXzwwJ6cNEErynbHrdtmmj271tO3Uf5Tfyt00kxOTkSKD0tF2q681EcJFqJbQ0Tif4RpZ5OEZKxRK51GWDc+EI34VFTVvyF+811oZtxe5serR4ef8lSd9jjiYyc4HlSLk+Vj+KBhIoAz8NUoC93y5b5/BX4dwl66pIKgYZIzyRYrtupgzaripFOIuC0gvlCw4HMbwCSm91pm6JPA1dnuOGLQaDZ6qk57HPFajiMWfy1TRpABVFzd4NmFH7f0aPXi1iAnJvJln0fYkjhOnipOZhlTxc8x2n9e9yu/5DCG7bqNWhQiQNI7QIkLtcK7k1/rxHCf6U/UkX5HJO08vNp6HUVQGUHg8SQ/5RvDR5hcbsJ3ybPPHG6e4W/7FpxAMhl0UchNWU5cM1WHoPZtF0p0oG69O7mU17rVTuBZmAGHPv9NXbdV+yj1f2/5ww2AImR+QxzASqfAFIUi2mODT3bS33HcSxlJRZClFa59Iw/0i485hevltzhV56+xOFm5DhrRJ60jBQnasakeR/mFZz7+F7cU+VOU38TK25ZFy1dlVuUvrUndDiL/sY+4yKmFWwAGjnwJiQo5RD7/aeGTFSWxLNxpPxk3EJbMe4T2VYoTYDF26/R/fg7rMhTTWkQdxF/aERbMtm3BMi7WrKXscAHTOZOXfpMyf5L+FND8lVttLGk73vmXF8N99pfokAX55k5z9P3fNP/8gxe4v9/8h//4J80zz34ui3sab6jgyBiH5Z/OZC9x6tnH9H8X/9avV/mgVcnd5viJ482Pf/QCJ69tYSvkV8LPL8B3VHJ4Mtcat+eMYz0CIeT/seblOVuhwMlj5K+gRPfY9LXrdO7KlhvqSfi2P2UArVSsC7cJXmHLjRagq5ExJ9giqBBjHXuiTtwG8K4vk3975WW2Ib0CbeZQ2O9qvvKVLyNTH4RPMrcxQctbpIrjfcZ88Pnk/kcMx6K247c3cjEf6NyS3HrPlUY8eOtCh/cg0EblJnx/B1f3mrs/ggDfNrz73I8fnLqAP2D4jA1gd08JQsMeCXoReo9d+T71vcuru/cyWhjUe+s99qI/0qNFtnqtl3rgTQbse3t1n/JqTXc1SieI8J6wYiY0Bb7Tq7nr38TBNxYn3VYdLE4YGulALRwFQdyEFnx6gnk/AL7wbOxBocC0WC68kUu/CLwNQ+pp0XtXyBZ+Ont6ZpsPCKohvcZRgJrlaXFyl46fbBeCztsw9+7jMORh4A+QBy8FPMucCR30mkdg1VrBo/fWcLLAFCcK7MZHwyEEzQ3rOMJxDQIFpBSO+VALDCrWG+8wX3hSGKACBLkbq76ZwNeO/gqeXDW9yCMJqfMQPx8KSBWojVC3flA92zKqexlDMcdhzq06Hkd8Sx8nOipTmIoGnhhpRzFmTH1boHm+xb+Jw8HA4uQEQmTPxwmDWBRG1J/KrQhubV0Ku8OnCpsW+1Dt37w015RGWpyMMhgpoDmIOGFwi9Tyu1eZiIw2TxzcHW/p+3Zv5J29m0zMssrgsc7Qt1MgVP+jr0BrBfMMHyCYU7pbZLW2DdLeoaFtwQFMqV1hsQ32Y+iU6iEwk5TUv5QnHu2l6tKJsW2Kj/Zd4jpR1zzy9XfPN//0Y7bqnOlO1XHbCIqT5G7+S/c/yxH43N2Tr2WBpuVesTiQV1CfPouvmPvrlScf88BP1/7ytWL5yWsAn+c4uIVmlVcSJ04sRpgcqzRZBr3XYBq70SO4t21sttNPduGQbZLB+i5myyMoH6nC1Cm1G8E98gn5dsRM/hJVBC2/UKSb/xPMT4e/37gMNsirnoftv3vvYvgeBZ8Ny3pBmFHoES+aXJzDzjJ5DWTbsUD5H4FYfMxgcPkirC73Dv7iuzHs7G3cDn/zfgD9q9zEGSYnj3uvfKYt2GcWbNWhXXgaiGUp4ZEWTwHSfEz0GOEvpL95Fy9QiShr0X/CSla1VZboV2IVChQVBBouhHogZb/xWdSkc2gd+vPiGJYPFZ1frgrwdwjf/ldfcvfj4mtBhOFHt+oIrrIFF8dL8IqAiJXYLEL8d//55807719ofnPycrNiDGeD+LjQguHs2bPxAzDGyRlmoPJtnO0ws/iGsI7XMxGfwTT78pljzcsvvcBRtTNYhO5CYYrfpfFJeDQWKvg6cbxbh9n1WrbRaBV16fzF5gJ700cglH44rGPx6Xit+9o3ku8Y94scN3zmzJnwSFc4Xa2Ns2Bpy0qnvks2rl/DSQgjrITi0wSnf7i8ZQEAR5+Ytm/H2d9aLE3Ww1NtN2OsMMZKQX7mZJkJtuNZ+E/GC14UfNUEY2ouXpa9u9wykolKAqzXlh8N6E/cR2j/NgEdOddxxK2PE7bqvIRz2PnlHkdczmFVdHQ+TlZhcaL/gcuXXTHntDu26swyQbmB1ajb2PTPUhYn5ePkoooTJis6276lgoUjol1B1kpkSR8nbNWxwX2shRfKGreN5jhilCJanOivwJN0JjlG1ImMjmKN5/Y028gNrDuGp+oM264kG/Bfyh3xYUjajsS5G2zvKRlD4nrZz6W3edYiR7gPzE4/EG45VmmyYiUTMyehWkcxbk6jWPPI1qef2ofibDPbkjiiWj8mZms/BpEaBw0oWKl36t5xTgvOZSirIj8FC35IMYjrk8kSxM8j1H/SDLMik7r6QfVcpR6AeZzwBUnG1o0PKkFq0QT6h30Zxifo8TEWJ7/C4uT/+avvscuEifDKcVNQfngjOCXdA8q/ZP33CysuXPcGDUPqadH7gDCE9+BHPmLMc4yQv/hPHx1al1yLxUmdSqWywQl9YNvn5S+8OHkv2WQRvB6G+XIf+ObXkjB00u5E6zQXKmOtpjyIQmK5ch3zmOm1qzkpZw8HMezBUSmnmaEsVhZHXASifJLi8Y9uBx+z3ZJPGBiARFtglF+cuqu2ufHW1m9yQp6Ps2pocwafIr98+ZXmfRQoTz/3OY7f3Z36BFPGN/gtSVWi6x8kSlf6+hT8Zwqfh/ps1LJSme1//uv/zNbOJ58+EgWIWwTtRxtQyjo+uu3J7VDxIUg/VQ5NPbCNxlPSVsLLJ1CYZPsPeMnbrDPrQmtw7xfYTjgHr/OUH/1VqSQ2b8ukyCU9lMNU2B87dqz56U9/0vzm+LHmwN69zTe+8SfN4SOHUcq4LV+FuXzEMV2iQTYyULHslv1YuiTUPAGdnKtN5L2d2xil3rsH8ZBiw6trGhXSe+s9DmN/yqcur+7ey2ZhUO+t99iL/ukeu7y6ey+XhUG9t/axFCe9BPXYi9h9WyIonxLe+zh4HDykM1dFVoWZbvi1A9C/L/F1iaCkSHjv4+Bx8DCA1YXIVGSq6ZMt2D5+NjUbs/G7y7iuarYshWcDiOeg2D5bOvWrd2iEZ1kx/dcX38HqpCxO5nGOZn/KijozxDs4zSz4ToKGdCk8HgTfb0W/4EZBHKgrNCiRmaLRwjBR78Ppl78Lly4K/F35ZQIqTmQ0CtfTrL6pHf6s4Fue5A1c92ooaGhq53acOY7gUtm0fnqs2bdnW7NvLxPCLes5bngNwlkJkrEwMA/SOJF0IqbQ4uRcRm3eNXFRPC6KJVR4XrknFi+lgU14vvFZZk+lBcfuwyL6k61AiNOjfxtXyoFSFCevYXEyp+JEs1FwtL78qtltTfR5a1FxdSQDRnBi4sDq2FEsTqI4gbGrkumOF4wwTzliSZDZMHinrNWezdJrqfo3vMpmGdv4lMV24d5+HYWtYqDQ3FwnVk4G1owtb/btGGsO79vUHD6wvdm5fQZv4qtQcHU0NU9LzvBPPhk4pQ+M3v2kxopSCsKYpMZT4vGggC6exsFCPciJlTRKIEFVEGhGHRu3Jna2f0tth+voSQgRcuoAgFpL0/THHEf8o5ebk2dZkWYF+m4EDwf74SVu1f+H9R/hJEDnOZaUEzOgz+QkihPK50KfdWIcFUYEBWVR9DJZj0TQpy7vYt7Rvw2OMFi+byRD1U2+gZc+k6TvnEeAIoBMIcjs2ra+2bNnS45+9mjO1aPso2fwHsOyI47IMrgr1EDPCDjkQF41wHPsrPRM2YZ4FUEEWGXpvvT5TxVSDBe2/ypXkhbaCbBNWLHWOxYnPHZ0O4+PE50gh9pMCjv6W6cpfYgnIsDi2T6TC3pLbONUTREldVA0E6xXn/6+9/Fbkv5ku7D+CTBV4PfKCl1t505YrP9r8E7NrMdab/wKw7n4niLU22ODT2EXlN9tdVpDenRl/EVhuRDHc6xejUBs8YwSgHSOaXLd2sbY0gt0gzHlLHp2vyIOzReUv+hxD/2N6SfuyYt7/5nXBZdHqYaPGdlZEH+ubtJZ0yZvsuL4f/33bzevvH6iGZva1qxZP9Ncx9fqWfalewywzgvlFyvwU5KjPZHiGcnYxjPZrJ2YYrUeYRbFyYfHXsVChEWN5RvhByi2MM2+hZDs6LGWU3myNRW6aGkyi78M+Z/+NtwKovLY421tvk7GPfZYJ39X4AMnT51EiNVxoAqbWehLaaGTY9kE2y6m8LOxnhNR7tJfj777JlscLzQbplc3f/zF55ovPv8kx3ROxXrFifIdlAZaSIygQFG7lcke9RmaSB/y1XrB9j/PBHEFe/IjY1ifLbUz8Usc6SqpoQf3T9v+rLvbMOc4h/3Vm8AZb45xqs6LKk6wrJ1FoX4TC0R9vLj1xCM3S3EyxWSDrTrgNjWNQkTFCROQkVic4ByW/fsqn9fk2OLLsTZVcaKCXvqPYkXiqrFKEHGIc9hLHueJQ3hO1SFb8lARo4+TMfqezmGxOMERtkoZ/ZroV8D2rtJGa5hYpsTihOOIOQkDqiTv4iC8tpe0eqj+T4PIdgRykVN5yUdrRPJuAFBo0yuoxyhA7lLHbMnRp9v0mtXNnr3bmyMH9zT7987g34ajr7FWs69a5mxf5iHWJbK8Ni/ruuTQBAGTyW3GHcqjjMq/xBE+129T/5ZBXHJZtoyxVdZ8CP/t8UQidmnaWL89/BYBb/HjRQXpQ8aFESe3URR7gwZDxcl3OYp4GsWJW3WMB45t3XR4ea+eU+0gZRQGgBbWf/tlUfltgyBED3v48idJm53wHRM7CwzbquOX/pH0baIiQBlKP0AqSNzOJv2LN4I5mcnn3fZtfXt1+Xf3hPHJMuWZn4pZd8tvBdE6+aJsCu8yAnDmPSkRa9ZVI+AAL9u1bTPWrDuwktvYzGyCb2n1alzGHdujVHQh2cBuIdMc034T1W9cIVw9Zo5FcPBokfTWyVPKnfrVe++9Y/nbtXc3fZftN8BUSb2Cxb05lLf6QrqJpZoKhRwhj1LbcUh+op75/NlzzT/+4z/Ci9dj0fUMFm3IfvBqrdr0PeQ1z1hU4yMlIB/HU3nKx1jKqbRSGTzFthvp77t1oxIj4yo4O3+6gMWLJFAxqyJcOTHyLXg4nqjEvYFV/3WsiTyB7DLWh9JnN764Dh3aH+VxFlTJw0UvKWNZHSNXKEOjwKo5juMxnwlfUv7im6n9tS139Z8QCTK4ksngLQ9LBA3Dex8Hj4OHQonIw5CFzwsB+daP2X5dIihfEt77OHgcPAxyG4YsCaEF5K0fsw1eFASfYRhL568Ii74nsAvr7gZ2z929Ui/124/Rf+7i2tllvvW+dIyq8P637rm7d7nde+/HaJ/ppD7VVfA7wa4aVH3tGhlsCfyMlyZHWv7xntVrmYDIo1Dwa2dx4ladn6M4+ckvXmNFawZ9IoJZMqBxM9FhJ10Ln0ZM+R8efoe10KwHO4nXsETdmwzP4btieiMOsBZew/Ir0MgcqlzE4l3hUK/18XGCcChj8OrgKhzcC18KwXKhy6PAL/rYoU1NroCK1248dcukt85MN/t2bWp27dyEI1jM6JgMGmcFDD1H7im0iJwXsPWXodGJzF9hUtEjMmwGlcJcuots2h9wZUAZeImdQpBjsIF2oWQLoMWQb23bDc7CGl5Spiu/8SW9W3XKOexphE9MRvFxMo/FiWWgtMYCvgNWxRexuZSPAGb7Cgdu1VFx0jmHFcHsewd3rRusIwdfVyk6+MlwQf1bR1X4B7U/J9AMl8ghxagV/nWYKJ67dmxvDh+caZ47vL7ZNYMPDU56GGVClnokRmnAi84VBjZUiAw/BC+ig691Xn+SXTRVmoieNPO5xEbufMsg4j1YSP+qI165DO1+ETroa7QQ/uwNXKlE69mBmHjAew2Lkx/g4+QjFCeuPOuYzb6gSsqIVe/Vpit3S++3lobcL126wIA4kkHOT9Hg09j0g+JAn4uCSev2Jbcu14X9X4hdGVr4vCsAV7iprD2x8K5iag6hZQUmsuubfXs25ci/DWybmlwNbKLZP9wHv/Kuqyg6LiWQduLecC8nek6+bmOG4+RF6MHUaL4V6ITJ+4Ts9/DLPBiy8DKHfvsrHlr5Gttv7teWJk74VJxIKoJqq04UJ7Q/gNtUqotVSoEHLPnYs82tQrjlki7p8bwZk9+W/pWui09ZgL+Q/n4bpAJL22zBr5yG0CoXQ0Ml8kLJ6MoYGdIFc1Sq7UIH18IftMM823aXhm9Zi1K99pcwMRM+MABrNVa8qsf6WjEsv0pn+9YY/VI/Jjn5hbrO5Jpo8g3buKuJ1EZytgIsr/lGgdYWWuyDNuF5zt0ymKxCTFVXYdW1lMRpU7UR2luVv74XXWnKlY3A+IvFhLwRPO3LNxGa/8//9u3mtbdPNRu3H26u3WSh4sIlnGbjyJVOLc4mHcFniScQKMhuZOyYxq/GNSbuKk424ONn7tbZKFmO/uZGcxErEyopStMpLE82TNcpLefPX8h2EBUWozmtAR6OgKwyRPqPoyhdx4R7JcqZG0zuz58/X9uDmOCviuJG+pOGifoEVj7bMAvfyilW69dN4OjvYvPOW/9G+Cp8Qe3KltNtWzY2a/AHlcZD3Qi3nAxSF5wGogPlUppU/5fq9mdJq5DvCqmUr/ZjsG3XNERgHDG+7cJ7XUX/B7d/0pC+y9NWouLkDHulXuRUHdwcxs/Mi798JwooLU5uMMFyG4zbF/UvEh8nWJlcZquOvHUtfsk8ovMG1qwjrMCOQa+rKE5WoMxYi7LkEpMHlfVufTTeLJYp+hxw4mF+4q8lySUVJ+QxzQTGfhyfQipEUJzM4iBYxYynZuhnxW1W+qtxIqRco4IlihO2c91kdXqKeqzxvyhja5OCi/u/sIuO3b3Pf+qrPNIJYvG/lv8nO+qDOnVUXclgdmfuBla8N1GOLMvpZocP720O79+GwnsDJ5KgeMIixTZQ9UOqEmgyvpRy1rqxThlng7C1U5djj6vjmuFHocZ3aWQ0sazr0evfdMPy9/t/m2V7SzsMTwlibRt6TPBBwFyVDdI2HQNVnICZopR8WHr4/RpK1fg4+evvoDTxVB0WrFICfs2HdI/S/i3eJ5Xf70ocVf9iymUfDPYxzYUAAEAASURBVA+r1/pdSH+VDCUzOK+g/cBnIougOPOENJ3DyntULuorw20hVf9VZvNUoWKaTwM/47mDitSNwq1kgeVsw3Zh5u7cNfr1Ck4XQ2GCP8E9u+Fnm9c3E8jiI2ir478kBYP+ll8FHvzJMdxLq3v7hTxJvKXkHVYIGX4KJPfI/9LKtiMafuLdqUcU4cpntOmPcLz9zntHOR5+OwrpNekTKk60NPG439v08ZUoGtbAoz1ly7pQkeEWwpuU5dj7x5q/+Zv/F160vvnaV/8D1jLb2GY5FWu1u/QzraozF4CezmU8LS3b+lCc2K7kJVqc6Bg825lCN+nPggR/KkQuIBdab467a7F2ifUheV1H8XKOseLs6ZOxeNuENb+nsI2jVNdSW76khbByg4RSOVYKnBqbbbBFUX7TrqC340DvCv37/Q8KlBJVmkpwr7b9tTJqW/P1qf3twrp7pUrVkLruCxIseOnH6D93kQp+tYW2rrtP7b1L1d0N7p67+6Ikvdd+jP5zF+XTwaeNyjq82kxz6wPoh/fiJc2j/wxy5iEywCCLPpxBLL72w43c/zZI/NAPSU2RLbXwo/Twvcu3IpAfH9O2Cl6YcWq3FVBprGHORPJfMiSVHdpOtcA5LKfqzHGCivHVNOZUnWgJhe9M4ZPhp/0LB0QCrwMJeqLlgJgOZJyUpYrQFStIiXMb7MPi8tvvI3TaEclHfG/FOSxaW7rpNE6KMkG3y8IIjZ+YINDB910nZ8vlhJQ1jB/kM6ghKAjT8kp3GYH3CuKbM9l8IxfymL/9MUx6JQ701qHR3tbs3bsNoXMNk0EsGuDn7vUEAKI0d+lK2uAtVoKHkeg13edShsg6uZwwGpdwJwzmMaCMWfABVu4DV+VpwrI4WUR/YpADseq3ezNlP0j6CFt12bc8Veet8nHCpnrQMS0MkRjdyqqJtYqwTjNxII40NHw2p+rg46TdqiPNyseJA5CDrQI35Yd5d+i3hB+8+uCqS/61dFNT76VPGSvXduXpDQp6dxA6zXcVg+N69oLv27MDQX8XW6U2cVrOGJN24JFb2oQzXy4H8AHATPqLyhbLeJYr/Y8XqbxcOqRR+k7bactvXjbJUmoqdBqbK5XJ3TITfwhMKhlfPIjkoC3tTEZZS9Aqy3YHuNffPdv844/YqnP2OrMt9p4i7GeAJ5MoaWzvIixBev2vVpLJExialruSUKfqUK60PWDz32f7X6dwIBPwoP8QbH8TsZTY/InPekLbPsVVsMRhi0L1IU4w4tVVhjkmhXcxmZ3EN8I0qx6uUO7cuaXZZh9B0QhEI+auEOW7zt009yys/UbbV7i2/yB0CD/ETh83uUSz7fGcZLRR0weHZMNXvrW/KZtwvepDe68Xv6TEPGhJZUpxRJql3ZUApdAXixMED+P7zagpCXVqTqlSP5oUfIr/JDI/KQQf6nvutq209yH8NLU2jvRPgtzaZ7+14R38fv2HNIFfRTCVYeFrCnf0Gx1ru7Vj1K0jEhFcLb+wcmKMjySyzIYJMk0i9A8ChFB7CHJyqcDnVnUBRXiIxQGwkjg/PpMXSOtociwr1jiVU8hDudbRX4jyPttFFD2+aqaXfgNs2gT/uchL+lvvYBHeyGNQzleeA9e4yd2HwqeKZQ4Jal9NkDJUqJ/aL4Y7ObROY2ZmXfLa9j8wRXFxp/mrv/2n5v0POSmFE11OX7iOkz4s4OA3rszqbM9SRjlEuukN00yeN7LSONucOvFRM7N+uvkcx7muXIn1A7B+/C9vNycvXm1WYo0wyUR9HdtzluFo8ty5cyhNLkE3FBYcue6igkdSaoHpSqEnKMzg/8Qh5CxO/FwtlCdYf9JIlFVWrmJ71DhtYOOGKUzKt2SCvAaGeWuO02HuYhGB9ckBJiFTCMrWc0YKeSjtRdaREM5gXs7e+M5/jnKuPN/6KPNteZbUcYxoL4kbsuaHF1qQefLPuqv2TyReHqr9t/UvPm7ViY+T7jjinKrzNjwNHydsN5llYjJwDot/AVSHWMKpEGE7JPA8qUIfJ54c5NYlHSh+zGTHyWDFuxzrCo93d9tJLE4o/2i7pcfmtBHfJZex8HG1doo6s0x1qg7Hh7K67OQkzmGpo6s4cFRxMoEDRwiVyZWWLhPZqoMDfFamp9d6qk4qD1oBwP6qws5yVzPkOR2B17qnn0HPrIabisq3LxWX8p0+ZCXafxk/tQSINSYKTX1DjI3ggwor2v0cC//0U3uavbu2s/UYp5Zux6JWVK7YkLQ4raNZpb64hYwpM48C4kd8W5x5Ekf9L2QFHIs+L7+CTdJ92vrvwzfPgtzB7+6DD22ECn9s8AdlkB7QlrqK9bdgoVfGLeVNSuupOi+/fhofJ99hfG9P1bHNU5dy49A2lAFHEUy98yzKXbj137b/37b8XbZ9+qc9pSTIcvxzsUXZoGRcRBLnFZRLHuNCk0d7e+LLHON/5Bnialko0so0g4aR8lgMoXa9nGfCnWirHNCCOBYg+v4wfq4ai90aOMcc4M78LLwPBTQnI+7eNYOVyXba6iaORefYXRJF3GcsiQzPe2QECej4Aj45TYp8zV5ZxL6a8vMuG7MO/U2M0N9XP4TwuSn5udDinMnjgI8e/aB5992jzXb8gKxbBx/GmvL2rfkcUa5vK4+G19G02/HsrirRpZcLTFpMv/Lqq803//Zv8DW1ofnTP/uz5sABT1pDWYmCW4fUt5lfuFC4ku2BdziW+Apj+SzOrFVOjzNWjI9rDYdMDG9X/jB/+79KdLfP6I9G/jOORaLKGuXDFQhv11D8vnf0ff7exerlFH7nppovfJGTgZ58GnwniUOxw0eghuM6dIEdwELkIWzLgUw5FEKS+T20pFVKvo6G/ceQNR+tGahovvwR1M1/S+ggwDy79m8eXb0kvPeeb4/+089G+IVVL99/J/B7ipOFRBgUsAsmoF9QCV8V1ZG2l6J7zN3OGlbT5cS9i9ALWvR4TwwCHhd8RVAFTmEsaGi+WZN8U0gdVirhRNZRZb5Q+MKviyMrcOWOji2z4+3MxWucqPNW81O8/q/C4kQfJ+kDtnq0t8xgWvgKOoUJD+0lYyHvFgG/ZvwMfN78D471veiryWcmD+aQBCSG0Yr4PfQnbKnyy3YdZONHQfj86czuhsI/3W0KkzY4KPlDhSgkig5qyctcuMVb4QDmJNO/rQKEfHSolIk8sO9Q9hJwVSNowkx6UPW4LZm45uKrliPITMyXH5ODHJ26dSt71VczEaBc5GdZJaWDhtdyhP17y5/sP6H8pq88zCc0xQJkyfoHdloN0R+V/uZtu3Ax9Zvf9Ri1s82tO27V0XlgtauIW2kkMmLKKIejjHqAzxGSDBgKBzewS3//Nx9FK+0gaTUzYvGj0GYVgT10UUC2/s2dbHpX0U9jTlMBnu+0YfI3fuRxB29wWU6+yzAlHmGivgozzG0bp5rnn34C569HWGlAgWXzJf2KTHJaKOJj5Txi+7MaQn/bV16GKFvN1f9sl+CM8HBP+xcRC2M+FF2FgN7baX0pn9+iXIujROqCMUfhIccRt1t1GDUpzEisBmrt1uxsn9DFdhrghAkHLP1z0nvl8gVo0FqcAH85A+QdVhqM7iSqLvkDT/7Rf2LCSiGSFXdxVgix/uwL1HzKqBDF1CLtR/8lK1QepF5c1V2OBdbW5uC+mThmW82EzDwiuNiWWny7FhAFIvCX6v/i2Ke/aHaX2XT0N+xR238/L3ESvsJiKQzEF5qDru3JajvLaram9dIeySvFEH5oDr0JJI13eRAtnU8L+39FNaF5LAXf1Iuviid/E6fh1wXlJ2GVv4Nf7UJlm+G1H5r6ov5dDXcV3RV1ryh5KLd9s9umZHgpqGh6dlnHkLRv2w0B4k9dztturVMIJX76DXLV2bSIai1O0oUWh/Amq3RryCSWYL2ikB1lA0+xL8oMKRGa8rok/wNWUsGDHrX+I6ABr66C3L3dc1dv64AQwbm++hjFDbS5hbD6d9//SXPq3G1OJzvRXJmlL9mfTEKptFSQH46yEjjJmLGZU248AeHkyY+aj5loHz6wq/nCcwdpM57wsrL5zvdewoFk06zZjB8gnPyNsN3k/NlLzcVz50OjEbZZwUai9NbJqZMN/XLMbMGnCRP5C5yec+H86QqnnsN2qEhEaCb8V5tJHDRvWT/OMZVjrG5ONLt2b242Yg22erXm2FvIA6eLVJC0l7+krLZayZSLj9aXzL2tyGH7q5q6t/0TsQYGc+qSVfpe/VfAABA4ADB901TDS1y6/m+2t1klPsN2updeeY3xfao5/tGF5hcvtRYnyAyzTLTG2MfvpMLjN92qMzhVh9Y1zVaD2dbpqwo9fcF41KYTAq1HzuFP5i6WPZ5SlqM5seZZxQTIfuTRwuLoFhy3Zn2MImsaeoqX1rHW0TiK5OtXmejgN2TTpi0oZbA4YSVaRYyrwx5NPAJMT8HwxJ4bbOtRceL4mr4ZqoVDCSr8PyIP4fJj2yCdIM/2vlh5QpMcI4sDTU/Py2I7K945ERCliQ5fMXOCxzHZon2zCJ3txk8/uac5cnh3s2MbzoLZnuU1Hx8SKE6Qm6yHor/jGIX0P0gVfyr+NpT/8hGESZUOUfyx1wKSWdf/P039Pxg+yILCQvgLoD8e+O34r/JDcC5G4sk09FHva53JK13s+pihVB8n/+2vvouPE+qYyTFVRzq+p76N3L+k8rD9pzj8dO1fgPenP/kkQZ/+n1x+4ZnMxQSVA44B+llSEaE8torFA10YXdfainp1gcaJ+BzytWXVCb7KO3lItvnYagb1X/DNPxef9JmnvKc1RFkrM4Yh+2fBh/Zs413hJB2LKDzmZLFyJ1vkn8Iybs+ubfiKQsGALi6yBpiz7FHlBohDVOqfKhGNOK62PfI/18PwHyoobd2Gz+XNGVMU/BT43LlLzcu/eqU5isXJk08/kxNx9JN2CQW4Fif645vgNJxxeC4FRZ7CShDE9DEyx4llWgD+7Gc/b777vf/BiVWbmv/9z/8CZdBh+ijzD+iqVWFZaul77U7mQFfgQdJ3DdYtk2zlVIGiEl3fJWNYibj10EUT+YunhblNaBxl+BoUxhLFcIn0wfH3m1dfebU5fvx4thQ+89RTzR89/3yzc9duZMiiX0iYFlHld4ZZ7c+BkLZy3/4fEPfpf1aA9QBtSd9SNglK0cVje/mtq64EEZB21n3nvfiPJTJuL0X3mHv1pAV59eO2+S2+dVkMwgn4Q4I/UJzcUzw5S0eZAfbDh3viDz/x1BW7u7cfu9fuPohnCjtKj7yfIXw7th3Af1a51wB+h0KYEM1BGqQTV6kqJrFhcBFprU3zCLegcSv08nrm/HWUJm82P/3Fm2zVKeew8EN4u3GZzHGqjnkhKufmRNd8zK2yg0kAu4NvcP8S33w32T2XgWJn90uO9UtwrCYSbBz/xIf/wIdV0HEdnAllhcTJot7nb7ASFB8nbNWRMZjG2lIjTrTgGDL4hXx0OGgetxn81cgaV3P1bIGAbvM48VvmRBth1jLUlBCaxOyQPYYjbANh1fwLz+xDk7yaFUD2j2O61rVHJwoyQdP6z1IsvIId4fcv/6ev/4LU0V/gDwPfVCpIblMt3+Q4Yk/VuY0VUhQnmh44GQbf+O9wwHTQY7CxvqBYypktPMS9HouTj6AJXrwRpkMPmLLMz/EpKw/AcmLV1ZX1EVzbZmF95StCZo4ZhvY30Y4bOoqVz0rqloqiilg5RAO+Fed0B/dtaZ4+tLs5uGcnKw1McsWN1RybMCgLqr1+j/RPQxS+7YKRO/3KYYfy2DHFUd4CbTwxqBQn7HVtLU70cUJjpdw4UaVlps0ZPxeqr7aMxS/45ZurAjoaK4sTBkkgx4+IHR5LkQgpETH4Qv+SVYhKUCVDq9qJokKB9RcLHwbu8AnycgXNle47MW2/ArnnEfLHELQ3cOzzRsxltzdbNmE22in2ECIV/tP+ASL2KT8PEWIA/rtu/yFb4PNE+YVvuy7FiQjqcFP+w2e+n+Oo05s4c6z90cVPpLn/cpHe/u+rtW0Jh5chbfnb8HwNfGIFvunMq82D/GwzuWwziQPtDOK1bsb3kqZ9+BVqM1GgcWVwBZNvj4bWQeVKhDW3GKTdufoG07SPJg/6r3xU/lvQFZwBTrtZ4QkZdiwyLvhOEvjGd+9aHgVV8lIxIE2dpOnDZJTVefnwKLA9WSN5i58NLGUTuuUpqD6n/PVKHN6hbzf+GNyV3tIG9wH9Delf5rWI/m3cgBuAHTwkv47+dpsgSZ8IfHEzKsK+yufb8Lqf/ert5qPTN5ufvPR2c/biHGEqTvQNYD/S0kZngAjOONVTuXzxwmWUG+eh053myD6OSX9mD3FcQbzT/PMPX2ebDttztm9iMj/bXLqA8H2VffzkIaX1KeW4GCU1dFnDBHsdVitufzqF/5MrnKi1CkWVgvFtrcBSALZHMTNez0R9epwJyfw1xtFLOYb3S194nj31+8HPo2dR4FJ34h2lGcVMFTiw0tOlpOVOGEQonvEp6E9uZPZY6p/mA71RbqI4efFXQ+ew+jjRV0xZnHDcL0qlVVgxuk0mihMme5dQZqhY9FSdm9AqjhtRYHgSzjUtTuC9sThh65UKqkksU1z9HZyqg+LE0yxsfzP4p9GZ4zUsTnTia11FcYKyZpztVtevs+LLJGoz8fSLouJkAnN5T9vT+qS26qg4mSX/65jSr6fVtvTv+ojkD+n4RXaxjbpl6yZKn5rUEoG27STL8bniytNd/CgZxePgdaJ+F+f2y5ezpRIZcGZmbfPs0/txALml2btnKxNTats+Ak9cQT3L8yPfOQ5T6X4x7+565P5nQjL4vfX/xw0fnhj6UKj8s27QbMmZY50GzbQ4lX96HHEUJ3/JqTqjKE6W16k60roUJ9ab1G05EMRW/ir6+9KnvAWRlO33ez/l6yPJ36SA65OjODsq6ngVvp0xkYZIf1GB5tHq1/RxAky36ril9ha8Q+ZI00t7E7NYnMRc2Z5qy/GvihHlK8EZdxloHZf8qpNmeeOI8j/mDfPI/cuQ/djdGQuTp44cwK/gDvoSJ1vB6xxr7CvOEWjtuQN6QCoXixzGQgd4m83Y67dpfxkXQRY0m6PHjjUv/OjHKE6OcerMV5uDh45kvLqE81y3t3iMsBYnVk9ZyYMJSCxn8c9FqY8/vtz8/Be/aL7zne9gBbiz+Yv/8hecaHZI8oWeN5GxJ9m2Ixu+cOFitmGqMF2tg2rGVucyq7FEvHTxSrbwrJ1miw+WKvpzuowyV2vHyQksx+BF+kW5jRWe/MATx059dJJ05zIuu939wN59OKPdwBjiiVjQSj5A+/Ny0dn6iZyfkPrp2p/4VszeR/NYPP762bj3Gf8F0rb+FgCRw4/6+Q6fg2cwG4YNn8RIrLt7+6V77e697/fkZyP/A4aPjN7WzLDUveJQdD+3BRiUtxe3e7zn26KARa9JlrAlPvSDPgv45mnFprJaYDYa370G8Kn7IQOVFsaSGxDTwc2YCeOJTpGvdE6F/TMckfgzBImf/PyN+Dhhvb46Q1aNVCb4Rx5yvLbR3he+kMRFWMAVbNIEv3oe4OyDTAo0M+b6bnph+Ti42rAuL74qyqN2Tv7Og+V8ak3jGA+8p+O5Xu0P/yljh5PPmZPKwNW44vg2R8PRSWtPNqppBrAMRDAEzcGdJI6yEqcvqVs3WV1GqBxBYbKN4xcPH9zdHNq7lck6WnUZdA+WWKcchIlIqjJPVey2uJ+q/NLIy9+CwfNgAOU5dV3wE4dIDwvfYdGUKuHLxwlbdbA4yXHEVFiJXEx+oDlDX+A79qWtWueW19N3aC83rs9hpn4KU2jM+kIHHeFqAgwNGKQyCMPx3XkhfubXla0bwcQ/z7STWU3bSadySssd953fvY1DYIS8dazcHTm0q/n8c080Tzyxk0kAudkFgOWKtgAqL+lv7dS3T9P+OhzNw6z9eWT62yj9L46xNvGhJiBq2WNxYjuFbipObFuvHz3HccS/wscJDojjHJYBl2RMldJnUy5xoaw5HQgEBWMjEz8p7lYdT5CIc1jqzS4IMIQd6hYhupCqmxhVn5Fn8KWdMem93b5h+VcRpoVWBGgnYqwAzTH4Lkcpso4Tiw7t380JHHtwxjtFbC5wi5gQBZzv5BucfaT8vARlw5Pg3v7/WOhv3gKgkJ/EfxQR4zOJ+J1zWEmhYHQB59o3sDhRiRInwsSx/1mxZTXEo2ESkwf7ljS3vedhSfh+JFF75a0NMtRHf80rsZI/dc6boDv+a3cMKvlOEu/e+FH5I0oKvgru+lxQaexRgkQDVfICSe8KgskrEKEGE2/HkbJAW04aVkZZxS74tDWAljNJnimsHFsCZXwmv1WYLbjCP4oAN4aVy4ir11Uobpaqf/1u61/gC+EPabb4gzsFdbCpk9r0udQvuMP/LI7OYd/94EzzwclrzQv/+mv67Y2G04f5Xsqj1W5d0xHfeDmydvvGWU7FmceMewTz9h0zm5rtM5ygc9uTsMabD0548s6G5s7onebcxfPNuTMX4XFMvnUWCB+1jYqLK4+r0CJv3baNyfNoc/bM2WwVkQfYEFXrLHeSjIXeSt537JzhpLEDzdVLZ2jUnO6C8nkXq7UHsXiZmEABxEQI6/cQxrTVtnknu5ILCCW7rv2FTjQYyBE65D6o32r2eZXQS7Z/wgfXp69/YcTHCX20U5wcY6vOL7tTdZgczHJKjT51VGDpk2QldF/LZO/y5c7HCVt1UD7MsmKrFYmnTOjjJMpnrFEuw09VKOlHwNXi627pYcJi3I+xOLH/lY+TUpxkqw5UieIEJcw421j1b6LJvj5OnNBoPTvJVh0Xf/RxorLGOtZs30UDLWIyhSbztFW6T7ZeRaaT5vYnnHBTb/K2KDzDCKgh5CYXl5zwxH8UcTwdh4rH18J14F1gwomFydZ1bMtym6tWJnua9SxAWI9alCYd+TumZ/sVfcB6zCo7cX4b+eP33f8fO3xII93omdCYcZZxYYUmQTyHp9JvVQDYK/Vx8ss3Tzb/vbM4iY8TY8o/5an2sOoPZJAcU//SvJURrOaO/sbxWtz/kiYf+HnE/meyKot+aco/nXjFKT/fVNy6mHMdxQmIoFBENnZLCH3DuYjHMjt2eOkcWwOcNEfKZ7stP1G0bpHmCk/hyy0stKwbt8S7EObxwm4zcRzW2euRgzuZ2LMtB787E+M1xpYMBRyyCt8SLtVgc40ftlSMFAW21WO34WMLOvDBLnCtKx4rrplwtUEJ79Nffy72C8jTvPXOe80//MM/cqT8e83Xvvb15gCKE63VMSfCqTenrLGoaEOwb3upwHCcnGcxUKXJDfjJe0ePNt//7veaXXt2N//rn/0v8XFiXDClTaC4Io1+kjw9xzoZRxmzBkW8cqQWZ1LDfzqsVkkjLzt//hwLxrfgWxPZTqVi9403XmvOXTjb7Ni+vdmzZ1e2EXri2FosUSaRuaWL9eH4r4NZLV6USwmKMsvnbHcCrlQjuGjEQ+jbvkPmhCfCI7a/yBCk9xrQ3+dWVlkc7nv/6qdJ+KKARa/DKEt86Af9ocIfWJwUEfooG9J/7z9X7AWf26AuTcXmV45jTfcva9uGcc+1GEb/vf/cJlwi6GHgR1hX80Z6sShmwktaYdvhU7gWQCLVcx5lcfZ4mKoNzn8OqGEjPOrn5DSrMT97ia06P3uzGZvcjFNQt+oQ07FVo185Wy6HasLN76HgFwZiExS9V1DwcACxQ9dAZZ4FM1knBoMJLws7SsHPHmlm29UBLR8mrQgY7hUmBtYf6zGMKN241Sp8ilSMkWerNJMh8ldD6mr+XVZOLfR8bFfJhW/qjsYRqm5cvYTAovB6h33f+M3YPcM+QyaDW3GShLOpFcStiYSMGQDgoBmigoTAZKrZx84XL3F0qBShlD91RCjwTB1a+MW8PnX9V8ErvwA151wPgi9WTohg7c3f/D1bdd45i/DZO1UnE0IVJ7YGp8AKA0ChnHcQ2vPoigoC/fUbKE6On2JiNYlgam3VPlUHMk02o7yiPUbgIjcr1EFfx3dSIhMznl2l9px7j+TVQSEL1M3F82cwZz+X/fhPPbGveZ4tOUeObOcIz/a4Z1Dy2M9VrGRbGv2hOMjE6VXo/numf9tOKCjEoI9JT9uhRLD/5Ts0osyOs9LoNRQn/+RWnTMqThgYEfYVYj2Czzr1f7U/crGRp3b4dQTPc1mc6CRsAude9r9M/KgD27/7kyNQiAMV6SqC7c/VJfOW/6iQiX8JPrliJBin4K783HG7262rzczGaYTtA9mSs3kzx5mii614Kh9Z4dXSSwEl9eCdzJhsCjbllxcY1sL3MaVpy1RlNI6x+Ji/IFJp8iUp8qX7TMRcle7h69++nT/gL2N1Taeq2XpCFudUnLBCHLyjzBUfYXtJN8vRvot/ytbSNPgbu+Kk/5s8+Lf9P2WuHNtPC8r4yeXvwydrMwEHSiTbpB6sc0x2OeVDocqtHUKT91lmTxyRDzkptA9FiULamGjDg609J3hprpaV9lj485t2Y9syvhZ+OJNDyTKOQ0wWrYABEGMbT8ULTcDURaO2/L9j/ifug8tCtfQfhOWhyqh2ORMBFE+DeKRxRVm6eKLCK28ea97/4Grz8mu/aU6ew0cFq8r2R4x0mmlW/zbiZPUaAvJFLBd0Zh0lxSjCNGlXQsv52xc5YvgDrOiONOu3PNksg5e+85s3OC2BrR5YCy2z30DflbZNaDXHKuUIvmK2b5/JhPsUJzKc/uh0SjXJdg8n/svgpddvXGCL6Vjz9a9/hQny1uYyVi53795gfMNhM0eDr0FhUuOPfXa+GUe5NadZPPUYbhOeYyMxHkSxjUs6y2+d+UxcP+XH9/Z61P5n+mSduuAlAM3FnhPg5NzC7H2zjedUnbNXmpf+bWhx8hLOYedQkMxqIYLifTX0WkWFfHyNYzuhqceBXuE4Yn2zuEKr1cYssoUrxFp/aBWiknEt8S6yVUc/Mm5JUKF8QwULSg5Pp9KCRew2cWTo5StDixPDrl5jWxA+CfRroO8Tjxqe2bI524Vusj3HU3VcZNCS1u1BOmF0cjXLAtFarIMsu73H8ocC8m7ok/GTcVl6aZLvpGo5NonKUvZn607FyTx1uZzyMQzTZ2/CCzwq9BZ5r2p2s0Xrmaf2N08d2RlfbU40pbn/5rGuc1LoEcmMpuBoBrQJzcqCkJGrbgq7fK4K5NGr6s10f3jyj9g/Vv4T2pFlEZGsoXkWKLC8ZcUp/FRZCJ5xHYXDL99wq873OFUHB8A45LdbOY5EGUId2+Kti47Gbe3nnY8LvvFibG/30F/peMH4C6CMP4nf1hBw0peTCdnIY1r4snn5vuNgQhVS+JbTJXnWKsv2psXJcqxQaqtO5WdbVP4fweorY1DypC2QV2Q/te2EuTTjIpvPt2n7K5C/VyPP3WZb4U38cqxfvxbfdfuazz23vzl0cAtyZkmkFhaj1/gJofnDJ2nzKKfMKn8qrsBfWadIY2EohQt4jmFVIr7zLC/jZrqHnX+VZV6NAcc+ON786Ic/bt566+3m+ef/qNmz5wBZljsBD00Iz4a/zLMVaYS5RGjFWHuNLXsXLpxpVrOFRku2H/3ohWYrbgC++KUvwa9QlAc1F63YYghvuc62eMdvFbHyn1KWEon/jq321wm2A51jW+eJD0+gWF+R/FR+fPCbo82LL/6iefudNxPvT//0T5v/+LWvwoOUMRnf7eNUVOQ/sgtwKBIYAFA2VR4oIgmw2oTAfcuPBGyvtBfa36fp/5VNwejyWwhg8TdiLRHUBdYnfrvxa5gpYYTTdu69FmfYf+8/tymXCPpdwY/ixIY8nEiLjVdbsCBXGC760tKtvlWa/nOF5DcJ+VmSWNLxdwwfeNXdxQn4cCsnRBVGw0vR/UBTNK7MOczAgpim7uai056MbMTzm6aB8+Slc9ifvYjFiVt1Vm9CFsQMy0ZEh9HuwHO7qwHBZEj7ifAZDIr5tGQEnBBlUYW3+SrSBHmRylPeW9xCfz6L/iDcYrbwtQ6pzmrOXPDs25ixXmO/HiySfcSeqqMwW6jLpN2TrYWN+QnfjzIaGVYWyl21h3YRLBDo0wRgYNcRaps7N1gBWt7s3b21+dzTB5qdOxQsOfOAiYSTjtUjHPPFICKcIf0pswMBdSY5FWyrzEP4hhMomsNyJoAQ/3flbyM9Wv2Tvs1fOJoqPgx8Y7mNS/3RN/+erTpvoziZZ/uFR+MxuCg4U1hwo13QhkTUIdjSza9QGKgwJ8I3ZuebYx+cxo/pBBY7IgONGFRdjXbAzUkltmnonboJjuTvZA4cahLPd2DpCd2KnmNV7NqV81hN3MGaYUfzxc8/g0XDbqx+JhkMdDloXqYxPhN60lr+EX5ipoxQaVsU798r/e2n4BVyKiKksqWRf7ZCcbTtqOGn5uiPr6HE+sELWpzguBDzS6R3mh11gpIq9Wv/T/9zOxQ5mZVBEZihPfR1q44nSOhc0uv/J++9vv68rvvOQ/TeKwEQjSQgEKySJUok1ezIdmJlVjLxxFf2ZOYqU64mKxeZ/yBr7uZm1sxayayVSTITx5ZlWyWWSYpFEnuRxN5JECBAACSIDrwA5vP57uf5lRcvQBIiCVg6wPt7nuf0XU7bZ599aoEMzUmXq6SpVI5dOA6alkwIlYFyht9bbdwBncnV2qqWKgjxbP9MwllLtOs2Xt22b/McPFdwL5qdSbawXQXda5IFbWywDrSdC+hQzvYd+FP5FEnZHRA8Pgv+D8K68qtewo/Aj3+WPx0bCNqcUfOJzzIOi7oskZjoAhP0kK5pbfB5tQwpqo8w9z7mW64gNJE+/E0qn4KDFoOr/zWd8XyajjwJTP/flZ+QAFDlVPlJkHJsW/KVwkQX8rmOmN1Cd6UsX37TAHQZ9uz6Utmg6xutomfNhcmu1v6vyrdv1xECjqo/RfOcGews8nYB6BWRNdGy7iTN4kucWj/h4ul/Xj9r+lt0yheESW7y+O+4IUlDa39TYT2qT3Tn8S++fx+36rzb9h3kBoUTM1FZd+cPG0McxZiL4GIau7H72f177733whkavp4zZ347xQL5rDYtDu9rB/e/xqT7tzAyupObec61F954gUn+Wa5T57jdBDSA/TyscwYBrIvsVVy96S7h7j1vc0RnX+x4zUFQcIYJ+gR/LGM4ksLOJMPWDds3t1tvvB77GssRBMxBw4SDfyyCxbvwMS+nZRb/nTx5FA0hBC8u4EImkGU/Bs/V0CfB8AsKyEF2773kT4kaKMnP/P00nzz4vRD/d5E+bvs3/wkWDe9gh+jJp59lLOBWHTROHn/iBU6YYRxWjRN2rueilTPTozosQNT0WYQW0KEPWPjRPhZhYPwkx6I8guO1wLktJ0d1WPwQ5k6tV60uQkPkFMKH4+z8quWjkMUrWa36CmwTHKbf1caJR3XsT4+g4RVNEuwbKGxR+LoKI7J1qw47wQhHPGKj/aQ6qkN90TY5xTxHg7VuBoRXbX/QQCzGthKLckZXvp3DSBvaF2NgHalhyKCPioxDXCPoPsuc6Sw3kKh1sgmByS23bGs37NyQ2wBneRwCwnr9sIDMhO7RVoHbilbQMPSD/jDEOSeZ8oNA41I/vaDDR51/WE4SkuZytP9PvHzwL/z2CGLFPsJxFq8UJY+JOzcxj548y606ezAO+4N21SyMw05nzmU82mt6UtrZh86/PyL+0xathOjOg9+P0f6c93rkz2Mh/juH0PAsAmRvSPTIh1d2m/ki5hlXIezwunKqH0GB8xHnNTMYS+1jqnz5mCQyjXyrZ6AWcueLNfc7e+porr3ejOHXG3dsRytuLUZTOWrifBI8ilmz8SdazQr8tX1neWw0KSySr4pPfZKC76SxTDsyxyH/x19PXnh81P4n89qkmYZGmn3PL9rzzzzbNm7ain2QLRFITFdjmGxjUsAXyrRYhbD2E8eOa5D7HMKhpcDR2j333gucK9utt9zCPAuBE7h3I+PoEa4cpj9S88+r0+ej3S3utAnj+kNNOjFyHEHThPHpd85wO9aKFStyPHQP9rQefvhn7amnn44B6ltvvbl97RtfQ6iyBpzSj4C3GOUXDaBGoWl17HzwquA0G5JiifH+U2//IiOOsuKsGM7PELajv5/64xLU/Y761PvIb7LkRxincJPH/4pyZZcfwclk5ARP1n5qOKcAvZA4SBJEXTj9ecF6jBDn0y7fhY6u5h8AmQaGHy0jXUTC8e8rkleR4dBlWn79hONtWOlUMtsjPZ0TLB/DaT9jIvHwIwhOvFUnGickodF5RarXEScLB+MPK59k1XFVvVMTK5AGZYg1GYZV5agjAHZotcoD179W+fx25ZdGDLsqzgrizjKhOBmNEw1VLeUWgaiTE6bWQmNB7xJG4YlSbJ3YcHcpCz9wMpMOX6msVT1D56WBNBeGs1kgbt6wpm3btpFrbRe3ZUuQ6LIIcOGkrROr5BnOCJsoSvW44N1CCIxaK+XW9YtTw98RWORNCf+l0b/PLpUi3wvjf7R84RFXiEXad7xV50WMCXId8RmNwzpBZgAL3okYNW3iloFE0jHoKHqyrHMITk5E42QPp0rm0wnrDz7Ama+ZALBT6q6E2wrSQ9TVjpkQOxDqS3wGvnb6KDTGbgbHcrxW7tabd7Qv3LSZm1kwlIhKw0wyVf08RlGhsVS2oztt/uTr4K49lDpeULjo+W8U/o49SE36/MorvqdyA17+8Pb3IfgPsF0c6i3WMyDZGKQVdXeSlWiyOfB5q87f3s+tOvuOtXNqnDCwRXAS7FVlhUwuz/hmvckgt/YEIWe5du69qJlrxd3MnVx4a5G7Emq8DCYU5NnDL+gEUEVoweRYHDvAn2Tnx3O4Xtm5edM62gfGXzct5/aFuZRRu2MhL/Ns25U9ziniq6mSRbKw6syfdydTATiew59PBf9mL3KFqwroCTYouIffSW+0b+R5VpJn1I4QVZCtbtVR44S86HuKs1wa6RPCFfCU5T99dJY8LJ+SrEtXHYN0VX6mzJU/5fd96ZD/iJjMkoTXysS0ckKEjyI/5ctl+pk5oTRyJ1dAx6062jjh+AwLb0uWF5wA2l5dLMlrTuzcopamE0yYYlQYms1AAF2ZVr5ET77asnFXahY7aR41ULuPdWnysf3Ldshb0m/I7znzbt06179+lvQv/FUF+vIH6NXDSgezPBBYyDvnbL9dpDpmB06B7xQCxf/r332Ho7Avt+MT2PaZjW0KtElcFC/FdoaUfffdfajocyUl45eaItkFZQVxGqHJQvq1hbPROjiyj0n39W3PO+faeyzMT3HDDR0qwkhowp8TW1urhkdXLF8GPaa1AwhiDh7AsDd1EP/nWNyr5r5g3qx2aP9eFvLvcDxnU/vd376z7eCWlIXsbMrUtSFR4BT3Qm80VMoeAf27a/Ls2FK2qJCPeOQHHAQ94oSXNGcRo+cI/xk96MIvCYvA5ZnMjFFZ+rxU+ptVNE7QCnv8qefB7YL22tvvNW2cnNPGCUKJE2iczEardDYLjMPeqhONE47gcKuObd6+TRzGdgnzg1m5LQdbamqc0IdGcAJ+FvKuxskJjLdqM2aWghM0U2ynGod1AXUMzZKFXCVsGzymcViENfPQODnKUR2P4KzyVh2O93gVqUd1FFr6rvHaOQhY1DbRxoMGa8WObUaklXCbOQhb5t6yJF/pb9tTi8mbzRx/nbe4+eDC6RxCo2kIeubPPNPWXb2EY8db2vXb0KLdsAQBmoIkOMoOgvjT6QcmEHzPsmHbr0P/s/CdilbyTLQWJSXBo/Ofoi6/l4n+V0T58jjwF6Uca50/ub0D/RhHMv9hTJVmuVUnGidlHLZ11xHbQ9svZGOj68vlq57+wS/ox4vX9Pp++WkpKf9i7c9hWBJ1CUwap5fOfPIrv3XlO+9QWOLi2pspHScc9ZxHazNLI8dUmqM6GCfNUR1nlXATZWUByvt02iCNzKG/6k2HkfryqzM/JuL853pheHU2uFu/bnm76QaEe9vWtZVcla5tIvskWVNU86hxiqdCP/Gb2xfJciZjt7XIWsjIowD6bmI9C+BBsN6p1wj8F8O/o+wpcKNQSaHm62+80Z755fMIRle3pStW5ajdCe2i0X7sV2fkeKx2GksIoiFX672UDeC5HGc9yFXzP8A47DI06b98x+0c71xKX8O4QJ9h2BzsmXjNscdq1HQvWrChxoA9A026mbO85egofx+wCeoRm+IdbbMdR0Cj8NgNzaXc6rZ502bWOAp3OR6l9iJ9xhx3xJg/yyO9prBCrhzX5ukNPNJUvMGlwRZfQeNH4b8uSZ+AlOXMQzeKf+lsxOTfR+joZciHubEkYx/npzwvWI+uoQxeTXYFl4/APEvTEehS9e57BMTOezTUSPV9gcA+V4Iz4A8w0eciG41OqYY5VlLj6cDgBYoo7wsEVuKkHS0/HYflwpRp8CnCMshHAtr448xXN1I+r7FDkXhVbnjOtKQ7y4DngKvGiYbrHsbGyZy5q9opJg4EVO+DDRCFJymfzt6JUM2GzK/ytNR69zlaPgM3ZRlLlyp3TDeW1DD+zNqXUfjjQaFVPs+ufHdIneB1qdLJqN6qgUMnC0uR0sZYkSubVMmlJEOTk/6ujmouzUBI5G6n0nM3i9VsmMBC91V0/DMZzJbSMe/Yfi2L86Xt6jWL63phFwMpmUkF9fXYkFJXz3baUeqn+l1gAuhe+GXlO9/EKYALgpDRdObHwFTXzllIcin4SRAciYQgk7CL0n8S/lPnyq8qh0f8umwIsrgzGRi9+6Ju1VHjZAKNkwkMlTm4uzuitom3pTCLMgvYxcT1zMKBeOgf5Nzla2++jSbTfHAj1ty9kgJ0ryC8V/H3XGQ2q8hGkHqBTCrkbssptH7OfcAxnPkc/9jMTsN17dpN3CePhF0bG9ZHdojQBNw4cJzkT5VGJ/+ZlKBqPQcBl+rRmdwRX+fjsuC/qNmVLy5p4Qyy7hpOQ/sjNJZfobc401jac69wVAeNkzqqg8YJk30XnnVUp4MGHPa7Jn37cxDOwEYrcKKvATdVaHXyvcIlWcqdg+wKuTAKl0vX4iP5mPGRTybeGWiPk5Y2wu0bmxCabN++ietKF7QFaKNolM2mYPnRWEg7tAcDDvyyEKcdiX3/SSNnTwW/6RIzvJY4pPn4/d+l8f9U7U8JyQR1SI3duc+Ehzox8T0wOKpDaHR9A1Z1oeA1iQSvUMmbtK04/CbYT53NKNpS+JYfv+B/cv+XzAaZ9KmTwyBv+aanvyFVvnHJm2rZM3uERmGGV8d6HbETSxfvtiXLlW9ygwHYr2NzSR7/4ikzZiFAAvM3b/tfbwSY7oSLPtGnizAnYF0knoUXi1GA0iMBDrjs/V8m8amQ8BS+fEp96zt0+KmWRxuIYDKB/AS0aseqWP/b//Dddu+Dv2wnudFl0bJ12ChZEeHUdATyXte+/92DyNvgCYSSGuc9zSJXoeR0Jv1brsEg5zVLON//LmPz0vbTn77UDh1vbe4itJ4o6PRx8K2mGO9zEYisW3c12gmz275393IF77ssjLmxZy50QHCyCU2wdQiZT3IcRVsmyxBufuVLt7R1q7hxh4W5dLBPPof9GWrTQWyrwzHDDz8RJ1joaJ1+2rYq8Ts/n3Qj0FFOF2/lfBIzv/3DEPP/tPpf+S2CEzROHuM64nPn5qNxcqg9xnzn7HSNw6JxMnod8Qcc1aFP9bYcryN2Z9yjOqeIdxShxVzw1AtEpNWihcRDk6SO6tSVn8eZg7iQmcUi5ij5CfFKNEnsd91FXqzheuqloVg3akavI169clU7BH28SSc2Tmhb2jhR02U2gpOTGoflbwnaSnYWfZt0rHWM00y4GxbpM2iHaoycQCijhHQWi6UZ9N2n6bvPsECjmbdlGO7evnFlbJhs37Y+WjPTZ5MPu/SOIRSbhZ3am/Z/rNbwp007lttubSy9s8OPHx4GyU+kEX5dUb5+u4/4f5r0vyLK7xBQ2hCO5/Qk0CjzKPoKedSdeul3DG2yx1/QxgkaJ96qMwNNXxDkXOsMuFbcUjgmUweLNJweuz5FfPf4iPhP6pTBm1lSlxJ+mU9lNuX6g3prvNh5neyQxTp+M+A5Nxw/wOCxGouxcYKwzavZ44jsfEY4nIfAxAGjL7OOe9vbGodZKIKTqxD0LUWD7trclrO5XbN2CbyLNqv8Z39rnZ1PkKdaIUr0GLHCfzlqVKCJPRx8Gr4lPs5fqh1QfT+LZrPjEF7lyfuU8F8E/wpq7J89+uyc4K1db2M/5AWOQXKV++LljIccE0SQLe7cdBWB2mvxyJ5CVPtYNQe98leh1LO/eLZ957vfRSi+on3rd38Xjff1aKVgGBwhuusPbR7Nx1aJ0KhBzwN8uFWoc859Ck2TI+0dNBAPcCRT+yRz0PxcuXIFY8Zajoyi3QRyFLrM5TjPTJ72UWcZi1zTCD8zBfDJ+Mx7MAf80rHWPQrO3KyDF/gzwqfV/sdPnKQmwBhq8bTgsQcf5cr7AoEjkc5bfwZa4c/I18fk2eUVH991V1750C4toqo7WueqcX4v4D0S4+KvY+lHcFGpKjS/YxGHeV7AexjhQ97G0vNhgx606PoYwE8fkXeJZTOXUY3i8Bn68eE/08v4dirpnJy0EqFu1SnByUMc1Xn40efazP5WHTsj86MDqlt1rLhMWh1K15/qidO/8h8vP4Hn/SQ20ftkeXafVbNKkn7JwBSmXyXy1842Vwb7ThxZ4yST1Bg4ZBK/lOuI3RGtXtU4NWCF/cmv5u1gCg1UOwn7rnNoNZxi92kGVzAvZwG4CS0TDeRtWLeMnTp2fLyHUZwyGDiAqRpoXeRKM7Sjsi6Ff6EccaYL4EOwR0LzavAF4U/aIfyi5MPpPywhyT9i+aZyuuzY9uffexAbJ3s7GycM4llQOyDRSaa2BXMEJ+DD4WqABSYHxzi3/fpbu5j0IQlnkJQ/XOjLpQ4a0i08Q2fLnmiEXQ5a8us0w3y/CoOIs0+2DVfP5YaJrSzQN7dVaBTNYSLoTQ/ZIUVIQpbELl6389Ogqmrvp7BzUmds62pprbFP5T5z/HeTniKLg06GemAAEMJicZ9AJzLdPIPriA+0H2Hj5J19HNXBACEAgk92FpgIpL2C3+K/IHkApotccgLXZ2PMUFpEcGJ80OHCWVyfgSaUbKFpdqlLQuAIhWUQTzsmM5kWLGRXcsN6rxa+mjP87KCz2y3dOGYMXYSBzE0rgIotgY9P/sgLeqe+BlFe0ZkJlXBbvr5Jz3uYHY/u2yw+Tf6v0qsM62Vh1r2u6PMbYRVzE+aCCIVK48QbMexPnfz27d8JRg+/ECavQMdXgEgw+fHE+ZBK3Wdg7Ms3vJBnTsThZ9j+K031/9agAlNG0hnel9/FJYo+wmE7VIB5nInuNDrC3KpDOjlBg35HsMUAB2ayFepA21lM/LxO2qM30lENPWZMKS07VKy4ZiKgnI0QbYbaZFbWcNqkwny5zf+msHyDfbff8akTxgH8en5G9E9drQDlx/UViscQ/1aUJiGiqGwJFkziYs1243HSEyw0//IH9yHsfBL7BfPakhVodixZmUnyexh3PXYYzZEz2pXhCI9XwDJp11CkxqLnswu5Y8fW9rlrV7JTuJuJ6fR27z3PsitNT0mcE9ivmNY4y04/O2/uzLZyDeMUtmMOYNNkzzt7GJ8c1MgHwclWbpnYsmVtruacMxNDphijXbV8EZp7yymfvqdrd14jLT2klbuJam3aXWahDPwu4jWK6oKgKCUSiCCtfLXqIs3mr2eHOx+G916EnOcMviD/J62phvgf8r81oTQ8qvwq13fzm6B/3IvgJDZOzi5ob3JU59FecII2xglvY/NoDTx9iB3Z/jriD95X46Qh6OBWHXZeXdTMQSjl1cBHuS1HPl/IgkWBiPzvAvEUc5CTCugVnBBXY7M6FyiHMNyosMRFips6EZxQprcpucOsUGRFf6sO9VqosVnai7vQOeKGIMZ+5hgCHA0Ki9tq7UIPwlmgubWYG1uCTPiQp5qzamlqF8zbcs4B70KOiG3BkOaO69fBXyvQEFyElslccFUUqDlO4ZEcMtdKf+0mE/Dk2I99fIhgYdQBvhn2f5RlxM4Z43LQ/4oov0NjtHJFSsZB2w/0Ad0ROqZDn9YO97fqYBx22owlDDL9rTrgnaQ1SyILcStS44Jdvs/n/z7GWDQ+QptB+orl58dpf+aRY9d0EHYZ9hWOIwrofHqjlGOgi/pcm0t/In+6sO7YLItu5ySWLfd0sxAqyDyS/JxRzuMq9M3Mw7ddu4Enm2VLFzLHKAPjwUMYi7yZC2ZeQT6OV9EyMV/iulHT85/Qpr3YZvJB6ZKFd4VUpbFcOL7U8UdNzWopXJCAEPTll15tr776Wtu69TqO3qykyyWUiVeOyVK4toi86OAYdlyshwZbvaJYPCkUeuSRR9t3v/sXHK9Z2f7g299u127ZmuOd3rKlUFbhi/bueshKWCRcaLYhYHn7rdcS32uF97y9K/S49bZb2xe+8HnmcOvSX1mu80D5sja38Mg8Dbzgb16lzSatjM0cG5pq68k5uO9i0XWRwUQfuMSuJF3KQdDgxeCPwn/y+Tj/D7KQjMl/6PPx3sbS9/W3YnEVmt+xiH24GLnyymce3ZHiYrUzCogVAhvKAOYhbBd8G2Zrw3MwvoAbRjw/widcfg9yX5B1ChZs89ZjxFUYngP4jeC3j5qSBqrUUSalAyOuGicKTn726LNY91/DzoKnpe2ESNgd1UlZNiKzMtu88G4UPnUVNlp+H9ZH7iL1iflMevPznb+qn50Kb53/SGBXRsW2NnSX/IoM1KIjOOH+cnrxJYvYWUNKrfRbiahnfGPzJJJpUlpNJiVe43rGRQ9aJtNRWV28cHZbgzHLjVxnu4lF4XIWg7O4QcdOXShq4ciTTk/hSWqSihMHfIVeA/xbK10Hf967byuQdHz77B6B3y6XPMbgryiVk/E/Ev0vrXyHrzPQ2kljf6vO6TNobqA2GrV/eYmFgrjQFobQZVeEF4YrwsAv241KqE+gSvj6rl1IsZGcE1eVZxebqjm7s6c2glc+29d67GkWtFOrQQOH54i7EDXm1dzwcP21S9sNOxBkcaXtUrRMciaY64UZozNoZ9dbHA3aH+U7muOy00EBZV+DBP6/EvAfpFEX8OxA7+KcWqbOmYDCnwbaRsSZg6EaJz+6F+Ow+xGczIQeDJiOaVrpD0sM4McTD5Li5EvzZveVCX5snHC0YuGChXIakwrpRyg8LT2dNNduhTwIN0CLq5jEzKDtTKcdzXdnm+uFr1mPEcl1SzkLv4SBGyEW+XjWP8dwqHT1opRue4BfekbXv5hbf72Ny598I7CmDP9Xm8Nj4BJK8k+T/+0ZJ7e/mtJVfc5pQ4CKFL4VnGgcFs280FGkE8h77TCSxm89rHfek7Jg6oL8KHwRBq0+tP/r0/nsXLwG9HfBCM3o+2yP7gY6mbUuuW6dyH7HsCN8FsEzO+8eJ/Q2kGksClUDfuutN9pffvcvIzxx8nSGPnbuvPntC1/6fPv8bbdxNeEKKIfmA21Vuzkuqt3J8tYmj/0oQMskTtiFkGeqHKYXN/WtHxxLmPx/Pv7xjDOeeXya9LdvkNKpZxU7/hs6llcJTohv2yVBjdekpf/Tftgpdv8efPiXXOm+t+1/H/XthtYANkmOHPugHeEqSI2uzpnGQgC7ISdOozLNMYxZ82a0ZauWco5/Ztu4YUXbupFNAG46OYH9g7vvfoZbeVDF5ujpUYTSM6fPbwvnLsztVfMXTMeY4H6O/hygfK8F58pJrpKcz/GcW26+vt160zbOraP1gtDERfJ8tFCCdzoXet3wpv2BvJ5sW+ZWAABAAElEQVROFf6oHUeWawKHO80EXx6JIVDaam3EyM8423HHf8nXsTbJ/AkDJFooXp1eEb8LKrQm5Xntr0uY+n4c+pvnQHCCxskZjuq8qcbJky/QPtE46WycyPNqfxzFQLK3xizAdslRhFqOTwsxoq16u9cAewOUAhGN0NtmtHHi7Tgep/Hoozu5aoR4XaebBUex82B9V6xYnhsvjnODz0K0WTzSqvHM3saJghPTreZWHe2seIuPghPbqHn3xmEVrqhV6808GradifDc4z5uEp3hBie5VvsHCidjJJjmpBatwpJz2Gmbj0HIdauX59a5bdeup/9e1laivcQMKRTKzTwim7Gzdpfxpw624ewkG8v/6c8pTTpSZrUXxlz77zSEsEPoNUZk89b3M6K/3HTZyxcnYMjNnP7pnNWxNWjoxnkF79j3bE8+t7v9u/+EcVgEJ1d1R3XSJ5KPo/wo/xf2hbGD0sCPgf+Qy0xwlRd1nWr8qSih56B86u1iOccv4ZdecGK/r9bDYTQcrKz87kaNY0SdGYApAdz58jTGjDP4Z27BWKHNwNMnvMaYYyPce70OI9XaMrn+2qsxOL+0LcZgtX2O/YzzFueNuup3O2wE3fRizlF7qKojCpLk1XpxrDYx2A2NKj/Hs8njv2XouqyJyEeHtwQMwvCkLLUvEocUe/cdaI88/Ghu1fn8F36rbb12eyWmH1Ww5NXjaf8avwUH2r+ah+BE7U/bnMfzHn7o4faDH/4g15p/85u/3TZu3hgcqw06H0NVCnEhW7rtCSbuJUiZHiHJAw/c13720/vZMHsPYe0R5n/vtZtuvrH9d//sv29f/srtuaFSYGzj4su5omOxINbcpcMvoCm4S0gHv/1TBO1CxLuujnvzDbLEV/2OIssAvvUKQhOp+9Tj4vhPosq4Eo7+ypzn8f9ohAu/99WxYtWfXSDuMOL5Ea7A8pkLVLM7r7aprL5TY/NicIZafYZTJu9S8xgyQp+ge36K5WcSY8GD8q0kHzzSD/s1qfwuRlc5GgER7aTpqvHjzxbGMxonPEtw8hLCk2c467uGJdwsBmyi2Ct5VIfd5SrfHLrcu0cKuWD51RCrtMQc/ojWMWdZenaNdxBWfpPhT9UIiv0Q60QHc4pF2zHOF1s1jbvl+AwdugOWHbINMruf6aDBi7tnWDFvCE3cGV3FtY8bVY1GdXX1mkXsytEZKQjg16pJhppqWgK4UBcZJ/6VClt3nb9+9bD08A/9E1g/FZE0hFrPvPns8ukLxqP4r8uloiePC9N/HP8ftXxrL9Qudb/z/fvbsy9oHBYjZZ3gxIGlF5yIERAQtWw7rLJxQkmqQrKYV3Dy6q63MJaHfQOEJc4RxIfhLug8KiA/OgiewwDdNBZ7M5noedXwkoVz29ZN69uObRs5i7+SK9j6K43FDhTgIUyW3/OOn0FeQhKa8nzzTxxWXJ66yoowAj5r/Hft2ioFK+wyK2wK1QizzYIQ/ByIiQKSclTHW3U6wQlILXXJLp05Bc7QaJT+Alo7G4cOvc/ieE52ghz4LDvGYFnsqi/kBEiaKMCK8UAWF/gyoE/jqNQSNLFWt82b1rb1a5agiUVpEhXX9w5h2QFexbkQ4GEALn1aJjJ8F0HwJR8adXakEkcf2kEyI6yLGjolr0F2aX8Bun7yWyWNwp8STDnuOo/kS1kXan+2sagwWyt412qnyZPbfnazj3teGc++lzXbyq1gMH+SJP/UZFJFqvyidcXr62tE4TefygOPeHXovCD8Tt6c1FrXaHZBa3NTkGxOwnSWPlPhivkfO1E7hF7F6k0f3vLywIP3t3/5L/9ldtudJE+wiFtz9dr2x//sT9of/pN/kt0ud9s1su3CMzvj8JbGRTPpIl+buKyUNl5V51cH73V4KPjFmHXz2cF/OegvAQauqwkP8TgWZByHRhbBXgFs/6cLXcIjaOywoH3y2dfbwQ845/72B/y9zxEarrTHfok97FXeZHUW1WiEyCdOHyY7tB3RAlnKIsGrYRcvRniyjgXUmffB48z2k5+9ym06LIrRTHGxvHDO4rZi6QrwzRG8D/ZCp8PsVM7NkYv3sJ3ywcG92KxY3b542/b2ea5pX4VAZgb9jG1b+zPFQ9bbxYwU8b3wD+cAm31BwVTwV3g1X97DXAJtMqlcTm89I+DOWxelgoe/ohc3pL9vnxz9zekUUs59apwgODmL4ETjsIOjOvCzR3XmYo9EHnb+ICnnI+g/zipW4eMCbpZQ80rBhkdm1E7xitAJtIkWs5vurq+CFG/VsU250HGHXXsD2jFRu2TFcgQnHNtRS2gh6vcKZAYaJyx8PN6jUGVVp3GiXZLBUR12rBXEzKFN5qgOY6qCk1PMXaZx3Ej1euHMtI6ycrsShL1Ki8FsHnkt/Cw0TpYgINm6+WpuysGWCRtDS7HVZr+uJoq0LmboaVi0DR07GvsIWW23IKl4h3Ips3ifjDIKhGU6qqa3TzrTduTmrXOfMv3lJWp7ecsXPYILwvJPIStzI+uW5gX+bE8Ky48gOHkK47D/bnAdMVq+po0Gg8tWcdxl2CMz+DdSUWeI52pLPfzGGDjTDiPWRwg6bH+Jq98Y/btEPKS7Y4wCEI8aOpdQEOCi3b5JoaGALViAxgk86qZZqs7mjE/L11Cx80mPjyGWDS/OQptuEbadNjMX37Fjc9uyYVVbuhADp8S3dHk178Kty9MQ/3T682fdg/OREKJUMvjcTLqo5pr45JF+jxzM7VLnH/Ybjn/i6NU3drW7f3R3btX56le/3m697beAWzswV+XGLG0nqcnnhGIO9kzm0W/Evhi4nUV/43HPxx97vP3V9/46t3Pdeecd7ZqN10Swah+jJixTaQSldWudt4sKm1osj3NTzr/9v/9Ne+D+u4FHYFkzcfTpj/7oj9q/+F/+RbvxxhsDaIRf4B3MJp4xA3/SFJ70K8EJ3/6v7AJjItdPl86Sxvmv8tN/xHV5GLNoNan8KfhPso05CRq/yQEVyyKmDiGgKz8xp4zUpeYx7P8q38HvFV4+64iIZqu+U2Bj3GsIcLA2HjiAuTBXLCXezos25jHyMfLaZzbu1X31nv2zjzx4VkAf3D/74AyGjmUhDqFSTzcWEWaDwUawkwiyYFSsnOETaEP2X42wLnLpIvjce+Bo+9ljL7RHuFVneFQHnNiXMLCe4+hKlW/aYfk9I9nYVG/ry7dqVjBK69Q7SfA8j+GJld09IphGMIedvH7kbOIp4M+ihSDXfS767LxPnDrDBISdIOqzhJ0gbVm4A6ShqpwfZ1dco0gmmmAxeJZzy9NZrK9cupj70a9mkY5Um8WgE4yZbJ/nRiE6penO+JNrX0vwptAkABlGFYF/0H7iE6wMOxWTVtSE9j8Xg98yg7wp4DdomB91HcF/5X3p5ZveCbSipr/46wcxSLqPBQAdtOdtPZtrKAN5xXLSBmAQRNq5iLBi0m0aE30noq8hOJGIM2Mwj1skWMxp8VvaaJjUo1IxAItmwwRHpRZi2PAartLcse2atuParW09Bnm1W+hRlvAvuZ0PP7UBJ0Fxnh2Cejzl2fGi6TtaXFb8gyPLl8bsF8JDSvqdSEBww2QonJrwXn/rfOPZlw901xFz4wO7jTFUygCMsuSA/yoVfGv77pkyT0SgLHKdwHuzyWKEi55FdsKj/HAaR1AmPIaiH4OrN0mxrYGQ5Wxbjh2TldiX2bZ1Iwu5NTEwyRgfNLozKTpD+TBBag1stGHydVBPPeRjXDglbQeY0m769s/ky3okljkKCU89uvqnPRjeBflqnE+S/4s1uvKBLeSwI+5422sjU52KiArskUx86gh3B2RXwWj6DOpvgv4vFR+Bv0Qu1XoKpgDZlX+p8GfniyJTKu3Ob+ktbdzNmgH+sxsIvY9j48TyF7HDrnvl5Vfa3/zN37R//b/962iTnGKx6ETtts/f1v7kT/7b9jvf+h2Eadg+Is/lyzBWSSE1AWN6T77uj5ZgrsqPTSr5Ibn7Yz3sN8vHX8ek4ReRf0X4h/xP5sOCLTyuL192kkzC71gycGN8NvLRvZbGiYlUze6KICzjFrCcYgHx9Atvtb3vTnDM7p326psYbMVIiW3AY4YgCJX0WdHk0BDnqvUcp0GL5zgL573vvI12yOy2ejXX2r6/iwX3kvbGm6faB0c5UsqieBFHbZZhL2UeRqI/eH9ve+/QHuxwLWtbmFTPQrCy+43XY8dCTZMbWSwvWVQ3v01n0QsrFJwOpOBYatnv2MP2bSytUjgTJ0AR5lP4emDFlX/SsaddvqCt7aTid82BeONuKvybW5/nr0p/hRYTwKew6omR64g1DnsWw5QnvC2HozDeJOVtQd5mMY3+dAHaeMewH2C/uwDtHAUn2grxiJptQEGHWigRnLBQOYGa/XzSiB+P0qh9NZdrQb1ZBMzQPpblmuGB4IT+RE2VXM+N/S9vtPK2nOFRHYy2ch2xGwsnmau4gJrLMSyvLPZmnUWLluZWtdNqmYDimdiUyo4xdXJuExttjNETGB1eunRehN03bN/QPnf9RjRqF9Gng2V4077eoxXWUeoVD0sBGARYomHKuEMwJEFA6NNXYw7aibxj3OKhiiFHjY4/SVRBI7+fNv0vN/+lfOEFbwrefYl2Wi84EcegMpsL9BfDozo/gKiLEYxxXCPtz00Vm6LEwJkfZJJSwTRM0M9/jKfvpzf/tnDYxzKzaEe7kL4smr3A43FPNy2PH8HgKN/OMxxrvM5eF27Jwoa5H6dL5J3DRw4iQGYcQXv12o1Xt2s3r0WjdRlHst3AJA7Alf264jXLt18qWPUjl/CfJehbIXkGSQq3DSsnzuq/fAuf8jHo/w3IH3mIZANkfBua3zqz7159kY97/CeYtufGoDh59fU3EJzcw1W/L7Wvfg3Bya0ITsDRiZMKY2n3aFzblnKzF9ombmRYnEclbWbzQcAzv/xl+7M//88c81vWvvH1r3PMbksMsotTbSmdxHaSRmjdsPB69H173+GK4bfao4880u699+72+msvsvZx3j3Rrr766vY//c//Y/vDP/xDNBBXM/5zrJaNY7FQ7V8IBK6DN3AKMFgK/4GQD4E/cc0xeKushvjiu3OX0v6du4yXX5mNVamP0Hv2z77gwbMC+uD+OSm4+xwJHXnt4457dV+9Z//sIw+eFdAH989Jwd3nSOjIax933Kv74jHUOOlj9M8+5RTPinLhiOeHTPYZfg/e+pf+OUW5vVdFuXDE80PGfZx4FxMXK9rJhCf6Vt413mJzuw+DzUNnQ4DZM+G3nzGR3uQJ87lI80jGuwcP56jOQxiHnT3P64i5Oz69APERmkTjxGR29uk8+Bgtn+LkZSddevfl97VInYlwFROSUl8nUufssJKn30ngTwdUXoff6Ri78p34CZa7mVbGTuoEk4gTEZw0JomL0glYq+lApIGlaUxSz5xFYMLi0TLnM3nYcs3VMfy6cePatmYFFrpZtMfR4FUXjDV6OjkBTF2FjjLrqlzqKVwk8Fi69fa9h1+fqj2/xqPTHYRVRL4vAr8ZdDnUm/lbOF+j+OezsvtkyrdO0jKCk++hceJ1xGfROIlxWOEglIXCVQyOWc1bJerjcY8Qxfo4GMEvx4+faq9hUXw26v0zWJSfZqKmtpC70wqkzqDtcxZhSSZ6EHQlk7ybdmxpN++4lp2GFRjjdfgXR2bqW8HveVYJ3I9jBouWPHlJ9PgJjVi33kbg3czEX7IybwPqe5hLeY1+J49K9Mngn/ZgyeJDHZ8AlJozOQKwCE6crFJfUe1q57lXuFXHozoIO7W4z0gZOxXc8UQeo+0PGiVfcUEpfDj/cqFwCGOGCk48a69ldo/jmf85rjQWx9bG3UqWfezEnsUOgjcurEcTax1n4TUcZp3JLI2P5TELJPGdH5hGmPKuZwLAOUD6aljaQP9BvQoLRE1gYZloXT56JhO+u7h6WU1xUkF59CmTP0G6Ss2v5fwK7c885bXwEsJD+T3dKn4H9mOkDY0TbVpUjBRtivwLrixfeqe++Umk1K//9IMU5pFnChx+Z2EqkvQahZ9vsse78GOwzgmGCz77+xm2N2wgePY6KtPEVsvBydpRbj9wd8rJreONGknPPvvLdv99D2Aw/KH25ptvkuYkGhCn2udu2NH+GyZc/+i//seZvGV8ok7eyGK51txz0qNOnNk/x+AvnWfA7UG0ziE81UwGfS59DsPvHr4UNAo/USs74RcPpilXqfn9GPS3rkM3Wn6Vk+z7+ttsbVhdO423MMIg9qEnmfz+7NFnuMnlaHvuxXfaW9gmOo4dA1WrZ9GQIjxmrFVLYdnyZW3ztVsiGNnPTQmxo4EA5PSJg+3A3pfRiLyGdorRc+g4eyFHLjZczU0w3Nxy8BCaCO9ji2NOuw47JuvWrqCNnmVxfK5dvXJpW4v9oVnscJIV9Sx4ehDDYmFkIYZS8GgxVwfrGC7FLwCbhSiSzkYfuB5vUsB3Kd/9fgz8j2daOVjEpdBf8ti37T1wuD3BrToTGIfVxkkvODkG35+c4AgOghNv5/DWGzVJFyC0OHbUNt1pnHhUB0HKbNQz1NZTO2SCBU+uKkbI5YbNfK9dBScKTlwYzWZBo3BEPHmk7fD7hzlm5ZEeNE7wM2wGwpr5tB1V9dXOXOW1xdiG6G/VcWHqbrRHdeYgiDlJucfZjFiEja8TJ7AnhR0EhZWNDaAJhJ+zUGFQMDbBtdGn+dM2xBduua7ddOM2jABz/TVzHjb5bQ6howsbNy/i8hjFt+/8JXLhvxLhLQOFeSqK1JZulW0yik/lVnn8Kv2vmV0K/a+I8u3YBjh0PLdnYMzUP+3NdlRt5ugpryPe2/6fyRontDs3rXJkTqBEeNqzmBHzfo6O/6FUF1bli/+sB+JbP8Fp1/+mseZntFFbWH0P59/4Ub686b9pHDlxkwblxXQvdeuc122jbYVWlpuYammhvhUNLo9kn+XPeQ9ySPj4A9rb7LYWWx3XbV2LLZNr2oY1XKsLr2pLT0GmLD6kP4ALP/BkLtf1LVX/EeCMklTWlz/jG0l4/cv/fvyhMnwXrGbeOz07+HmmR9PLKJPwX/WrMs9Eol7HxN966+32k5/+rL2KnZMv3v5ljsntBB6uKf6AI37gx3mw8zGvL4+dO8YD2WUG/byFzGfu/OKLL7Qf/vBHEXR89a6vovW5OuOxghfz0vj6HK6mP4BR8Mcef6w99NBD7emnn0ZY82J7/8C7bFpib5Dy5iCI/+M/+eP2z/+Hf86cbhtwMC5blmucuCG8k7+lvzPw4OlD4O8yS9TgrKNR8G+gKKWoj8V/Hf7Ty3Tpk09f2BTPgmYyTMOI54dM9hl+D976l/45zO68t4py4Yjnh0z2GX4P3vqX/nleqUOPRKGT9zni+pS9d08NnpMQ28esxP1X/xzJcuR1EJqXwdcUMQzTfRrlw2ppoF35PVwprbqF4uSufDsEO+O+KqpUwp5KfrME1Z8Oz5bvQtcdvn0swh5C4+QhjMPOxjgsF/CmTBcHHpnwnHM6Gjr7Pl+Z3gmvjD9WvgE0sPL2WTEKBL7TgVW6NKhExD+DizD0ztCafoXqwmVc/luEny5IJpjkRAIJPHVUx4mKGicYT0vPVre/eFTnLIvGc1iDnYdRwzWrV7ALs6rdsnNbWzR3Op1WTSi0Ep1z/+KLWw88o5+ze5lcUCerWlXgaR1rMe+56GrQhA7wT2TSWdfqY60/HwVQl0ulE9Zx18FveZdM/0sr35rINVC9/eX3H2jPPI/GyTm0Tdz9EF4qpMYJYjdqbYcLf1GUvX3OlcI4GcIZ6DWM9fobb7f5CLK0JK7x3Qkmq2r+zAQ3Xot4lr/FC2ZyjnVD+9Lnb2g3bL+6LZhdogBWFrArwi8GF4/6kDHoKzqIIetp+eV4B78uUB14sh6A/w1W8yqYhqcTOysH3y4j/qsiHQTOOiS0I5KVDKDMT+EM4uWGKAa5Z1/lVh0FJyzArmICf47Fr+fStXEiJIHOtJmc+az2Z3pzPgWu3zt4gImMghM1TsjcZA7W5K/AdIJ2MocFwqrVCzEcuL5tQmV21co5LNIowf7EtoTAxZsWpnHDjNfdlaNPMR/bgu1FOlkoTk5JO6U+tQvR1ZXC5RaBtGlYf0OslJM8tRQ+a/5P8ZSf/seqyULSpipGVRVSSSVj9Lfq9Ed1TMCfDwkHUE5MpGzg8yNw8sxLRzX99BpxKR9Ps6nEROiTmbW4HHiYsAvsyhezp9HicldqOoKTOWgoeS3kBCrTnqs+zSIsdhlcPKIOfIAJ1s9//nS777772sMPP4yB0Xew+TSPid4OJmDPs9A71v7pH/3T9id//Mfthp07s8voLvdcr+eIs35UrCvfp+1QFnD86SAlSgds/+zqbbK8dnmN4v88+BN1Cvgt/1fsf2s86alDQaluV+fUrf+h/NyqU4uhAGtk+x+g9ujWcRazf/FX97RnXnq3vb0P7YaJWbQF1M5pazNoSy4gPOPvFcJL+ZtBm36N/vLI0RP0l16xfra9+/Yrbe+e59rmzdvb4qWbGaM5SoXgZMmyOe29/Yfau7t30x6Pt60cn/vyl29pG9YuRyjTEDyvRBigUJ8xnHJU89c2V+jT4XyAb88MSCw1KUOIDmxpMmBePkJf/NJX9TjxyV/oJ/y0DPvXTCCgu14mMULSG9/I9dRfSo678fZ3qfR3t96rw/dwVOcJj+o0jurs8jril6jmLIRYXkdMf8dOr8dhFGA4fnhM5ihHdRQMzueIjAJ/BRtqiERwgsaJKu/abziJbQLT1VEdjSyfYEFY8Y6iiWLfupSjOl4zfALByUKEMt5QcwwbEDMpM7fqKDhB48SjOhqUVVjiTrONXy2TWdRvLkKSE5SltstiriO17z1F21ZTcYY7Nwo3WYS697MGzaPrEHR/8Ys72yaub50/3xt2FMA7gjNe0G5d+Nrf238Psc9baCvN6OMhUzVTY5iydy78paCtu3ohKVrEls568wP+Lyf9r4jyYX75PzgBt26SaGTTEU5eE6mOL7aZowhVn3z+HQQnapxoVHVexn7P9Cg4V7ht/Mpuiv6vMvtQ/FdvbEYSKsRKnsOfi7W/qgAW0+BjeIQ+ZRqLfNuZY+V0b52Dr455VIfqzkcgICN5S5gLfIUh3lI1wfEcNU6WLl3QtqMRd9OOz6FlMgcNOg6EWASMdxWCBcevOczRU8++qvZV1Dv81/k57xjnP6IUexKvIC7elgaDjJJL9T+mN4FhF4Of4EGMPp/yGZZPjsBgF3gEgcUz3Kjz+qtvtOu3b8fe0eqsMd7Z+27mtBqCVcjqWKpdIrVCFWSoZWbbcbx87fXX2qOPPsaV9BtjW8wNj1PgTzQ5D1Do4obI3Xf/bfuL73y3PfTIw+0Q19EvQMtw3VpuzeGW0Z8/9XTbsXN7+1//1b9qd95xV4TDZdOIHtiGPuamhj+bChaKqxgXgp8In1L7zziSGgR6C8qXNcp7+KDz7sDqY1bE/qt/dsknPQaheRl8jcTq/XzqrszyWa/ZkvrKVlWn+p06Ru/bP8dTOsmtCdOgiVUEovd0KC6dOv1oblPH6H3752gKyrhA+Q5sOdsuTXivhVTXG4ROHbF6ohFtWIKw8JUodNb0Ir6mRdNBmK/Lyb0HuY74kRfaw489j0RyJdcR09HRCdIjkhqNE68jNqETwizkhuVHCj0o0ZJHy88nP0ncPYkThNZ0yRh+Cqf4tzEq0AmZ/QIvFl7lj8NvfDvrUq1WcHI6ExjzXIqUez5G8LRUfZJdJA2jeSvOMm/L2bie6/e2tC2bMFpV2Qe3yUfBCXVhjsEA4CSTTtZOxQrhz9kecCZeXLTzTR2FbnoGQksmCX/6DZ1fNXUchAqTMf1P8IXgl1aXTv++Bh+vfI0aSlc1Tr7z1w9g2JDriM9hHPYqFuoSgjobmqVt4C6aiSNtnBga7SZGjeNct/nmW3ug3zkmiHNz+8NJVP6PHn6PMexUW4oa9MaNa7gtZzt/17W1yymbMk6TbhaDrFdCB37ydELuuFa2JFJMABTfvRNS69dP8YLjBApNQslPAaBtQeTzX5AuwH+fKv6tuH+CxsTctiVXnFWDw7ZmGHzj9cSynrV/7rV3MQ77dNvzLpMSFsLnmKRY9+K/JOiypGUXuCnAYwA6j+o4MXcSrtFDjbjV9YPsBjGZmcZiYOnyxexUrm3br9+MyizXSJORJ9y8dWUWbSjXOYef7QfkFR7OEHxUFULvfBueitBu8jSCFevw330JoD7hKYKFPZMin8CfpNJoUv8T4vVIzNN0VQKPzlner9b+qkLiFB6H51OG2fJygN1sjcNKI+vcl28PoatxpS/fNOAKgHrePJ//hN/M+Ru0f3IliwvDb0nlBuXTdylsdLKj+r43crgQnD1rTpvLTR5zWdxZ43379rTHmJT98Ic/bPfde097e/fb9JPLmobsvvGNb7Svff3r7amnnmj7EKT83t//Fkaab4gGi/3tDPrIORgxzdBslXXAFm6wIn4KKeXWpKt4XNCCqK6y8rCJxIn8Xrw/Cj/xPyP6x0ixFddRLUkR0FLXrsIViuCEF3nSClel0wbs/+T7k+Do//g3/6k9+cs32wludJm7sGyInWTBDlVYtHLigkXxjTtvYAPD23DeQSsB4SUCmXNoUC5dNB86cWxn97M5lnOWK43nL17WlnF8h4Op7ZXnXmr7EZx4Q87tX7ylfe1rX6Q/5ZYF6sKahMqr0YfQBJx6FK+0aez/DDNCwZf5h3gHx+lvu7AB/MaXhYG1ECK8XeI+o47fKymBabTEScSe/w2tsPB/l92n1f9aahmHPYSNk+dhoe464idfohkiOFFbBKGTtktmY/TVm3NcoGqY9dgxNE5YxMxHeDih4OQo8TCoO3s2NkmwaTKBMMrriEtwgtFXNEnEYx3V0cYJ2lwIG+0Tly9fyhFJBCf0sQsxIqsg2YWlmpe2RQ3FqmWyGo2TQ/ifoq16XMhbSzze41GiuQjVTrJY8ghRdvEZTxXmnENjpm48w/Qwdqg2M57eefvN7bdu2cJNQdAUOmgnwfE0x+Uo2+Ym2TTYrhvgH75xLa9mEZyQSEP+p+0SVv1y9WF9v28eoWdSFMX1K0ei8IZP87SA9Ar5PL//u7ztf3T+KfHSHgLCJY4/6csVjIgNn/aBGP4UJ8grxXM09+hLj6Fx8vgL77R//x+9jpjjWNPZsAq6oAdzXdux/BnE4X9J8++Pgf+p598CAh2xf3ga/nSefBV9C6/wEVV2M4d59DE0GdWoW4LNrDkIQ86qFXWK46DMMbx5ZcniebmV75bbPtc2b1qJ4VeMizN/nFAAQ58lhjxGlnHBeWXGCPknowvl08dmLB2gg5fUTAiHziQ93/V8mL6JAMPIO/1f/Cgmc0PTkMsl0l86x0wCdfXI4t69e9vrr+/KbZ8al9eGkTZJvEJ4HoJU62H8c5gVkN5q8YrLY9zyJb+8hMbJA/fdj+BkQ9NOyurVK3PTlUdhjx453J559tn24IMPtD//8z/DoPubuZXnxptuar//+7/XvvnNb3CkcH7767/6S2ya7GSM+Dpzcfoq0gZ+aOFzuP4qoeivAv8Y/kMNysozCLcwvnj3v8U7tnd+H9b+BvSXRFO4roRJIb1v/xwPHtLf1mjdOtfVL99JOnX6PrrPqWP0vv1zNAVp4LOaJ37y5SN4M/tCf4odfAxexmszxVcf04xsJ/1zELWPMPAYfxkLHnwMXsYjT/HVx+zL7Z+DqH2EziMI5V1v3VgwH4FhxFcCJ24CaIh+2VFbUHogPmkwdjguSBSnvIvgxFt11DiZMxcVsKvYQbQTdEx1J0rhCb+m9zlgIt/F4dA3YYnbI5c4Fl1tk6aSuuhHQt5NatTK1e/yT1D8ByFGGpZkOv5VPoaU8SknPsZaws0rGtdTy8TF92rUlTewGNx4zWqMWi6jI+HIAesGG6yaK+blwk11ctMXvnij8kpjC0Y7VQuur9HFQu34m4OpcSPw5zP+4/Bnx5jIPW6q3C5O4pvyAvAbYFUo8MPwP8zjo5UvjGJEjRNv1XnuxX0Y2ENwMp3jOvCNA74aJ9E8sYPnXxbF1qUTnDDEgywFJ6faG5y11P5kdksaV26i9TCPxdama1a2nZ+7ru3k7PU1G7yZBUEJ5WozQU0nJ3niJhBSrnlSMv8CevEVY6hcXliSQyswPCXtEmZ4ddwhr3xsfsJ5OfFvVQflUz9g9gxpLVxEJoHht2i5hg+f4VaduzUO63XEs7Bxws5EBH2d4K7HjekK9sJHjyVVz49g42SWi2fOzKt54GJcewtLMIa8Ye2ydi3XC2/dtIZBfXr8nbBJB4UasTNDfMuptlrVrGMqtkIoYJ9hBNpWBAmJzCdEkVbVbgv/kse5kXnZ5gTXiYyuSsGfcL11Ca7XfHwa/F8T22GZAcWS5ZdUojROQhoC9+dWHbRwElh9pPDXaNVBERhsf2bDU3gzYSvcJil5J0za4WE98Akq+/DywbvzuDD8RKFCGoI1j9woxoRw2kwpdBW73cexrn+gPf3EExzJuafd/+MfcxvLuxh7XdVu+8Kt7a477mxf+OLn23XXXZ+dL41fnmTR54RtHmrFgZU6uPBP+5d2VAbwKM6nMEJtSFl9LHCqvUdw1t68GDcsnkS8B27SdfzVgVjxzBenXxc9HxeGn1hm3rmUZZ163Ab/U7f/6kT6lOPPyeVnnEQrJLSikJRDJPsq4yo4+d//z//Qnv7lW2xILOaIDePrBLdnsOCdOfMsGj0L0X5cGY26N5jwnkRYglpQzr2f4WaVrdesbzu3rUEQcpi2OLPt2s2REBbUy1Zjv4tbeA5yln06xLhh+7UItLbQr3KbC9pi6TcRmpxDMBq6BBdMiDmyU8coCy69jStfWvmu+x4HeuRrAH8YtNIUVUYi8WofNur8siV8FPwnTuJXDpW23gfl+8nHh9HfnkTBicZhH0NwghikvbnL64jROAGfx9HsOIngxCu43eE9TrswXwUnRxm7cqsOt3ucUoCh7RImDXO1ccJ7jMOyQaMWiDfdKBCR53JUh7zM0+M49gMahz2E4MRbfBahzTIBzbzhQi3M+bmO+DiCzePcILe6fcCCU8HJfATb7gifYpHlFciq82ug1nq4GD2JlgryFBah3MzEwnTdmqVo0F7fbtq5ta1dPa/N7jQIM/baFsG/Gxtn4TF6hkHfYA8hXnXBtW2EPpyYxROE9uE+08oZPxWayFt9mGkHtAph+pDe/7OnvzDpql6XqXzQYPkuiGt+BF45FqumkP2HGLT92WaQibUnXtiD4OT7jO+LGd/nBfvMsOAjMW/v3TnzNV0oUL7+Busj+PfbedHk9leDTOVRNSSmFen7yMoy5SXPLu8BxelfnE+oxaZQ1k1MWUbBicfjjh4+lHFnAUfRZsOjpxGa0CNyC9iMtm7dWvj0WjYwr2GTE1zApG5WnkXjxhvgmAoi9EfzRDgCERl2b/ESotRVmhoQrkyc/PTJ/Ogr72sPm/F5t//zGUQaEX/zK7wmpyngL3+jVzxeOqr0xdrO1Oz0GNNVaOPs3v1Oe+mllxGSIDRF2D7BsaXVaILMZvNrBoOo7dxjtI7XmhU4Tfs/gqHv9/bvT3+hYOQHP/gegpPN7R9++x/klq5XXn61PfnU4+1ptESfeOJxbtB5m3xPt5tvvaV96/f+Hlold7bPoeGycvkKEDvR3tm3Dy3iVfQZs+hT0P5hLqAgPf1/B2XNPwqXQqXrYfJ9CGl9XAj+QkyfQ5+HuIUL8a75T+H//PLl1CrLMkffzbE2yrsIBo7XSo8Lur7+RXfrQf7Jo0vSR7hADmPBg4/BywVSDb37mH25/XMQo48w8Bh/GQsefAxexiN3XxGcjIZU9PFEY37n1Wo09VTvw7zyxk8a0VRR8Rsrq4sz5vcJlF/D3TjzpCioXUIVSoT5qzIyokxfy8tMXOUKGjGxA0w1CToh4pVxWDROvI74cQQnjzyLGvDqHNVRGJlBEWvsCk7CvA6WPZONlk8nFrjTERkhBzpSP+tmxx7mNJLOKlmdes2zK6DzGYAzwH+VPwgmg+o4nZRrZ8O8VHPzjLLbgMsXL2SCcQSp9vy2gSv3vAXkGq5OXb6YqzJdPDCROMUkRhU3nfWbBr1c3LuAM99z9P7i346tagRkVoQ/MerUQrw6SZ7eTyLG8G/dxFnXEVhJHekvCf5KXb+j+A8uzHcy/S+tfKvJcMigWLfqPIeNk4nYOGHBlIEbuB30UfV1IHdYdsx0Ae4+lThRkKQB3RzVefM18OzRGzprFhmrli9s27duaDfdsInjUiuwoM6NBqyexbdG7rL7Qn5OGFWRVmVdLQexrfPXdpnmRVnFXJ0/nwp2nF4UbxAx+Jfn/c83PXUGiD6vQaZE6MrIm/FNPhrPuLpPAv/hlcqurndUSCmviQMwKw+6yqQCtkfr/SyCk7/tbtU5x3XEXht8FjxrNE4BoOnkzmqRppU++HfYUz37fWxYzGBS7bWyno9fsGhBW8YNC5/bupy/ldiVEd8YC6M64sEz8U5qxLc/2RkxR3jAI1RqCDjZoVBbAnHlC8q0/kwaahHV14A8CPWf2Umn/kuBru86w6RRdh/w0le/gfsk8E9mg3zzklJTRJVPmV3Zlu6gr8tRnS6hUBzgqOMxryPmnwba9BOO7DiZvvsL/iwRj/QxidXBlfymKr/ymhScig/43zyTLzUgi758xwc1G9wZdxHnFYfS/9XXXm8/uR/7JQ8/1J589BFsmryXIwg3s0P1O3/vW6j3f7FdvY7jctz24W64+S1g8Sh83gYQeyjwnerEHgFxEeiC0nbmYiqCaGmngN7U9LVOdkpwSrQAIyLFZuf0Swk8eA8mzI53g/wbjf6JtL/RfFNASiVrIabMkfL5PM+lPlpVpg3I3SYItYSX9miTPYng4t//6fdyGxbytXZigkU3GiU0O4T3C7nufnEWxK++9jJ9pUKRRYSzk8sEeDYLkuu3rGs3cKPY3BlHuJp9ARopGCHFlsbchY5VLMKx0Lxm2Yq2ctkSaOiZeg7acpzEHctjaDtot2M63+K/5gaCQZ2pXLW2TghoN2HtRThPYRP+vkX2+A/tjIpHz2tknixtH2PzD3O5zOOfcxxJtNdbdZ7qbtXZ/X57nPnOGY3DIgw8zq06cxA4afhVAYjafrlVByO8E9ByIbyvMN8jMjmqo+AEAYaq8dLQNnYUgcsi+lEXQCcQMioQcW5xlDTyRgQnh7pbdRIPAbb0UeOEdqZgUoHIKq8jRiPQHWmFN9pCcRx0sePxSoU0EZwsXgB9OMJDJ71y2WIEZ1uxY7IewfdSbloiLjRx91o7bdM0dA/hJjCeL82mUTcXdLBmhCizELRVA5NeHf2hpZsjHnWSu3v6S/qKLD/pb96kJ6lpx+gfBoGxzItgP+P4kLeSRH89BxF86cKMY9w+OCHdT8qquoXHktkk/rtSyu/wZ28oZApPysgr9acdikc8EguWy1Edb9XxOuLGGF8hajA420pDBSk9/kkLjwfUrv3ZajMKdfBnsSraLF4nTnn3U9zGOy+9TxdmHOOOxuO9K4wHGozw2DQEJUh5wvtnaTsR1qtxAh9r4Pg0BorPnTqCdtastoV5+PXYMNm0fjlC+kUc0dauBy2EeZ63s8lPjjEeZXTq4+002kTxtrahs1IFc2pqH5OqT6Y/0cLPXUrj4LqoI/BDB9EqvihfmAM3n1PCbyaT8N9n2vO/pdT4PyPt/LnnX2wv8rd27TqOzazIzWdzF3BbGlnFDhkvjgnexqWWmcaixZ/HphchbHn5xRfb9773vRzn3HnjDRiAfb/de889XFP8E/oDLrlAcHoDR53u+upd7Vvf+h3sl1yfozhqnNoHzGJ8pithzC7tUOfa4th6ewOS8HsbkvAHB/7+CvBLx/C1GZORuQa9PP34Vfivp2lHrcrQfHFjftUoKuAj/VbqQT4j8E+VfKysLsKY3xVS/iTByRDIDltUfcRv0lcPQ2KMR5uEk4sGjsQdiTd4Hbwk3ujXpZZvunCZOfreOdiSf7Ijriuo9wuT2rDpQOXW6pjIJtHpUl2FObCS3onRPlTNH34UGyePaeNkGX50gk6ACctSGOGJpYXjuzr0ZU1ZPnFcLPU1tFjmEynfMl00pP/v8krWZpQKUjfi5L/52HPZg+ryCBDd8FGS2oQRRwOGnjVWKLKEWx5WrVhMR72ubd2EavPS+W0+Wg65jpjwGGFSs4FOWmeuwqsWRV+fweKGth9P60NE//x2matzsim2Cl92Gcbp4CeNYIlyn1PCn1ySEz+T4O8RYXCHhorZY7fzJ+8e479q+eELKkt/2r7zvQfbcy/tjo2Tc0w0FaZYzvToqFtSgMriOfxWQOJvpzwNjYbj7e23Xkelk3P8yxaiYbKt3XbjRlQ0oQkq0HNjzZvYXU/qEaBc5eZqHbzKn4qvnJzV7nlKDNDWMy510r9wF1/qEQx1eHdC6/ze/BIuTUfwWRkZ8hni36Jw1ijcJA7iFJ7g2zUaJ1rBOzA9zw1HP/rx49zScYi5CmfgmbQwtwA2Fke0q+IvUe97YcjcnbQpTDnJ5OR9DE/OQOV82ZKlGKRcgZHPDWj/LGIBNqctYTfIebQGx86hNmwqm0gWk119HRDVYIC87Ig5gSEedRQGW4IaJELS1SZ5aGPFBtDXL42hoCQFsBIu9tP+fPMD+lv+ePvHH2dIpeDDwpKkL9GkXYwq9lduf+bcFzRQlU4lNQ57KDvXgmh9w8rUR4zEx/fUYwh/hDDVkVS2xCycTOI/44z1f0aU3heB33qQzAmZdhucEKnhsGvXmxiOe6L97Cc/bU//4im0FfaxUz2zeb3hl798e7t+23Ys7q/mzPkyhCzzWEjCCyz+ZsIQ7k6dhh+02u/xBduS9bVdyb+unXqaGETHmomSPOe33WoW7yYTLba/jp94iyt04Em+Jgoe+xdjmK5zfVn51D9JPhn613GWvqQLPavQ6gYpFxrVDqbY0CG+5UVV9Qd+8mR7ZddhbtThKM67HMkAjwuXLmzrudbZ4xe7397NYvkQPIKpRf7O0fbE+QwYass1a1kUryWjA+0ki/MXXni7XXPNhnbd9nXt6rWr2hKEYRo9d86rXaK0WHZ/nQl7o8MMCJO2LHJpo5nMK9Cykvx4ZCD0g6CiPUSz+nn303R64AA5cfqPQQAeiTOCf+ibry7N5Rr/bJNqw+3ZV7fqnOGozq5dB9ujCFHOgScFJye6W3UUThxDC0SBv8dkTmJnRoOX89lskY6lcVLXESsc0bD2Im2cIERRy8Sjj5Z3AmHJdNrdbAQYXkcs3pZy9C22Szyqg8aJmy4ez/GoztDGyUkEJxiR1Tis8RSc0L5OUj+Nw85VcNJpnCxbMh87YNO4eW5zu+WmzWhsruQoFzv7MIKisOnp/zn+SVuNZht0phqG0K7gTed/NjU8Pf4xcGmcfMvP8qIBru51+ahXU1TL9oUvGMN/+iWM7yuB/oIjz14u/kv5oMh+koYaFIqZ4IsOwvBaYIpPbJzAcE8/h8bJ//vXCE643joaJ6YjYpwdZ2F6iP948WMcgPVh++MpLcbh5zvh/Ei3Mee3tObpf6P0L8YzXZy+/NH5uch37qHWiTdQWeZMJgwK/A6j2XqGo2VLsbPhNejXY2B++7Z1bd3qxfA8gkqF756VdwJj5ibmaX8V/BiUOiIIIv9h+VWLVNL6mZTKpv/lWygy/pNfaZDz3Q0rlhAQRWM+9Bi6tBURZhXM19yCMOKMxO+4uxLqbzQiGN/ytf1CpWMLbvfbe7hG/uH28kuvtB037OQI9Pa2mr7/KP2Lx2eDQ3ARoQn2JPU7gf2xc/DCSjTVTqOh/dSTT7T//Kf/Ga21D4Kf3bvearvf2QNfnWlfYey+46472l38XXPNNVxZzPX09CvW26pHkwVBaTY9yNNNs6sQ9qsxF40TcYsLjnvkdGAHZAHShT712sNa/vxOgv+Tnn8NyrEo6RNidBXr8D/mNx6j2gLRE3UQP7lO+rlo4EjckXiD18FL4o1+iTqrHb/RgJEc6/WigSOxR+INXgcvY+VPEpyMo2k8yUj+I1lcPM4wTcUbxs7b8HMQcdRr9H0QYfBSoRePM4jcsUSXBoz75tw5fUjPM/glRvedsQuPLHASUI3G1P1OkN42keTDj9829Hf2HURy+Wz7xTMvMVDTwOYwYDPp8pyiAsPEp+M/aydnIluJ9YIT7KyMYDl2Gf7LApqnkbOzn4HXBRax+JuBVoE7NO8ffD/3lq9YsZIskzEPU7uUYrCgs3OiF4NmKYZhn4WiRzi8wvYEarZ2OtNdZFOGi/QzSK5XrcQw2vVb2ueuQ7qLAcxFizH81mmNOJG356ynsNiDCpEw+UXJlCkUhcuCzPjCaz2dTOorXoSSV16UkJMLcQAh8TqQEteYelua6fLBY7KzHg6TKb9Lk3T8JN1I2hTTfacueHwS5TtdEgY20dsP7v5p+/kvX2XiCO9wvMOJXHBOnOlIsc846BG3BBIdzvSwQvxNcFxq/4G329YtG9vOnTvadZuvQd1/SYRYsAE0DpGzIBMX0v80eTp2ev46xzugh/wf1QfjGBGnV//ucG+cHn75cgz/1MX4wX9++JjCfab4t0ID58gu3dWXoRbgod4RR4BihRFq9Tz34pvtwZ88jh2Kd2mPtFXO5U+ccarMyokJsfW3BUUjiLzNR6FGjpuB62NoYalptYIJ+g3bt7VN7ADZXhZxBTeG19ssquEtC0GRSNXB05mAiGz+5I20ZeurKko4VmCsM+Xz2uM/viFEcXShvsuIFHHQpkKtr2/mY8ryxSu09YmXxcQlRvdtmMX09PcjxSZ6n1dH/5E8KqfhrzWZqv0VHolnH5A6JuMIBD84dIQdaCdA4j09bJch5abornwe1k8Y7NvSRjrcVMQEUb5PsMCLuDB1+hZwXRMd2hXadWLJvk9w7CMH1wzTeKLyS1kuzt546w0mXk+jYfJIbstR42jZsuVoM2xtd951Z9vBTpUaJh7B8RjCbDVN2OFzomUbz41kVGKCMixP2gdPPsUzfV9w1sEX+kNoQbNuBQEQ8R36B4cE9A7/uOQ1Nf4vdfwz30H5vFf58TToPBdODImslLUvygzIFJ/6SbWd8xOtvy0u9E0AP/CJgpCfM64+/9KuPPfQbuewcL7++m0cg9qKYGt2e3ff3hyTeuGF57iW8whHRI5jh2tD20H73MbCePbMc+3lZ5+OQcXZsxa2m27ZSR+6PAv12dDnLJ2zbXxWxkHx3LWySDapXOojXu1ZpBT/0pisunGFOkCHZoZnXNdPMIyRBhaQuvzqXVImjrQj4qD9QbDgLMHJIfl8FPxP1f4ulf62pZPMZd7Zd7g9iXHYoyfOtIPvHW7Pv/gaC7w5CJBndosRjVQuriNshw4jlFoDXjHMu/9AW8QcYi6Ci4O0G+FbzPdh2vzho4c5crCOudIEi5h3clvOQjSGvO5djSFvLXOsfA8jjd6WM2fO/Pb2rjdzI44Lmw/eR/gN/ryu1WM6tsv169YjKClDzYu9pYd5zlGENLNpi/PRblHbSwFObsu5bSvatKvampVLyJtbs0Cz+BXR09BEqU0gPuxIaKOZrzEDlAvS//AclZmYVHIG/9IzWckNNf6GZTp6G9cYRX/y40UeGdC/44eKd/nof/nLt62Jf7HYjasiTVzZPkFN4ZxX/E5wU9MTP3+9ffev/gvzWzT5ZqMlSMcpH6k9dZYFdtqsZCVh4T/ZJJ/Cf5fp4MELof727S99dHmLos6FotSzYlPDpEk6fizPQjpweCnN0uobDDAmeSiUQ6iiXS1veLrpxh3tppuubWvXrIDX58CrjCnElF+dp2SuY944ywz/GcY/a2S53qzTl59KEdfSBvATxyzG+E9AcMYztB9/ErE8E9LHqA/Kl3fD/5ZfoR+3/4lhZaCU9i+98gpGW+/FTsnL7dbbbmOT4ssITGdHQLpr1y40PhGaM7ZqE0lj7fMYHyLMpPCjjAevvf5qe+yRRxHAP4g9sr30Nwhs6Rt23rSzfelLv4XNkq9ihmBj+iwFsQrpvNBDHPf4jLYncxOvOL4qcwa1hBWqsm5xst25Pr6QF/4vDX75nf9x9iDBP789/3VEqQgjvx+lfOPoKtdKPPpePqO/w/Ir5WjY+e+TY+d7igJGvUbfPzzH82OM+lRewxzzNvwcRB31Gn0fROheaF9pmWMIG4+UIR+0jmQzeB28jCcZjTspZEia8bTjX6OJPo3yqUXADstRmNAVU1s/GcH6lDOkhA6JZ2DGTDtth0sb8rCRmMa8Pzh8nEa9u732GmesT3r9HteUcr7aHRWlx+bqtcWuzSyw4K/fvjZ9Lap8F9PEt+XQ22WcIFWe+CvgePKpp9pP2fk8hzR1/abNOUt//fXXc23fMnY2nY3afZoFhvJo4A4qNZlj8EB1b95MDJXiLzR2BGdQGTyJqvN0BCQbNqxpm9E0WbVmEeHgwzrTQQN98jRvUZO6+RS24KXqWD20pRsPCMW/capGXS7mMMR/NeZx/KcQdxUs37x4qQmreevMQwIZofKLV/9u/klj/PHy+/K6VGaG+2TKt46OOU44n/7la9xD/yb4xVAXu9MK05wUKryq++Z7nE4qv4PhHAu8duZIu/EGF+mbopautkKllx+lrTIRFpzBs2AAVXpd8d7h/yPAX7w45P8rH//iDF4HxiH9fe9pbTgcAs5r17i1t3bta889/xLHQzCuy4A3C8HJaeybcMgpQqa+/ZFL4RCcmn2uj2b0P65aJ9l71vW66zazO7EAoZW7ztCAP1t77dKA+6S9CP67kXHIAUP+C38KygX53xL9K/7P5Iuy40b5P5Mw/XucFJU/Tf4f0OIC7a8Eq9adtsAiRwOTGnZ0cnsVgt20nvBwolQbBs60ZdpRwDT5CPxVpq2O1O5OWjawe82jWM01seA7k2azcEfLONDI2FFtJk/pH0OwCJVfeOH59tQTT7XHsWPyDNcLa29h2VKOYyEouf1LX2q33HJr24wg06MCCsoVxsxGA8Ld9hy9oc4KYC51/CnwoFTqdZH+z4h9mxcvo/T/GO1f6Iv/C48fufyx9icFrIQumB15xnPgmy+Huj56yjd1pXPB41j25lu72ksvv9xefuXVtmfPu7GFsW3b9W39hmsiMJ5gnN3/7v46p753D5sYj0KHWe3mm3e22zivrmFR+WzD+nUsrDe0lWtW0YYVjtFmqXtva0uB5xkXGGweWCn5x2Ok9rXuOsp/xXM9XH5X+6Pz4B3/PLp3vi86/oQfC0vBReC/svpfjw6wJmkHPjjS3nhzHxolp2ivJ9J/zkJIOB9h4QeoxLuIcJNF7ZP30OZbu24NgJ2LUcdF3Ag3l2NShz54HzCncbPREgxfHkVAcqitu3pdqL1799uMbXPYpFnWjqA9pO2oRQhiNBJ88L2DaJwsacvROnkFHpiPEVkXSYc4uuNcxhsvTnGd+cEDBygX7SJotx97Qwsw6OgxVW/fuYodY43PnmaX2N1or2zdwcbQbDvzzlXP6IxH/seNkLfmH463evZuNEJH8zy699BfNMgvH9b/VppPov1d7vb/yZYPGZh75oIFsCh+4kKHfs7jnDaBCOIn2mvcnPfoY09A59MIsumbYWC6iGgJKTipWWnhu6dUl2nyh+PDQ31bTvdmKl5Cf+lpwjEGGc2pf3e8Ka4aoz/CgBxpd2N1wE/yiHACIZdJ2Ha8gW/pkkUI5q+jr1vOcdFZWVM451FzMWkRsEyHt8X5oErJk/yspnX25UP5T3g+pP8Xlj5vo6fErv/r/YeVSH4X7f8or8dUshsp3z7dtYjD9Cuvvtzuufee9srLr7TNW65tX0B44prGtvyLXzyNzaVdGX9XDPkKlQAAQABJREFUsvbZvHkLc7M1WSO9e2B/+/E9d7dXX301dgLff/89+qil3EC0vX359tvbV77yFWyYbGtrr766ih8pPx5Wbmz+1bd/8alLBJEc/BfC8e5xIXzhmU8I/z3/kW+5SyvfapvyfCf9i2MHMSwqkQcvk5JdyN9ofVj/rKTjX6PZXZnlM0d0RjmCsgEEg5dRKMbex2P0AHZRxgL5kHH4n/5lpLghIienG8tgrNz+YzzGRy+/0pPaemQAq7qFpqmnYfyF2YfYsTxLSRhPh9LqmgqgMBhpzNJO26v39qNy7izDzvo0z7McDp7BwKz01B2FaaoIjOJjivKrvuf/Wh+TugDUSNrf3n1P+9P/90/bC89x7pi8d95wQ/vm7/x2+zwTxavZdVnCpCMLOiro0Q0nP04MrXAm+FjEnzF9duqnUMRdnxiVoo6LOAIyBzsNZJs/axNVvmDBmkSmnY7baV7RuwBLNy1cLti6PlVMpvIfE/+W27se/vqGMlRusFgcCbwSylc9MnYrgP+9QyfgDa6+pb612yyLlLBDvGnAa5T/enjzBF/Itdsc1CUcQD0yoGRdOk1TpRgJuDQabdZXAvwZXD8T+oslkDxGfz4VnPTl9wgFT+LGW4ree+/97DxOR0VW+wUsifjznLGTso5Xk6kZ17cc37cf59qzWDAs4FYG/Tzval/hIn1Wz5d9+WkWXZ6/Ifx/IfqLClGg4ES0yv8uSP2bUKcf54JWZ/Co87tQ6S+U7PFszEFg0ViVZzV5XPTa7ozgv3RI5gLNDFXD5SwTT9ufmgsucA4ePNiefeaZWNm/F4OvL77wQnbAl6Hye+stt2SydfPNN7VNmzejabQ6O91qlqg55uLRnS53wxRGl8COkqrKlN9VtKrENy+OPRcYf6aEP55AcxH4L4T/z6r8tCNAE9zAPgCkg7//5rNQ4otIIX4XZkwFYE6gba+HMfh5CA0Ddxe1mbV4ybJcS0vCLEK8oWgfi2VvLvqP//H/47rcx2nrJ9ilXdM+x0T5Swi6fvu3v9m2bt0SNWzLUdicLRH5kZIVkPA/dExNxXH3l29jBaArG/+fJP1DB36OIxg8ccKjBWJsgv6zzv1HfR0tDhenMzhe4/GDY0cxzIrQQmGlxuZnI+RQjd2rvX3OQUByGs2AUwgnteGgba8jh48yPnK0jc2FUwi5bEtqlYhvbZSYZjZh7yMsmQn9XVSd1C4StJtnfsT3mPHi3MzDbRrkPQehuGOut/bIiC48z9DmT7HgWr1iUfaxYk/o16z9fZL0h+Xjev7vvz67/iccSPeguGNYCwUoGoKPU7hnEH+8oWUw0fYffC/857XWzm29yME6Oy8bdjLd+6XQn+LoLarQYbVSxzH8W0HqXoV2EdPp6W16+xj8DeInMCY+HrSLRRxf80rcGheZ+7nYcMYCX7tB49g507kLsNU/+ib7UuGlWLyTL4Fk35Wfd795sexLgd/y0meTvsvWkqz/J1F+bpxjjusRvz179mQ8fnffPoy6rshRSzoS5l3nsrnxzp53IuReTtgKhKhq4z/59FPcdPd4e+ihn2W9Y9j2bdva7bd/uX3lDgUmHPdZtfa8+XfQJWoEZsSNgIjv3+3+v8awgCHBRl/8mNKdB3+4rYs6FshH+IoHr8V/fZZjEQvJfwfKH2icjNQ4EMEGHRqGbz2og+ckmHv/yd79dz1Hf/sUPvtY5Tcsdfg2Gjvv40kGwZO9++96mp8EtEvxrXd9LL8NsVO2E1PiO+wQFBZU7+OT/yFyTewrR9LqSZiTPI7VsqDlllMWuqcQmrij5WRMabfOsHIkoMTeDaHmLeV3IUYxKq7v22RGr+F67bU32n33P9B+8lOOgqB9svutN9tKDCft3IkA5etfa5/jvvGN69e2ZStXMhnBsjj1sz6Blkxm0vEEBjrX2nEriBTCCKHlqMTQ19J0Dl7lxFfFMbYpjem/LDD5MmZvR+BS8Z9sJ8FfNZJfCzUGD+t4ZZRvfSS5k7qzWQiCGSpaO0/UlwgFQQ/NSPubRH93xEW7U3xZzYmf+aqt4i550GMc0/HvNwv/tpwek+JS+OG7/tnh2QlEGJqHuFeIOBg8xGlwy0v3lFZxesXhQRrxX+k6/idMuyRqSoQmCmJo5CoDpx5kZO1+0/i/etKp279tQFxJLZ2qroPODaypWj9Ae2J0P9LEgJ42k7y7zwqmr7MEae6EmkcmnmmKeEdwQ2Ya1DNTd6Df49jj66+/inbJk+2+H9/XHnvsUfrtCRbea9tt7HB9hQnXbV/4PAvva7NzbaY5Amk7NBcYyz7W9uck1z7V7TLp7/Ej+8+h872H0pCpx59E6ZL1/X+PhB4V4zldIe1fXATC0d8h9EXEgl9WqFZSEE0e/8RpFunusLJY8DYFU0RbL2/yGT4UZQ6nmGj/9MEH24/uvhfNk5+iqfJK0mzZvKndcecd7Y6v3NG2b78Ow4oYcUal294i+Lf/pKPVWHS0lLoM6/hYjXI1/lc963cUEsu/QvAPVGLpkxh/RQNDTsadadiA0YXHfRLoX25p6OZBIK/DaGE2/R94yZHSYKjaX3bcQaKbDGp/pY2A/7SEdLSVv/TP7jrPCcqyv3ChaR/DugnaIR7lw3ryARG6MF6FX6pMJGJ3TE5YyMtjWbWGlpK/Zu0vEH0y9A9OxaK4FlXgs7Be6C4M62/Ip8H/liB94LrQU8pbH/uY+su7lTOqw4k05n2GmxrEQabqOjvz78pjvNaXr/+V9622tSy82v8VhAEygkqDHEHO2D/5TwBJ6Dhm+5s5zU0f07lJW/xPjE+k/ZNd3GdNf2+3UTjqsRpv4DqOIWjtvqjNqeaY2HL89XieV4of4UiOVxa//MorWQ/dzxjwwnPPZhN546bN7at33cmRnK9xnfCN0YxLf1GQMeco7Mlno+u/FHKZ4A8zF0t84viHxXD5Sd5j/If/MCTB9TPaZC7i3Uer5+jvSKIw1bCUK738CE5kGBE3rCwA9dAOXkcAHglLxHwng1HcV8puwBtF0eT3z7r8frHa1yOLmFS/74SAil6mcCKEdkDpdtIJJbAzSjUOv61N4jPxYiB2l/MMAo2ZGI3UdwLL/zZIb82oyaGdd4f1j1w+RYhv0oVG/Lgzq9aJE3XtAuzevYdJ/iPt+9//fnsStfL9+99N2DYkqnfddReGC++MpHUZqq5ex+eREes6G8No/osjewcW56QKebx+OLd9EDgYq0I468GfAHYPd3WrcoYR0M1iuyjJfvTn4+G/L9+8zaV79sSykL4yCbsyyi+e80gOkzUn/NTXxXpuyQBHTgQzuvMeUC7If0V6d+TORvpWA6xXzxUq4CdRAoO5s/abh/8wRfHFgP7ioeeKMIgIirBRXIkn0efkWXyJf1/Fsew73v4rn9AqIcQhshOtlAGvZ+Hvh3+qdbrwN9+uVjzG3G8C/4vvyfD3/XAJrcRW9b8Klns6FBZH+1/wr+dI/xdU27HqbyEDh4cEJGiC60KTDuLarxnPqBFuEM0Jlzt17mRrT+ElLO7/GO2Su++5u/3il8/EnpMaDdov+cbXvtHu/OodOVKgpoP0l2lcUCtYmekuu8X2cPAUVsuL8KQrl0fcbwL9hf1CzjYovmxFaSWSbaz/k4R4Eoo3SNRW1KkY2nVMMihNlx+P0hjPhXRwTfubxqaF+D/IjvPTT/+8PfjAg9l1fAYtIu2frOK66N/93d9rX7r9i+3GnTcysV6T6ypnonlmxRRGk0Py9TdNvVuw2M+mzgWARSdeKmWa8J91PJ//jan7u0Z/YbH9CXloxndsGACrBqsViNT5f9+Jyi6x9Kt/oU7QY3tzvnFmgnTsmrspo+ZIzZvIA1p63bZCmhnQMxoq9rV4zGKuIk1PcvRCmzS5KY1qhVbiXNxX7SCBdHDxSMMnKIvLMIwJ5A0paP7UFYFmhD6U2ZNUyg3nf/ib9Xn9jzTG36Lz4wd/v4b0v/zwg1evpwvCxbEol7D8+ZTOODcuz2HEzPZ/xlu3DM7UVMFJzWenzyiutL/ox5/Q/bLQX86BrwUhsBRowiJ/yqVCPcHR/6uw0RSDpIlL3am/449QOP9WiJzUv0b8d+IE2mn0JTOcTwFexluEKYAcg+s+TzEuHOHI3zsYeXWj42/+y4/aT9Ew8cieminXXntt+/a3v52Nj5u48U4j0r3gXfrneDsZudn1m9T+7fuKd2SoYh0YaeQ1jUwUd/OoCise05PIFaUP4ImH/NflMxIw9krX3CUvHk/gSF71OvI7Ena5yh/ROBmDZWokDDDQoXgMgD79CPoH4YOXLt+R7z7Z5Od5UUY9fsXyaSCSaMxJX4qwFCntrr6Sxt5niBDLZteQoIyd52dD7gzRRHBAtphkw8spJu923vZpdpAWaIc4LMOydUTquamr0bB8w8fh9+o+Jxkzc4WZzZ8YdKJOFB/iekyv2PrBD/+m7d6z294GeyXr21e/9vX2zW9+o33+81/AyNoqhDsc0WF33N2eCSYyTnSmc+WeVeE/PywsMpiAG6uHV+DPGDWEIXij49ZbECpxYvPOM3lU/QkdOssh2JgmujD+Da/0w/L1w5FYfDqhuzLLB3jhp+Lipl84unjzb+hEsMDoVxipZ8Vxsg51RvBfKbMTanwGlisTfur2qdPfHcqaOg35T/zY6CxfXMs/LJYRFqrlmluhWICpbOB8uibQvpjO3HSm6OlR/OcNPZ4A8kv1UaPMZKKvwblMW/hOKhf1KoH/xvM/2BihfwkNXbCAH1Dbt/+wvj/6g8FqG8Ek37rC/3j7JwOiXLj9VwGs1ejHeq0PppjQXdq5cNMmxi9+8QsEJvfRZ94dzQTb2rr169sf/IO/z7GO32k3cPxx5Uot7HOVI33sKW62sh17Q4eLRe0rnMbfmrvQc+Fn32zfKYARoIX/5KURNwK/UPy69X8lTATekHGUliM46F6L/kQdQ9FoGtXT6eOlN83KcVYa2v+58D5neyuEF6ckH+jQdQFHsW/x/AsvQed729/+7d0Y+H0YDaMTuXHlixzf+b3f+33Ou3+pbdy0id3MeQjCyni6dNHJE+knrBL8M+i7rQ8lXrnjDxUcaX8Bpv8BtL79CcXF+E8Ii515U4Dk+NXlU/AjaIKI0tFM0b+jP+z7X7z0I1BxlIIKy0ubhH65uYLvCeY0zr8UTNo/iPPUK+V0OCZ/+1/L1kfesvxooXTl161xVsSIPC8Cf2XBogz+uRj8FuTMbbz/SSVS/q87/S8//CUgyI13drS9ox+QFwr/CBkc36G3/HMGAYo8MjPt1a4b+pE2R23DFDJTnxG5yLwXmH/18MtONUfo0iV98Ubag9+DlyH/hQ8roC+w4hG/VgTyH+/nlV/5CYlHj0f5z/HLI3MzXGD8Gs//PK5X899h+7ONe425x/xeevEFbtt5qP2XH/ygPfroY1xbfKStWb26ffFLt7c/+IN/0O68406uL15d4zfo1I6axqG9mtljfJJM0mvn6nz8h5jGCJ1G8Y+n3n9n2/9gDAsgBcs4i4qZcfjD/71Xn67DTbwHSQYvwdGU/D9I372MJCmfUY+ujY16DdJfqP11FU2aKRMOcsjLeVFGPZzH9as3qR6yj6Knex9NMznXhA3TGjzE5TBh/9Y/k83YzzCPPv0g7uClTzDikdf69lfXp++oFL8+ReLYMoxVvVMHNQhnwBQd9sVO2owVo47pRPVzQuGkgdm3yY2XrOzuut2wvBlSfWqF1H6FqmZ2xqptRy3VDjulmElK49cU5ZK1X135XYld3C4WkSbOIDhxsKfDtGOZQOrqNVxO+Exz7OgH7cWXXm0P3H9/+8u/+ivO+T3GripGDZcvoiO5q/3Db/8BVqm/0tZdsyUaKNpgsWxrpwagJakpcRRDsYs5fxwPPPUfwm98UvEf6BKYyQex7P+TVwUnXReB+GLO0j4q/sUQ8aWFqbqMh/ivehl45ZWfmkMr8MEk0fn98MYQcSDeP5z+TjKlhztjDB+Mk2RIciejdoA5+3pFwp9qhm6h0iW1vw+jv/iTP0bpb9sEV3gbSlB4kkftbPtCYOzQwE+wfwm3kspIlR++fnR/9e6ZWuloWEKM28Uv/pM+Gp/rJkOJWfGrQtan8vxo/c+HwV+1tKpXHv8X9sSctXTH2farDZBeU0DND2e7NTV2p5mFkx0I/6v/vRD81cd2mE38yfCfwY6Bxrnt37SbYE5eB+xk65VXXm7/6c/+jD7yvvbO7t1tFjYYbrr55vaP/6t/1L75zW9wE8u2Npdraqvuao5hU4h6ufDr6a+GgxM47Q45kc2OoKt1nCrFrrZLe6Gr5afC/wF9SviL/7twIfkMy+/bnZxeOBQHo65C8svPcPwlTpdIv4JBwUXXjgnu8W80XbQ94SGFmCmXdMeZVJtW/Et/tczUDtq390AEKN/58z9rP0YT5cD+/dgpmtc+d8OO9u1/+O32+9/6Xa6TXstke2347wz0PY4hdhf4XrVr/2vG1gYKxwW+1FUOI4R3w6ruPPtY8TClNScugHz4/ONC/P9Zlk/rFCYkzUdRl5/DFcEaXNVPJ0TexGc/6PjmuxBq+0TnHEUPNS/ZGw8+HNMmFJyonQLC3Dk2zSzydUyrxRIjpMgUV2aUvpbS0v/af8cX+ssf9Ph8T5+hTZQh/jUgabpp1MV03vJjePGDY/KsbtGUKk4x/7sS8A+cVDqY6BjLr98c/hN+/8OH0rxvf2HAwovhjs22vxjjhl2Cr+4ZBuo9oscB9rr2p3fS+zDPlOFH925i33V9wyYBvnF9+qQzI5MRcqH2X6HmWe2/zzppkrjCklFyKv53bp7NWbugzB27CqR+llfwm19fzcrTeNaLL+D7u9T/KFD1IgX731luXgDCseNH2+OPP4GZgvvbd77znfbC88/HttH6dVe3/5+992CO68jStFP03nsL0Fs5yqulNqOJcbvfxsbG7v7MnZiY6YgZqVutlleLIkXRe++99/re5z2Zt24VCiQAkiAI1iVRN2+6k8fkyZP+n/75X9L//b//R1tq30rTdM15wZ+zrtBLY3WrZRzargFX6R+uJoduE9TPIW7P/t8wrf8SB2Si0Ievyl058OWpedgZ3/zyOCu7GvGKq7wd3PQTIfXwyl05SoKah53xzS/PYMCXjFj1BLBaeVwCK4woRhVUORyj8dPq3/a7eOZ386uJH88Sfpm9gP9uPHlDgaIgC1Zt8VfFQeFYFWUEnFY5kAX+1lQeRnAjTocYBcV5IHSYgc+hZwx2yH7TN+odR05bBNPf8teTVSCRI/8SJ4dwfzg3q7AHkHVs92Qo0jHB2AceW4bu3LmRuP98sw7J++zTTzWr+td05twZGT7j0/IVy9P//J//K23Snv0NGzZ4FQozpix3ZGZ41GiWpNP9VHmiSC5X9ZOLFeUEl3o5QUDEQdNDq1J2oigqs3d9p3/OQq+mhwbA9FdmwBiC8Jl7u6vlhmNYWqxiIgvQ1VMfIg0HlnFwWfVUOFQUczrLn5BtGiWu4R95IwvkBT14vSz0zzJW0a7gL0GryZ9nwHPQfQkgnXcMLW/XEalMOnWskCmTTz8ho9Cx4YcAmw+awbLYQW9oneH7ADvie2r05Zb/Ov0hFvKP8caArztEIl0sow+DGLLR0fWMs+kPHYtD7/pTyX/2bIkX22noJDH7eN/Leffv25++1ZkX//4ff0zbtIXjyuXLaZZWk7yn0/X/8Ic/pLc2bUrdy5b5xhYPfqnTNZqOGLICfwsoublBh9nxkA3NmKuDTf2L1SboX9oDbj5gMJsOJToS/F4O/ee643bTlUE0oI40vRrfhba5DkHmwk74yIBHeZANd6wRJWdZaEs8HVaqAbISXpc/2jX2yqN/2d4K73doS9ann32aPv30M926cMDt3YKFC3347z/+0z/pLJQPNHs5z/x3niqf674LOPTbnzr+hf4Dkz9aefFE+HtrGjIt/RZ5cax2qdvU3agr6Fa2TACXDou3xBATmwi1KPev4scIt3/ijewZ6px8FUd1FltIPPMEkQZcXP8ioeCqEmWjxDdjua2LPKXUFaR6Kb+HD+44L9t/6BxyVzg6nxVktL8jVIcbPKWCUuaG/FWCaN/aT6V/ilCXtHzrcZlAlPxKHH29JPX/6eMPDS04euup6M8HiiBbzLQvVgzyy+0//Pc2HiW3Llf8sL/hGenz00b/FNaVKNW7gl94S16F14r1WP4ThwJRf7CzkU35KQ9ykYhawu2Hd5b5KCL1SHVCkXyIMna/DkfuH3xBcWbKvDyRub8yVoqjzzqNqriUgeAcoUrLt57H4v94+FwZTvvrc06k28/p4O/Nmzen//qv/1J/5k86ouCEB7XffHNT+vtPPkm//d3vfAj4DN28NYqb+bREhGJxTfmECUyCiNKiFQdDs0VnjAZjsD9YvcgK0qBHIFuhXTkCrer3ifn/ePxdkmcEv9hfTbyteFhje6/wRYkgVZCkNZ6/i2d+N79qQDJVhzB8D5y4mLVC5mKHt36hR0G5HlbcjbCGq4SVt0N6BNc8BhO+YAE5sKLR1VeGX3AtWLuE8izFw/Yraal45OQ4zg/FR3aRS8QLY92DI/LAfPYABB0CWk2abzKvMiklIIzYWSEpTjQARIyw8saHPBm0GaG9ef6Woc5aPr7v6XpijbsLnnJUPA5WQuns2b0nffX1V1I8n6Xdu3emyVOnpblzZ/le9N///g++nmvxki4rkTAmVAZlTpnitxV/BeZGKkIUjUfRTT8KJnekJgCPgdD/xYRP62djnQ6WHgw1BrlYzo8MmBowyU/gaJLJafqFl90MniB/zGwz+00QI+cMyHFYLI2wzz0hr5L+Jac/smtCiZisbPAsoydGYxUCxkjMVjMDSb2T0Y20S6a9Dx4iM0BSZDYys5njSSs4mA0dRzU8pfd2AgV16G/6QxtIgQEbB0D6w7SL2WJorjqhiMSjUc+k01fPpxHWcLXGYjaaztiZc2fTLz//rKsMP09ffvnXdPz4CV8pPH3atPTftPf5nbffSRs2btBVj4vTRK08YGaLAZEYQJY8qA4zi+0OodwYWoXnlBY47I++p1PBSetl4tRRwYb/o+nIUziQC+zinSt48TWhFG6MoIMcJCEPan78vjj6F3yK/Kvw+TF24S4I6isGpQLHVvyhP4Yu9Pd+dGijNJYR8QL60wliINRuEY0ZSuCzNx7SRVsJWCYFYrUBfmx55Zae/Qf2px/+9rf0p88+Sz/KMGc1xLSp03Xl9Jr0j1qB8qHOCVu5YkWaNHGScsg8IjtKApMworO/vfEiyMi4CMSM+EWXZPwVNYeFy0lKer2HCv9tzxhL5J8hiMCHtkinn6ujIgoUWuh8NM53G60BDyLSBsZgKB1E0oqG4iv5wFe2WKAbWJFCvUFPu6MIjUTbWGEZbngMCaMEDOTQ9onXugkQW4cJJZbhE+0hW6Wt1+EZ+j34z0HedMYUJaol5aaglCv7ZfYMGfpTtiJlLqPKid/LIn/iZsWhnvhDm+Ad8uEVT4o0WrYRiXw2nOSMdv0VNc6ykkW6oF1wHVoOLv/dB6A8up3KfKRM4KByhf0Baym8/kvGvWJK9YXJUvAvbSRumiTsF9xGQ8mMHh5yB2YEOmW8nzP+faE/OuG+btfavWdv+uKLv6Svv/o6bd+xPZ05e0b4j0ivv/5a+ud//mfdlvZeWrpkia4xn6pzG3WLlu1t0VboIhXY3cgFuob+Eb4+zF8u6IiuYdI4YkN3uZSW91DRv5QNaQGfzNbgowtbQuTFU8qfI7bjP35GkHftKXmXdy2ocjbCGq4qMDsc0iO45lEI3JKwxCjvlmB/NsIartZ4DukRXPPoJ/zGwEkdUi2/lpLVY8ndGrGdT0sSPpWsdPzahPbMpCeYnKxnQE+fnhBopDlMbIRmSULsFKeSpixl1lvZU5lmeYyyAcQJMBoQ0pKYseJQ2obqWpZLFDXOybyEUPVypK5Oc1bOLsO1Ele+qtRVoOGrpC5Ezs8ACNB/d+bCPya+FZegjANKgfxsXOjtmXWuD9Q1rGfOnk67d+1Ne/btTv/+7//mJet379xP02dOl3G4Uh2Jt9Lbb72re+PXpAWLFkrpqGPAIS0C/ED7+5mtYcEtxg2KKPB34Vy2CodMIs8oa1BnpE7jp9zQ3w0C6GI8ZWpGu5ETgQvh+TNy5yPTvwp4PHyzSsmeiP8uQP/h06BBD6Mh3tAvB2PLELwivOBo3hGopwf+RWqhth7SyYWyD1LYV/xXZuFs5JuL/Tzwd1mGAHzkn735o1RBvMJEBaOxtDRCQGgm/RD1GmGRSwIYtAWBHIeI1H+9qG88iDCs43F8vTHQQ5750B9ZPon+URZVRoZPphRgaMt/LmJb/K0XZBAyYGLjRYZfXEdMh0sz2qDY8tChZQCM5dTonjr+oaO0jVFGEstv9+gKYWanvvvuu/Tz1p/T2QtnNcP9a1qxckX68MMPfR0h+m7mzFnqEE/UVh11tnK9Qg7QP9Q2dBSy4Bk+ZrMz/QkHPte8+wYdyqUy8Q/j1zVWLLKfyvuy8d86vvBPdHhU++92wPoPohLZZHZ9ctstekJ/OOJ6GxXT9GdlgiUBhuTkdKBxwwv71eofclfaH4crAjObl65cSkePHE3btm2zgf69zgo7oTPC5s2Zp7NPluiA9Y9loL+dNuqAQfxo/8j8HqvU5EL+iv71QJA86fB4oA25UhPqmy/k760rens1BXgJP8XMHSC+lbPCKzSVvEIky58jCfLj2l8XroZ/aUnsLzjOF2UGOgVgG/geSFbBTDOnI61wgdZ6qMPYHCqQsgwotj/wkv6V8SPvsL/kZVDeFqEP819v+ALvgBG2BRHx19sP+WLXNJ7Q5wIt2IR6pQu8cb45rfKI+qf8wRV5Ig5uIgLkMfhHPCH3nOjfgQ/PG/rXPBPvQv7hC+GwVDIAXyWDUf8lFwyuqf6NwP5WvKwxIgEJHZ/PLAskxzvHiBdhFu7I16EFflPEnh/KCP2nlg4JdTi5+U82MWenNcufQnIEi6jLF6uxSB31xq5cCuVaMizQ+Y4oLvazaH88uSAgbr8hWIFJGYpbb2+P0+qz0v9ieyvh6D8m/uivoP+8Kk31mMHPa9eupr9t/jF9p/NL/qZB7T0aPLl69bKuoJ+WXt34avr9737n6+UXasKDSRCvGCkwAY8uyvhH/0ftNEIgb/QMfLDekUfoLuiVM1A687+WnxPCfwWY/jnzwdK/Txt+a/tsfsE3niI38VX77RnQ06cWvTgV6VHtfw94vWbaM6CnTwFaez8F+M0DJ84wWGIwrqBISy9PrZThrHmQpHz6zQ9eCJocyrYEOyACLasVxGcEP/RdKBeXxcVR5RE86kqB71UANtCi7C4ngdbLxHdCe4czlDNCEZ76VvyIJ7/sLgaCBzlcOQnqB/zIqgY/Kj6/zocyUgTitcBHI6M42ELgUVUZJrfUsbh86aJWoOxM27ZvS1u2bNUhidu11/9UGqclbcuWdaf169Zp6fpb6X11MqZpZcpkXduIkgMnG7EChG4yaP3YSHRHSBDVofcskYqEYQN8ZpOsZFxGlVrpmYnCj44H373S/wnwt8FnJj8B/wcIX8nEM/AKQuViyDfTTRFyUPhZHqFo7TGBaWRUfjJ0cN/534A/+PhT1qEBH7pJ0lSeUlVdb9QRDvpneurDB026MsEDEoAE6RRW0d9eClX9cxpFJYx4NsDl0EDJ0ME/EDeulFNP4J9lIryEC99GIPs4YuAlZwP/TC/whwrKvsIfcgEDQttd5L9Z/ojvDq6USDG8HmrQ4650AiSP85ri1HwMKW4N4PYa8vSP3pTCcLORdPbMWemyn9JPuk74py2bddjrPt0wdlGHgE5Wh3dDevvdd9KmN96UfluWFi1cpJUl0mdK6yxd5MHD/3nT/1nDtxiJppYQy0OwqtAaLpp9eFhgxM9BlL92+HOeyeWr19KRI4e1InO3bmn4MW3ZuiXtldE+QbfRLViwUHK00Qess0qJVShTp0yx/Hs1KSIvfOhQgD8rX1hNwSHufNNZp1pwOCHnf7jNyx14DxhAB7WdDBDyzWAMqxOhU0UmE/DJ6187/B9Jfyq5ylEenK5/lMyFk0d45ncUOvS/I8fivZyBkyiKdVLxM/8JqT057xLfsBRc4BuPHMd0ktvi1KR/lAB//TX0f8TDL64sJqxT/wdL//db/sTcMiAJM/Xl3+A/Axp6ELmK//hFrLC/GJwkNoMd5KA8zG/c8bTlfwnTO+QKjxb4OU4zfCSUEghSFryQ2YjsQUU8tFrL4XK2g9/UHNcqS+SVc5RQ4zJ81IW99WNa1ODruyH/fcQf4kRWzpcVz2QceikoShsacQAgm19tONR2nNz/uHdfZ5VokHms2nD0nvUlZXba4C3nmRw+ekwDJT+oDf9JN6L9nA4eOqjVnPfSEg2QvKG220cLrN+QlnZ3qV8yxeeexCCo8qKcFOUZ4g9+INksfwH7hYJvQolWwTKQCsLh39tjukZgOGseeJdPv/nBC/mTA77EC+945PGiwFfdRFuo3C30afaqfdWcBd8Bv0te5V3LqNmr9lVz1qL3ywnKoBsVCoe+6NiQd36awMDpwlHr5EysWnySOc+cPsuh/BrkDaihaFDUDBJUI8tFg7WDX2dQU8Fy5AJT71KkKAtQWuErDiPa8kdhoVmYWb3/8LZn4C9cOC9D8UjaqW082zQzywDKoSOH0o2r19N87fd+8/VNadnyZektnYWyds06nYUyW8vPdRo1eUJXKT5DlNurUQDh2XxDVBQpV8GkfBiUOMoBi8y82WAkXGk88gturfQnMX8FWeLoqXuFux3+kcjhOO3QT3/4TxqnA2rjqXuFuxl+zGzU+DFA+JltLxz+udaZYKbPAPE37evEziyoe7Wjf3Ce31wSw9dPVf8jBtnFyi1y4XFEvfWtuI4FE8KR380iEfCjehMhohffHLdkW8E3MP+UIH8ME/nvjf/oHwZyg55Bg6IzfFuKBrTY/gL5H+jqdc5VYPktnU2W/POYvupccr3s0aNH086dO21sMQi8d+9eb52ZO29+2qgO7jvvvJNe18Gvy5ev0AHZsf856r/qKxnpeZno35B/o/5s8M90bUCoEbmJ2KK9BKLJ6znJf5FHjP5bN2+n02dOpx2/7Ejf//C9Vy8hV6xmmjdvXlqzZo1ligOFcS9SW+mDSZWJtwVKTj1pINw8CKj2kW0onGnFyirO/6g6DyINKyFCEnGrXfRkgwZOMBgI4A8i1Z66V7ib25/e6t+A+A9c9Fbt6Qm/Ucx6GEn8rfIzyeRyPZX2N+vmkn+bN7B5CvyKjpBVNK/CcpzwcEEjEfYfiVsRavGK4GdI/5cePjwJfplHTfTHxqR61Okfdhe+PJCPA6aJyGBeNYLg0IhFnHhIk79qzhJa3sQoJQp3HX6Up8RweMnL8Gn/VLsfahWEvkuQ867pP5wuSh1YKUAFPeIYf9evSPJU63+Gj81Of6roNB++LPBxQQGjNtyABm6CzgQq/QO5OQPsvgZUYuuqVppoax19CG65uXRJq/2OHUu/aLXft99+68ncI4eOpElTJqfuJV3pzU1vprfffts33C2Unh2vQWwGldluGeeKeRhK1OgH/Z+K/hGamRf95v9zhl8mNoooFTGL79pXzVniDvhd8irvWkbNXrWvmrMWfWDOkld513Jp9qp9ZWcMnNQShLMWsYS18XKQ/WuBlbNyVEqg4RMChnC1f+oxc4w2Xg6xfy2wclaOHvA9ViTgVkI5+6wHrJOiKW80wkQJ3ZWrA4qO7HkMpgGrAlZpuAauESsS0jFzCDPfcvULfgZLDihZDlgLpZgD1BjUFw2W/IFisIpmqFJgnuXVN9eY4bbKUXpGdU+fPpN27dotI/HHtPmHzenE6dPeyjNZe7s3vhYzba9ueF3L3ZfpbJT5aYoU21gGUaQvmbWJsRRBd3nCGLQCpbGiIETQQ+eHQZKybLbQn85TeZroL89IiaPv+Dfo4GQD5/8A4RcuhbEYPK/wy3mGHBSVX0IVV4X3Vi15oWOr1C8Q/oNL/2yQi1LFFW8TT7QUje3Uj4gedC/vKCm/jgK15Sj0pxY16hs1pkDICZxf+JEvudTztw9xSubyqIeXlLXgGvxGXuTzIsk/+AQ9XOzASToobtkQXqrvQQdWnbGMNmKjl7y8WuE+70D5YIzFAPgIndl0LZ3VmU0MmOzauStt3rIl/awVJgcPHU5TdTbFAp2u//prr/rspjffeCOt0MqAiZMnWf8wMOOVJiqHCuDO68tEf0SI51nLH/q9if8ZpoG3/Pz6K9tdIn4JatS/wZN/zstALHweB8v7JX+s/jh3VufkbN+etm7dmn7avCXt378nnTt/wUVFtt7WwNzbMu6Xr1zlmdEJbP3SeTcY9Rj3lnc6EMqTDoevw9Tb7Z0AWv4ki7ZT9O3BQZUD+tV1SXy+OO1fE/8tcIFRUYQlvP/1L1JYip0J0lNa26BbXb6DiiHzpqF+ii4OjYOMPX/5y9RxEZ+H/A9t+JROf23tH/GvZhMTszf+Fxzr4f2Xv4ashZIbiPwJKgMn2j7kR3iVsvFtuQzhjPDya+TqGEKSGnylqYeWLJR9JfMl/0a8vss/kx7QmkER2mlWl9BuszCO9tQ6DKeBUC613d6K86tuzJLOU1xvp9X2yDPSqwf2709bf96atmiV6M8aODmtPsdirS7pWrI0vaND2znHZCOXV+iKYdptBpQfcKuoJlH4BkLYEQ2eNPDK5VAZnhb+7eWvRn9g6Q+8y1vOoQdffGg89ZJm3zZeDrF/LbByVo4K74ZPMy0acIurHnNowtfAibqrVj5RwDZFfiTi7eIX9Hvm2C42CkOClvnWPkYbwVPmrcLYDLd81XPMbhsnjXDgy44xnlY6WEt6okg0wXKr5cLbfgRTZl4Oyw69ej6OEd7R+tmNcnYHWgddEKOCr1yLymyCT3zFC7+G4IWKI0tCm5/IyaVXMOXHEARXhdgIfOiDY6XvdIc5p0ir48JyYGXlGS/SaITnyrUr6fDhw9pXuE838fw5bd/+i0eE2Qu+WArt7bfe0TaeN9M6XeHItY1z5s3xoUwFvs8tEDz9l3LVnxSnjUQX19gbfwsBZZN/dIqI0Ib+hDuEt+JHAiI3PQV+oVrwMOdfxYQu/eS/0vYfvlLQmDth8DKK0AIffDIv2/Nf4eThIS5i6vuFwF/ouw4NFv0LHBPLpLYxBRtEtYoDuUxR//AlXYSGK8tf5UdWme5ysVPZkMzYCOOXJ/jHOqPMyUHFH/iUM9d/yjME4aOT7uvMI+jvW6AQZtESIwyqQVYwQHegR8KPA0Lv+SaUs+fOy9jap9UlP2olwA/STdt9+PXsOXPS0kWL0ge/+UC66W0bXPOkl0bpelMe8ucvzkaR0cfZFFrVgv6rzzxDRVPQtOMrHqTKZdFvp/6JGJX8ZwKZPg35i5U8hWqNONSg1vZfh884QtD6+dGflSCUbaQGNZBCDt1mixi4gMkdrTbZt/+AjHy2g232Cqfjx4+ny7qhZ65WoXzw3vvpQ8nf2rXrtK1nQZo9e6baxfGebeXgbhv6QvL6dV3pqwNMLf8SdOTP9NIbmJbJISZ/pVvyZO2PqJjbRFd70bToXwSgif/ISeVHvFr9yzUxfCJVpG/IX3wX/R7p8Svyh/4hU+y/mJhogT/E6N+Ev8uOT/PzIuj/Ov3BoE/2N/H0D+62lz80Nvxr4X/b9k95DYr9p/K0hS8cgG+caORU/1V81GmR5NL+1f0a/CdfJ44fuS0JevcN/ww/p+s7/aN8dfqHjlcx9ET7TSFoy10iTXrEwc/3tAWHiQ/OIbl48YIuqrigYwJ2aTvtlvS9zi/ZtWOHVuPdSwulM9drG87v/+7vtC0nVohy/hgDyaxsiRUmcVC7D3cVgQgr5Rhs/APzxu8LBd8KMGRHXGt6mmQth7Tza0pUfZSYeNTdJULIn2W41xhF1uJdz6ldjiXneNdj1N0l1sDgq42mOvLkTP2qA6j71+I5Tf9/qpzlkL6olMOgwhfKYA18bwfhux3+lqAosZWZuStDhoQ8BFUPfhG3rVeOW8WQ4+EI5eUlWpghOcSv7O4NPlTT/8K5gBy/DRFTKWpeKC+UEYfTcbo413MxcEOZOWXcAyZJHQczRQYbI8d6yonUug9Q96Xf0ZLl62m7tu988eVf01dffpUOHDyQbmh5/MQJk9LqtavTGhmJv/v9x6mruyvNmz8vTZk8xXsbaRo8WyzFRiNB4Qr/UYBxSB8dIxXa5eJNvKBFE/0dp3/4G1imshspZTtg/g8IPpgwSMbhWVF2wzeSvcgf6Cs+hIAkNGoY8PCITiQzFDz40oCVuPGuAnKECCe354P/YMMXvIokmdminenvX7kUjsRBU/z98BKxg55Qyt1mGTPIa8S1mSN3o/6JFw1gSpPzyvCdDGDyJ+TloL8QzfjHuyf+UN1LeqGMB3VNIvlGJ5X90wxscJbSKN0+9UAEv3XrRrqmcyfOa0vh119/k/7617+mrTrH5Kxmq0bJEGPbxOuvb0gff/yxl/R2Le3ydcI+h8nUz+YM+lCDJXSEkQEOFR05Qvut4c6w1D896W/5Rzj72/7B0EgmhhU2R/7lqxZgL+t1ebp6ESgH8EkVTw6x/115lZDwb9J/OYwQnoAcv+UrQuq/EU6apvqXU0fNVByD048KF3U66vlDDaIgf2PHcvVuAGWpeRnguHjxkm572K2l5d+lzz//k7aH7UuXLl/SWSgTfA7Ku++8py1ib2tyYb324k+1P1ffjs6rUDj/xBMMdJwQQd7omEr/UCRwkP6nzBSVcujB91H6JyJELJI04T9A/seqyShAr/ApXPX0At/EhPqZqH5lt9OG+7H8z21hoYaTBsjs1Qv8HvjTtpJaP35n+HRmhxD9Kdvz5P/zhy+JyQSATQOTf7S9aoPygctmuNkdPO/Jf+IAKWSjmf6EtTxRsCx/EUbOzfVPHvKsyz8DLA3lmMuSX46c4edC14BGJOcF7Cb48UGMZvjyN/51ONlNEgMRxr3IfzXBaX3lrJRdrKqjHKwc9YHogoo/t27d03acK1euptO6Deebr7/Wga/fpm+/+1bnj12wXlyydLF05bu6Tvi32pLzVlq0aLEGM3VotwZcRuv6eLbolrNVymrRGJuJckf9HRz8oVBQNn7LF/6DQf8GmCeD77LyYxJm/huJ7K68KodD+/tTpZajt/a/sr+GMPzawEkzCSoEi3cLoiiNIqARt5aiOP2m+ljVlJz0LhFqXi3OHjHkUSf0k8DHGKHzCYyGdEcBjJPCMOoRxfKgyx7KiHcqaqniBP4Zv9AsgRoFpXdMPCWMWemSk95ZMRJC1OicRXif4AO+Bf4IwcfPj8Kb4auBUBCdAw5WxDAbo4ETO8wbEnA4K5164a4xEwyjUTrAlX2H/DGSy1DK7VvMjo1T0hHp+LGj6auvvkz/+Z+f6QCnb9PJU6d9pSMzZh989EH6x3/4p/Tue++k7mXLq5Ou47YBZeT+KOUNOrJ8DyVMR4nmDPp7UEsxeJroPwD8M2WqzJ6I/wOCD41BGp5LtsQrfMA/hMYfIVMK603+6KL7v5Z0ls4gkvo4/j9//AM//z4B/qZan+kvaIIFnQN/XPpTXWlQHVcMEhIW8UR/omJi+C2ZhCel/sM7Vmc5Mvm3p3+cjh/hkS956hk0/MF16ML33mfRjgFV6r4HdqV8xowea/1lzkgvoBvu3bmbLulk/R1a7fbpf32a/vznv6RDhw44zbjx43R2ia6K/cd/Sp988klatWKVrlafYv2LrqPDS17A4A+mYnyZteLl7Tu3fCXxlMnT4zT+R9S//snf0Ka/8Q/KIJV+nl37U68BQK5/Z+DVS9dFulZVHoixzh2ibYhS92j/6imcvfKH1+3k33mXnCwZFaAoVehfVpzc1xW2bKdhsuGBAklFG6qxfrdVyBLL1Llql8kJzjzhUOKt27amf/vXf0t//fILnY1y1qss52p5+Xtabv7J3/9d+vD9DzWxMF/yNtZ6BBhuG4Uj7S23M1WrW8BDfugf2nFI0AP/x7T/gVdG8ynonzgfQvTI9M1HvQeAPtLfkyUg479IOmD5Ux6/epuDMCVLCIUSb8f/NvibPvoZMPxBpv/z5v/zhx+yB7N7a3+b+c9wQX7gPxMh1GYzXnKCzOgZCvy3reHSRZlcLpWvYX+ouCq4y6p4bfHX6o5m/Bt5UScGan9CNf/P8Nm6CHxu/KVMPGzz90SEGl8ugeBsJ1aE0L6fP39eVwj/kv782Z/Tn7R6/fjxE2rfY7vsOl1AwXXC//AP/5hWr1mtydgJypNVd9H+jxo3VnCUnwasOWR7lIDSXyD/B/fv2I/rh31WVD/rf6HOUOA/Qtm2/RNOrfSHGW3530/8yyqdQgeqQ3Az+8B2eRQ/N635I+LWUhSn31G+ki5yKxEKtJ7vHjHkMZTgVwMnPdCrU6YnXpldzeRoRCtol3cOKZ/lrVwKKwYTvsYPcoPfBn5BSZXRMz7UJNECb2L7gTYS7Zhtyn7llZWHO12KE/jlTEmG/s7eDODELAZ5A0PxclSUXq/wlQXxY7VMrVzyj4e86GIX+GFH8Ak84+/1qCqQvgM3udWY8HH/nvJW6pEa3QXVuNKPKgsC6ohogOW+TsUerYPtGFG+oVng/fv2efb3K40g//GP/5Gu37ylrH/V1p0FNhb/7g9/SO/pfvUVq1elCeMmyMC8mSZOmqjcaAC0z1HKk+w9Og3NhT+DNaZPK/2Vqr/4gyP5/2qFAtrkAD4OsftZ0j8UniDSFxcfBgo/SgsqGNFk5v9gV3t68v/546/iDTr9C3+hB/CD/q5bWu3V4H+4zSNTUbFdFyT/WoXVo/4blcfXP1d0ZFnPy0n/hv5phz+zRugX71VWpcAgYoWb9arqP53Qq9evpYMHD3mL4GeffeatOBhfGFQbNqxPH3zwQfqXf/lnrS55I82YMd2Dug8eckZGXuGmOuLZKcV3/ZbhhdHMSj/zP+vrcDM7lg3sp6x/2+HfkD+qRi5fiIv139PU/88dPvWgUR0bFSJqRgN/1y1dTwm/lCZqbvl1VfIP9BIns7vhHy7iN7d/Ff65HL3q35KV4LuxfiXOJWFVpIuv9JSryBQyy7Yblp3HoBwZBPzzF86lEydOqj38Y/rs00+9UvOs/LhVZ+GihVoV9ZE7CZ/83SduayfqBrtx48ZJPWkFpvIlv9E53weaqQWO64YhPF7/1Nv/Cv+n1P5BjUZZwLn+9JH+rpOkg5uNlTTOqV/1D9kCQ2Pp/Hrlfy/4m2fIRsmiX/Bd4trPI/DvBf6wr/9QRxXo6dlfsb3TedYoH07oT9tdb3/E2Dp8fVCfw/6W27I4FPgfkzoUnYmxAdu/pG3C/2nTP/QPt2dCR2z18rCyhAsfgH9T/YBb2tJ/6OBBteGfp0///Fna+tPWdEFbdKD5GzpzjINe//f/+d+a/HjV5ySiA8NGAwc9WU+wPQedY72TWeWJXUUp8M1Hhw0O/hTQZSzI+/3i1X/0nu2jgodoLkKXrx7vSl/1CMEDipC2vPHTUz7Lu/IgCJrV4A1x+K9I8Chx01PhBa4EZwLW/ZsSEE9/Tdm0eLR8Orn92gTUvZ4FfPKktGZWBgbT+Oap4Ashr7YQ/rTLHuVVKCsqolOkd26wqzRkACHQ2SgT3gUWTgfakaGFF0nIg6fKqwf8SF3iwhYXJadBdTkPIqB45VHCPUMmf6chniKatnrHYAUDIlFkr4hQHPwfapDEAj2SE6+5PoxrE3N6IwcwDaRIqbHn+6qW4B0+djj95fMv0md/krG4fYeVJ7PCS5fEFWLvv/++DMZ/SNOmT0+TZCwajhVwlrUMAP+29DfEjIsRpgQqe/Yv9G/gn3mt8MaT/UgIXfSvz/xXEifLdCDPPsEvzCBBhqfxm4HBF40KH8itT/CJWD3PAf8KNo5nDV/0qcl/KGaolP/gnf4F/ZFxvl1Z8ZU/0qTGX51sBiij/ptVOa7cfeK/4JHeuSrL6nnW+Atshf/QhO/Opzoo3Cwi8rjDSFtz4+ZNrSzZnr77XtseZGxt/+UXXQt7xWdBTJAe+Y2uRP///sf/0IDsu2nO7Llp0uSJykOdSuqXkB7FCf1aHvDrK/el47QY2/RXkGhvmojNcbisVt4xVQYfxWfzvRY3/MNAbNL/ygcpUtRKv/I95PWPyth4Blf+oCX0rz+moX/qvsiBtq3gBX0lH436F1qasGdP/5jJVB/A22noHtCGchYOK0FgPrOdrDJhPz9L0OH/GM14utyaEUW5sJ3n9u076djxY2nnju3p++9/8ATDrl27LIMctL58xfL095/8ffrgww/SazrEePacuRYs6odvjkImhXTVhg8K/o+mvzUl5YIRenj1Lv+ZX47Z+HEaZZBrnvKIzPjtP/8ZKiGnnFgFbOi/3uE32l/iZPgC3n/4j8KfckUb08A++2VkQ7Iz/AHh/7LBFz3FvPbyR2Wpt/8hF+ZpYYBIXahN3Xre9A/ZpUSuWY+RP/BRiVXunviDjPJpkv9nJ3+cwcTDoC4K6o6204zReU3A339gn65w36zVoZ+nH3Sl8MkTx60LZ86cmegD/O73v0/vaUsOh7dPnTLVdoAHQIQYvKLOMDCDmxUsr2gwRg2CkYZSZXuvB631zfXFo3zItlIMEv496a+CUOAXDT7ClB9oW76eRf+7CU4d2AsEv1pxEmVuxaL+XXe3wbBQw+qoKCqlCa1UhdqBtNUY1QhshVH/rrtzijZeAqjAR8Nn9kjWmKMiIGXUObQQlTYUTcnLUuQaEmHUCjeyMpoCjxCzSKe81YS7ocS7KHA57SaEvICvZ2DwDb0qnjNyXpS7Dl9wMnx8WZ4XjU2U22wgceYFRqpHdOWlGPGrNGahFBZjbCNRXjRYhBr/wMF5yPOBDva7pbNQrt+4roOfzuv6xp06tPE7H9x48MChdPfO7cQtA5x/8qpGmDmLYNNbb+nk7EWJg5+8t1tUwWhkCd8IrWgBmo1V+aNIWR7IzDCrXZid41RuljnH8vsG/qyKAf8RWrZIeTGtWOoH/VHEeJJv//nff/oDvwf/5Tlg/juhkSJnPz3lTwDq8gf/oWaRvyeBD0SQUjlYgkmeozklHes5y/+vukazwEc+imn7tOEDkefx+BMH/MX7bOU36I88BEKU0xz24IliOFIJ84dRL/gH9EfAp70n0jOi/2PhC9f+wqdeRSpR1bqyGX8GHqC3w3IB6vSPehn1bwznQqBjhP8DneUA/znfIVZMURd1ftLNG+nwoUM+t+Trr75OO3ft1LLecxosuZHGTxjvk/Tff/+DtOnNTbrudXWaOWumz1Vi0MVZKw/zVph6ps5lQv4ER4O+hEIE4qD+fQKkymPUJLRmPXnIHXiAb+BM0oZsDID/A6C/5W8YwY9KZKYELXObk31KgN7wieXfPGZUJLU8ljUmcEYxxD93EoJBNb7BZQVp+4TWt0tG6/wHvEJd/zOnrbQUN4AqH+p/jkeuWf5DRsjXmUd8iqsA/sit6D/k35OwWf8ysHLzxi1NLFzWIMpxn4fygwZRvv3mm3RIB69PmDghzdBEwkrdxPOB5Pw3H/0mrVu3VrdCTavgQwjaXPQX7V8uveRZ9ojgo39b238mPvADH841cx1RXLYC4el2VXmRJ6taGAiiXY2zzUgFjU0Mu4FJQjijbCJML54oT/DMFHcHLmhC1Fb9RzC5ZQBVhgUv4DisDfzILyclCz39hU/2ztoM1UdGCBuI8j83+CpVB/7j6A+/zHRY76eZ/+Lto+QPnuf0bAH0R+H/c6A/HA/5QzX1hf+N8mf0lYp0fat/Vp2KD8xi/0KFqEUN+JAEXUYYtjgw7t27a9KV80us+0RDVt+dOHlct439nDH5clYAAEAASURBVP7yxRfpe51bcuH8xXTx0gUPKG/c8KomPXRYu+z91WvXeLBkyuTJsgWiDe8NvoHXeAOvYF3pf1SrTdwfoQ2PPsMj+W+d3kv9F8CXrf4F7XOFQAz81L/r7jbB2QveQNXqNwStCrUDXrrONXuXtA3fOsy6O8do41XyiCD9PiP4Hjgpgh/FASRPkDIoUBWjHtIgTombfRyp/uMs9dOWWMJNhCyrWgJxEj9D+IKHmWNYelEBY1YSPykegyZAaoK4PsA1d6YoGeHOA/ubvLDCUCkoLQmNwyMbw1GcgKUA/uvT/tlyGBh8Fy/DB+AA4MMvJe0BX7lFwKPwbw8fTDHT6BRhiF2/cVPK87yW2x/wFY6/6IyC/To4b9fuPTbaZuqmgaW6mWfd2rU6OG+DTtFeq+uNV6Rp06YrPQfZanBEBvBdrXbBmENJQn/KDJ3VJ9UNG+yRHK38pLxZuYI/Ay4YlEZQcbOxyQALvEPHMjATRnEeQCFun/nfHn/o1mf+90Z/iuFy95/+zwM+e0yBqzEpv4PqQiLTmTc88zYI3opuvjw3/KmzEFkFoZi91n/Jmsrevv4rndJG/Sevp1j/yM4FG3z+M9jFtjkGEFhu6/5lTf+FkWIGGv+HGjQr/OcKchtXDE6KpvcVNmr0GOtW6pqlRAbLSOWNyFzR4ZmHjxxOO3bsTJt//DFtkcF1+vRJXe16Lk2fPkN6oStt0lLe1958LS1euCR1LetOzFZNVCeTcya8jxpK0dFTve63/m3iP/odwvdV/w9P/j8r/N1m+gfZbn5a2/8w/sULVbDnV/8k+U8JPrdI0P5wa84I1Y0Hkt0rOuD45MkTaefOHenIkaM+JHGvtrte0cDKWO3lX7q0K61dsy5tXL8+vb7p9bR65Zo0XnLP7KpXujCBIIml/UP+0RfUP94MitCJYVCEazo9UaS4hFK3uQaUq5aFYGyNQ+6l56jH7mzok3r8OPw1JvRE+s/mNaB71b/Pk/8UrAP/+dW/waM/g4X9t/8tHk8k/8qB6mr5Hynd+LTaH/QCdUpZO0/elf0ld6nr6KSwhLDdov178EDbJJWA+o8NTUMtbaL2Vja4Bkx8hojacP5Zz0i3cRU7bfjWLbpZbMvWdPDQQZ/zRPs/d958DQS/71vtupcts20/TyvqxmuledgE0vbSRRySHXqs0/6acfwMpv4R/+OhleBBavS4SfGPnXjlkAjKv/jnyOGs/zpL/fSx/Y+kQxu+9CI9AxU1N7qVk9IXCgUmj/w1bUqSpo+eyXoEV0DdlEZRSPaM4LNygsedAoDw7c5BKC8aCwM3TYqTwmRS2aVvwhFufaOCeHDjR9wMIDwjmNAcswQr7jCFj2nHAAqDYpxXcPHiRd/LzozbLs0o79HgyS+6oefkiRO61Sel2bPmpGXLV6S1q1drtm2dbuhZoxO1F3l2+aEGUcqtPFDXnTXRn86drzeWQerBN9GSu91dR4PBHjFHOfP4FO5SgZl6pnzyBz58o7np8B9hFVX6If+IvGc+oak+4FWR/5ixEF0z3UtV6NQ/dIEIN4Tqv+wkTfizl1j1QgyieYB/nHPk0qqB9U04BEo+qGuU3ypPMTBAPWOvNPDfxpf8OeCNv+u6gevgoUMytLanX7b9knZLB5w+dUon7J9SPq+kru5lOtx1hfY+v5lWrV6Zuru70/wF87VFcKw6jcyoK5pkrNR/YAFTPwqx1Dof/VTyB4md0DEazo78QYvBkT+kojzwkKfywcO6Jtp/j6oR2g/9M5TbX1ZPenBAdacsRXe9yhMLt7Q17eDhA2n3rj1p809b0ratW9PRo0fdUWGgkK08y7uX+0yftWoXV69amaZOm+H6SZ2BjgwikicdECYc6BSN0MGMtL/UX9o/YKOjWVZPXDprZEJ9Uo2SMKg2aaCl0/5ZIIeN/FlASqWTrBRnR/8Nnv6riD4I9Kfp9aSgJhL9Lw90CLTrfb3/43ZUumDUmBgo9eCpvmn/1bp7Nbe30uY8GfxlQOTU6TNaKXdQbfj2tO3nn9N+TYweOXIkXZKNP3/eAq8KXbdxfdqwbr1vFZuvARTOLuGPCTT0kH6sf9Bf2N+2Gyix4PM1XPQ/qAwm/8MYEw2jgpuUj4OPOUdEU74oCMrdx6cpSdNHzwx6BONRa/9Ls2+69Uze1qcpz6aPntF7BA8Avi5+KGchFwDOJX/UQGTveiiR4ruXwFqWcW5G4UTJpYx3loiNHMOHeDxK1wuI8O4lMBI7bR0+CoE8qaCx4JxPYCgfuFZ6AZW01eArXswyR+a9wQdTZ5OzfKiBGa7UIn4RkuEKv9FBpmNFp0uGmfDnSlHwv6ctNNeuXU/HZRzu3rNPnaddad/+/enggQPpzNnT7ozNmTM7rdYASnd3t+9yX75suQ/Um8xVjtpH7tULyotG4FcZpmUZoUVFfnX6Y4g/0KoV6O8lzllJQP9fpTGI2+E/Mp6FtZ/yLwZAwOC1OAz/GaCidtNhgP9ceFD4P9zl3+Kknxe5/kcdVv2lUwUP+dM/VTTxUbLCnx4f2CZ/rgxW7VL9khxgECFLct/S4XAXL1xMJ7WS5PChw6rruz1oul91/fTJk7oS+GFauGBBWrlqVVqvTuH69evSihUr00Lte540aZLlKGa/BVL5eVmxBmDQIzy85a3iqFLzEcXy60WmP7gFQryFVMYN/F/E9oddb8hHPIVRSFTDt4QhOsUd72GCv/Bye5XtD5qhYn/QLl29ciUdPXYkHdh/MO3bt1erUXam/WoXjx495rPDurXiqnv58rRR9YQBlOVyL1iwUKszp/h2HmSEVZkMilD/kH+oi/6l/fP6MdddCMzKL+pTDJz4hiulYVsrHZhS/4cL/V8m+wueUdM6+g8iiBb6e5n431r/sXDBXyqg0j/WsbTTWlWCrDAgEhMScbA6g6wMvjLZQduOHc/16ge1rXD3jh3p523b0gG14Vy7fuHCuTR37rzU3SX9JB31+uuvWzctXbpU/nNjhZsGZtE/PuBaig+bwgMkGb7tCQrV6X/liiumuNHnLQ6Fyrb98yza/9D3BSYSUdy8a/DjC08/UaxSOHmVpDm8RKr3vytkhFS79r+RCfnyDD34kl1JsApm1DP+UdjGby/ejQiPcTWl54OnInCE+rcpYkTjtxfvRoTHuJrS68NKgwpKGeKjwh+7gvgEQhUMVaIw/kp8R3d4Tu48sjv7118ER7cyfK0XyGiYwrfSzgSgMrIa5L4UNHRAGVuJQ08MNDpCCrt8GYPxWNquGegtW7fYWDx58pQPk4J2KF+WLK/XLRosYV68eHGaN2+elu3P0MGQk80wzzqbV6qKehdFUKe//eEqDM70B765RxpYrAQE49fhf9/kn0YVOgfdoCFf1BWlF305w8LLxsX/4S7/Rrz2U5c/vIc6/nCRThSDFQxIUCGiutDIMZCSawf1RcH3tPQfV6xC0fJdzaLf1fLdK+oEMgN1XPV6pwZL9u3dqwGTXRosPebtO/MXzEtdXV1plWbPV61a7Y5gl7bsUacx4hgIua+lwdROyiJ1YX2hX/mEbFGRNSRXwS/y5wj5h9i5httnqNNfZDW9wXM46R/rZzhQmBRszB5FxvQJ/qGghw3+99QGeiUWdUb4V6ulqGDCFVlnQIWwwn8OlD1+7Gj6RZ2Uv/3tb2nfnj1qI49qxeYZnym1cOFCDTJqNnfjRs3urknLurvTrFmz0wzVH99S5XaNFV+yXJypai90Vf21xcdAluAHG4L+DLAQlR/q3HCSv8BTqIkEaIwgNuInD2OavRSED34d/CUzkgVI1LF/RQRkI4QjS4i+EawIqlSbA/MPwc+j/Sk2GMWgyNhh1P8i/7ScrIRjhRo4GS0pBgZLshLweYSX1I6fPn3aumifthL+pBVxP2uFyYULF9L48eN9sOtKTXS8ocGSWCW6yva6YaFH9HeXs5PQgdJFo9Ar/qNkog2r8QBOXOpcp/5lORpc/QPhkQ1kvPXpxbs1Wq/fTen54KngRKh/myJGNH578W5EeIyrKT0fPE8IXwMnWRU05R55V79EyRIdsxhVyGMdjWwxuxGGXp5GxJ4RnjL8gnIBRJlMBaaAKEftibB2+DukOWZopWBKzifQIq4UVxHOWiqcwwk+43Bla4ZHrUHQtGA5sJwyzpiRxrjzeSP42ch7Jd25dydd1GFSp8+ecUfr+++/1/LlXemwRrkvXrrkfGbOnpVWain/mjVrZTRuSKt0mN50XUPKmQgTJ03w1oAYkAkG8OtDKGv0p1weRBE/6OhBfwxFae5cVr3zE7xRLj3k3yElmt76fkn5H40kNBX5VJH455VGdsd2De+dzwMnNaLZGTSWs0P/5y5/1N9y2LLridjiVSfwEsMKMZcfs1AUlnpEfK4d5BC4k7p+9eTpU+nQgYPp22+/TQcOHkwnjp+QgTYyzdQqskWaIafDx6zUBp1ptGz5ssQBce48qo7WWwjLlaD4ERyP8efOnleYhGKxzPn7Ja1/mUBDWv9Q//va/tfb59ANL7b+5Twe2n8OQaTjUPArb3CMs7mijlHvShvFOWGXL1/WbO/l9N2336UtP2/V7Tw70rFjx9OlSxfNerazvvrqa97KQ72aO29umjljZpqsesUtPx48UYfIAyjuHOW20UYP7aNqnQZv3G7Kj/oOfJ7hQP/QKS+H/WWVaKaZffkTjw7+L4P93cp/BmTr9i+DE/UHHcSgBvpn5OiR6fat2+myBkvOnztnuxtd84tutftFt9ydPnU6jdaNORxivXr1GtvfmzZtkt7ZoAmPWba/qWvokXt371mHeIWJ/Dg7rU7/MrCDrsH+R/+wtdDlrxWwo3/CphaBTJtif2XN3EypJ7B/nF+zaDTypp3oAb8R/CgXLU1k2/f2v0d+QxB+m606udguLO721GwQpAeazcLfNnlOrVdDEFryeYbwpSoCcAWfQupDL2SPx0aNyx4I5BgKQTU4B1Om4e9k8ZPzcINdGYwITlCTZUtGPL8cb5jAh24cTGfjT8oQSsU+yhjJjltsiMOBUDqzQAQkPvWSg/MgDcqUJYIcZHdIy/u3aVng3374wTfzMPp9Q+ckMIo9WUv5u7q6rLjffPNNHaa3Vkv8F6YJEyb4EKtxY8dp9lrKmHJoZJubDVDSo8Zo0GZEbC24q1s+OIfBZ3KoDB3+ZxFuqX8ijeW3nfz7hhUYWB4EXQ8NNvyPDpM8Fcf5DGP5N+IZ/+FQ/1lhQl2k3mAEcQgzdem2VpVw/fiNGzc8I8WWgm0/b0s/bv5RVwnv8HlGkyZNtjE1f/6C9M7bb6WPf/vb9KYOe52n1Sbc8IFAvaI6bzlRx476jmHHoCoPM2LIWznQmW1BfDNRLwWjtxySU8hdDDE5i6AqpsIla0X+HI/gjvxBBNMp2l8ICjNMLij49Ns/QFRP5oRebdt/eNoP/UNuTU/2GCr897kBSKFkFjlFJ3I2EIPLRTuO0sCiBy7kZ9yFUGy7CcwYQAGtW+rYnNKhslu3/awDZb/RdZ8/puPHj2vr6zW3o5wLtkadGm6uYDBl8eKFaYq2t47X2QKjdbBsaX85G4wMfc0nNQXeU7FU36uBE/MsGMdvkBWqgoHFBZHJ/nKUZ4jRv1P/g2ewxfzq6L9QPCLIoOk/1ZLBpn9MYnIDJQe9ivO5vfQUphDnH/X/7p270is3NWByOR3Yt1/nLP3kc0u26xyyw4eP+EDqqVOmpEVa6f3uu+/6sNeNr72WFsyfb71CvugsD4LojT3tLT56s7JcKsU2A4M4iqi2m8sYov13Y25vDZ6ojIgmz3DS/y+C/kE0mh74ZL/WgIgFm9qHKCDz0DHbRsqp9Wrb/pNwiMOvtuoYyTbUaPZqIGyqNQc6i/iJgBJc3lWEJo/aR81Z4jZ75a/iWd4lcvWOgBJc3iVYusKVN5ijULjH0xRRlVi1mErf4D1VQJ3BwtSm+JEFv55dU54EM29TjIzISb7AI8D55G99Dgf40IbJKk7OYaAi9lHnpfxZaO7d1XJ+nXMyUQMc1pSKi6JFkTN4Eg/KWB0ldb6hEwbnsePH0t7dexMrUb7+5uu0XQfLsu+StNw2wJ7KV199VSPh69Nrr72RVuva0qmTp2gkfXSaoGWFZZbcB1oqW/NVPIanfpr42eF/X+W/DHRBQ/g0isEq0ZSG8oEOErt1504Y7rphxUI+jOU/aEDH78Wu/7Elh8HOGKCgXnMN4W1dNc42nEO6DWfPrt3pm2+/8VYCtuVwLTWDltOnT1Vd7NaVqh+ltzSg+YZmpebOmZNGi/9uoDP/OWTuzt3bsp0YuByZxiAfWdkyA3ZPK9CAT6cvntCVdf0b5mjIXY70UuvfoSp/HuQqDGrSs7WP7DR/JQh91T8l2/Ieau0vW1XLqo7e9J8HVGRs8A83f8Rl0KVBu4b9waDiTd1Yxw1UW3WY7H/+53+mL7/8Mp3SQct3VacEMi1essS3UlH/NmpWmFWas3SN9/17D9KkKTo/SHWO5pX20DSneukfDwOkw4X+L5P9Be+Gmvx36F+atajTg2X/e6JBiuAV1XNvpZVOZWvM3Tu6CU+2NStNLl2+qNXd+61Dvvryq7R5y09ebUL9nyIdsXDR4vTRb36TPvnkkzjkde58T4rSTrOSDn1RVo6zRTdu3ZEe6Yf94/4YckumkMjP8NH/L4L8u72pNcWFC81e+at4lneJXL0joASXd0tw/qyF1pwlbrNX/iqe5V0iV+8IKMHl3RKcP2uhNWeJ2+yVv/RqrDgpMcq7pGzzjii9R+wZ0urT+K5cxVHebeAWr4jSe8SeIc0+3seXayimiLs4RKG/15Byx4jwhkFBGSI3/WLgWEHgo6foRfKshk4JiBS44ml8Dzf4Xs4v/PNKX6POCd90hFiFErPJDfy5FhH6w4VXOPeETlVWnsxyYzhi4FV+6sDdVYeLq0y5nefnLT+nz//yefr8889tNKLEWQExbdrU1NXVlT784MP0oe6OX6fTvVdoiw/7Mu9rlcl9rXKhLNwi4BlsitTh/4Dk37euiC+8EfWR2eimI83WjTNnTvvAzxkzZgx7+Qd/S/MLXv8ZDMOwgpcMXJ49d9Yrv9h+8+PmzelbDVze0G0g927r+kLVG84dYpb7Iw2WvP3OO2nZsu40VTPdDIbQB2Qghs5ZdAob35YVUQy9Qf2/e1tXt+ofddPGnnp2pC8PeVBVlcADc9z2w0yVjTfiDRP6g2Kgg2YELWONt11us0SYF6X9ia1VLn6FQQOXmpf5F7gOF/yxN+r4t9ofngwWl1kJ5durhLi3ylinyi0+w3/0qf3hO1Kg90gNOrJS69atW97Sw6TCX774In311Zc+vPHG9RtK84qX0rNF7q1Nb6X/9t//RbfVLfbqzNFjxrouxTY4ZZnbX9dTs2V4yF9lQCB0uQaFK9MSkeu0/wNq/yvdFBVW1BVNX/D2D9l40fUPbSWruX0bJSOkerjd8tq1q9INB9O//uv/0yTkj2mLrhFmEJbBlBm6rYuzBD/44IP0248/Tq+/uUmTj5N9RTo6watPlQ8rWLyCRCafV4ooDPuP1Sbw/86dW27DR8m+rus7F4IYuTzEpS2nStL+d/pfohCCl1+uS3JXdcxuooT+f1rtv7V8CHwFnzK0e1CVhl8K2hIpwuuerT6N78pVHOVdT97ijii9R+wZ0urT+K5cxVHeLTDrn46iysW79pSUxTtTs3RoS7BS1Jy1r2bfWsZ2VqF2VF+1aMWPN8+zgC8x9ABJhlUExtAQ0ZoHbhqBh7wVATpwQjVejp/DTUZHyL5h7Oij9oQfhtJwhV+UZGzTCB42RAxaYQg28K+TLYy1Bv0jHWmgtxS/nND/Vx0PzlkpN9UI3NHS5Xu6e/7smTNpp2bAf/zbD2oItupWgn3aA37Jypsr0CZOnOhbej76+KP04fsfyGhclKZpZnzKlKle5WLDUdlbIVipR9l99ongP7yvMqjgrE558KtG2hnw8Ug+gzq/etvCGBmhaj/cSSyVA5zAo+DCipo6/kbKeIWYhITUPEDawsY78B+K8lf2qZbDvsIc/dX7ZDlM7O2337axDndpdIer/EtALUUF/+Bq+Q3utvK/nAlUl/+oR6RTTjbo+85/rxZRMYpBU6CXulbgP5ShxIAmcoWBU+BrDENnKBzVifkc6LpbAyY/p5+3bvOKrzs6tJKlviNVqJW69eqD999Pb2kbzmrdijNPW3KmTtG5CqoHzk+AyJOnyD/8L/Bdr0HOcp1LKcEo+jfSgrcK1KJ/ZaU16d+AUjKC+kHrnGt+tad/HX7EKPmQrD18F8mhObwQdwD8fxngV3IQFM3VJOoKVG486Eq+/KP38KO/Zaxmf1BPK3Sz/OdqY1qU+k/dabU/ynZIIlKXb9++pRUn93yT1e49u9MP2uLKYCe3WXFWCtuFJk2c4CtD333vnfTb3/4uvfHmG1oRNtdXhY4ZLT3gpfZaLcqERaZ/LP6knCp9XM+hhpyNdc3yj95C/7gG5vrvQVgNxHrFp6KX+l/V4Br+nfoHl0WkSiAyfVv0Xzv9Q71hUsgr+5QHOZXBtqJLkbVCf+tV868BrkP/gdPfdcV8a9/+uF7U6I/841cGHBgUfahBUB+eKrZTb8jTTWTN/sX+e6gBEbare5u5VvjGqjRxL9c/D1ggARIKtvEd0yHt2GFff/NN2vzj39LRI8fSzVs3JB/3tZVvig53fSN9/NHH6T3phO5lyzXpMcWrR2jLfZFDhq/s2/Z/LLGSv9Bb+kJV6R9yF22/0mWZRP6sxxDijvyZBNCL53nVP+txlyDzpBQI/uAeVv3/QKnSf8a7/PQN/1fUYZR0l8glcc93+xjFt7yb01FhQmGHMoIFfhS98MH8GWT4dJJ9UBEFopGXIvoVxUBhSqWOgoVf/q3Kr++IGIrN6awAED+QC/yAA/74PeQtf6cTXTrwRY8npP/Dhxy6F/lgNN7UrNv1q1fTOR1sxVaCY7rBwwdbqeN3TFefXtXtPQyUcBjljFmzfPXpOi1f3rBxXerqWp64AnmyOn4Wzhr/GWWnsJ75g3eWa/dmK/jykiixOiaMSeS7N/6DNjfNuHOc5e+VV+i4hni0wpdvVRrc8Zh4cvJGsPRWBs9T/kpdpzjgzx7+gwcPejDrvQ8/TEsXL1ERRTdwV1nBnwGokaPAASNdCY2OfwIn41ZhKL/yEGd41D91eYSL9gPbojB5QjYyb01X85YweM2D3CNDkqW8ogv+cwirO1MiTzmg19mqE1T0T0Vi5QL9Hzy8l86fv6SBkj1pj/62/7JDV6Ae8flCHEDJOSacbzJfN1mxamvlypXpN7/5SMv+Z/g6VIys8eMn+AwF5C/0q0uoIoVMdvSf6GDWifpZ/kUalMZL0f5U174juqLDo9r/EHGIZUklhbUAX43HxNNniaN3lrWSf2/6d7jSPygWs8x3vK3utleinOOw9YMHtLVuj2eav9KWnivqTCF/kydO0paeharXq324LLPNy7qX69Y6XSNKp0nKgzpM3rR/uOnQsWLzgSYsmFjgfBbetH/u+KGb1A5Cfx7PTpPWn0qv9hT973PFFEdJxZKO/WUptzjXZNqUf7z8u86Ippw/RWea285oB+AX+h/9a10jPyY5oL/bG+sf6J9h+pXdVd16PHzXQ+VpSRGfH2X/DNf6R4V6FP60taYTzNIgSaE/9wQj/7TjnGvEVnZsI7bLUEdo86lb8I3BTOqfJ/Rq9g92leup0t3SQe3nLpzXteYH0matDt3y02bV/4PpwrnzPsSdLbLTtPp306Y30yrdiLPprTdc/7nRbvyE8WnsmHGq+6MzjKjD/CIVjYev4WF/dfp/cBf92/vTk//ELb7l3Zx+qPb/i31geXbR25e/jk1rDN+q0+RZfVSOevq27hITfYB+Lu8qcolQeTQ7moKrj8rRHLnNV4lZ4JZ3FbVEyB5mqNx48zQF68M41HwhsOMW5JwqFInMXjUS+lUEGp/SmaGxIsy/8necko/eVZ7Zj28/Hfh9pj8NCbYZM+DQ/xWflcAtHw+0AuRWuirj8OLFi+ncmXPp6HHNoO/ek/bu3+fZt1MaSKEzyHaC6WpEOIOBM1KWL1ueli9f5ts+WM7MzQTmDbzXX+ETMgR8pgNii5HKoU+1h3rEbxkwhf+xfNECElKV5YjGkIffmDGIDxpU5a6PgMavYw5R+WPFT6E/8s8glgeQZMAd0QDWT5u3aFXCprRYtz8Exhkz4QMd63QlgtF8gfB/0vrPXmR4zIwTxi7cbuU/288KnTCQeMJQCvlnBorGD/nzKhIZUMxcIUss56eMGEyjdN4Py3UvnD+fjhyNVSV7tLLk4IEDnpW6fEk3eGiwBPjTpk3PdWJZWq4BkxUrV4iHi+U/zdeBc5Whl/qrOO4MSc8V+UeCqRsd/UfbIGqI/lU9hnemjH7yOzgaH8NR/h9ll6ET6vi3k3/ihGKwi69Mw0xbeb/s7S8qgsEMVCqDFaX+P7j3IN3Q7PK1K1fTJR0AeVT1nuvB2XK3XbdlnNS5KCSiLZyh2+nmzp2TXn3tVbWDK3xmWHdXlw55nuT2knYKPVVv/8xbNcSF/nCmbv9IK1n+gYGuo+MX56vkzl7mpZmqsg9H+X+W9Z8tFhhCI3TgPa0B9L9757ZuVBpn/Y8qjsoSHXDoT31j9ZEfEbxe/zr0Dw0URMukg1AmDMSMB5qh1fva/iP/5AH9i5u6VPKNfonyU9vNKhQGR1zXFJ/2H3vVMJUPbX+xAxgkvaJ2+/Bhtef7dqe9e/apXv+iLdJnvH2dmyrHjBuTupd2pfW6iXK1Vop2d3WlJbJ358junTV7ts8bBBaTLWAVAhF1PfCU9xPiX+VDVgFFv4YW4PKHwVRCW4vbgR9yY6oVuvRd/uo0r7uR6Hr/I5iBL7Ee/5SYWbyLaDcSlggNnyZXU3D1UTma4rb7KDEHE37jOuJcoihEKUp4NvmV0rXDoK1fIy+79CM7sleWNMFqV6anAL8sBq8Lj0GpYobyUikw+nOBbRDQOde/6Oih7LIhTBwecJLbSfRpbwMoPlV2Ff4d+JlOJiA07B/9qaHelmB6y3AQEzDMbNypkXHnM8dhD/h53T1/9vSZdOLUCTUue9PhQ4fSmbNnNbN+SDPu590wcbjlLI28z507z9t5Vq5SZ1ErJRYuXJDmzJqtgy+npwmapeO2nsJeGrC7arzYvkDjRrkoB/A5p4GZAxrLOzo8y/fYaxYB/3b8Jx5G54skf6zIoaGn835LnXI65jzMnpzQOTTbtv3irTpz5s7VYbFaiqq9rOPHjY8De9X5pk5xpa0NA9FSUvBC4W9kxbOB1n8bsuCNEcUjWlq7WHXwgy6Kw5MJJl6RbZHOBhTfSJQHWJB5/bEtjNs4rqjDhBF1TjK+f+8+yf9JrcY6qts5dHWwOk2s0OJgt6kaEFmwYEFap5s55kj+u7u6NRu9SHVhvs5I0JXfU6f5EGbZ6BqACSNO0CmS+a8fOFfJtQ96poA5jl36KfqflPhVj2jQ0b82xUWy4dX+eJBPjHYdoa5UTG92WB4k/8MNf7B81vJP/XMHS/XInRy9PLhfq3/Eua9B/cvaxopuPnzkqNq/g14ZyNXhtIUXNVs9caJWZc6c6Vt5li1boQHUxe54LdGBsws1AD5tunSBBmGZ6abO8s+Dt4JN+4v+ia2tYB6dPPBn1p367zaa+k5wefgmHb7DTP5B8Vnxn/YD+vuqa8GhbmFrjNaqAW4gZIuUD+3EJhL94T/2BzxwmfT206H/M5U/6iY2IvIvQFErRfMyAAIPfKGCDECv4HJ7KrtK6lBNObXIW69G6dyw86qjrCQ7efKUJqeOeKXoPtm0J9SmX5Sde0Urr2fIju3u6vIA6Cptpe1etiwt0cTHnHlztKV2qlaWjLHN5frr+ioLQuWpBj1LNVSBVaMNn7ium53+j+hhNsI2V7qB2n+RAXk8P/1XZND6ILRCVawmP/MfzPv6RGpi26WfYv+1y6EJVo7Q5DdE4LcMnEQRXd7KWTmavfVVcHCM5mgZ5fJ6ZGCJRI76y0ypnJXD8epfA4VPOltw5Ig7PzQjWZ1VRSl+riQINjGURk4rNN6kESHdv9FHy0NKRnH15r/CDQMHTy1+gVX5O0mUiVy8vYHUStOBz35eKXMRBnq6y6kfDsNCubOkMZSBqKpOJ+yBknQumYm7eOlCjMir4WH/J43PUbb3HD+Rzmow5eyZs+n6jeuefeM8lPnz5stgXJiWdC1NixYs9AqVGbq7nlsKpqhDiTBwzbG3SQiYBxPkRwMYt8wwqMMtQ7SADLDAUZgKL7ObEhZ54C3vIhPEGKr8V/vu23Nu3LquQ3uvei89J65zwCcd8+3bt6U33nhTMxuzPHDCKgXOl5kpw3zihIlQwR13tpnAGz8vEP5Po/6HDDT472Xumf82VCQYMRfUXP+93Bd5EtFYeXXt6jXT/9LFSzauTuhw3oMH90u+j/oGqoP7D6QLCrunW62m6IrBeRogma9rBRkwWbF8WerqXpa6u7s98zx9xnR1jjSwRZ1R/iPETwbIML5Hq9MUshr6ycqQQohvNq6I74EgUnb0X67ML63+L4cBShge8USlt/y8ZPUfohRdbwI9Af7oEmgYf+TWqH/UXfT1aLVV1GraKc5AYGD18KGDab8OjWTw5NDBQ+n4ieMaYD2iSYH7abpWny1cvCh1dXWlbv0tXbokLVnS5S09c7WFj0FVZt7r9R/IZUsBusKaQm90PM2gH32X52nhX/KL9/DXP/AT/TtCkzllexqDKTwcAjpOqw1mz56Txo3X7WSKzIH8DHzzlPi4O/TPbZkJY4JUNEGKntT+8lkmEnza02jNyRRgor3qLL7UNV9aoHjEZxIFfvkGLQ2WnNTAyHHs1cOH0wHV0ROqoydPnU4XNSnChBTtOQe1s5124cKFeq/SwEm36ul8rRqbaPuTCxfcEAFX/zzwpgrKgAl10+22ogQ14vdp4C9QL237B6WHMv7F/nQhEUjkEqY3HHw0faHTJbbhV8V3tJafRwbW4tbiVc7K4Xj1r+cJv2XgpIUwKqppV0Ot4QwU6og0wnq6WmP7u03iulfd/fgce8ao+0Re+VcUx6WjTaQM5eAPj/xCGPj2yJjeYYjIT4ZBY7yDBKg65UP0Wh76bHqsnBST2KEeA1wHfqZbjXamah/pz2qTGMUXVTUkH9scsCEYpWdQRf5ukLQiRIem0QgRx6tC5M+c+S0taWVmnNsHTpw44Zt6Tp48ISPysJY97kun1fG/qI4mB80+0Jkqk7TaZI6WNs6bT4dzgVeldHd3pdk6H2WOjBO29vDHypUyeOODuoSTG6YsM5T7FZXfMkZAljM0kf77+0WRPwYN2VN9Wwf1Xr92XVukrnoFCfhD0+2//JJeff11DTbNNl5sF5kkGk3TVdHjdVChSOIlom6wX0D8n0b9t17IeqnO/1iXFNtvMLbq+octaawmuaytNVdEc7bZYFjR8WEw0PKsQ17Pa28zZjQHt2JYsTVtvgZKVi7TtrSVzCbHkt1YTTVRvNMBg1olxcoWVqyEYZVXmUhO+R6lARUY55llyi0egkM0wGKz/MqDf6yh6eg/81mEgTovk/4P+QBrlFvgb1mOT/sVf4cq6oui/55G/bdAFLpAkwHgTyeYuokebaqHyov657NG1Amj7UEH0/7gD0toF0l7W1fHX9F2Hs4HYwXKrp07vRqFAdhz5896YJZteZwFRqeM1Ser1EnrXqbB1hkz02ydHcaAONt+0P/oAbaqAoftq9YF8kO30P7SKNYPx1VpXrj273ny3ysLqTjZhoDeMWCt1QvamvnVV1/5TLfl4tFkdZzhs20mDYLD/wfYRCxngPB6OvR/dvKH/ek6Kh5B5yL/9foHP2l/bUddv+YtOEzkMQC2X9vMadN3aHvdlatXvCV3gi49wB5dvITBklVp7eo12la7PLHChFVGXIyA/Un9c29FgOE/8Gl/kBtKQ720zqiMj6wzJC/4q1gijOLx1hMYOMdO/wd6ZJ1t4rT8WOeZxkFnSGjS68fpamnxr1ggf+g9GPSnjDyGb1ezO3vVXhGzHr8W2MPZGtvfbRLXveruHhnmkj46TiNVxGvEtqvxWUWse9XdVYTsqAZOeo9UqlUtRuWsHC359uZPtBJW3pG0+aue3bOAr1JYE7jJE7CG8qB8NTl2eWOdQmPWhhseEG53ZKxU6ilwK64jZH+/Shzl7zRg3IFf5KFweSD0VztgcougehhB5355XCh9O+wmbw7gKp069gT7PBFFLvDZX0oSttpc0JLHnTt3pN3aM3rk8EENphxJx08eT5cuXNahezd1Het1GydTpkzzVp61a9d6MIVlzewfnT1rTpokY4UrkCdoZcV4zfiMG8eBexwExl5vwQWeGieAYtRQXpbRUv54KJn8kSkiUbghJn8PuBKWoumHwax79+7wITxGpePHj+mWoy3pnXfeNY0CXxrikH9u4Sl4FfxB8UXCn618jRKbQca/ak0dXOL0rP+QoshfoaMy9IO83hd97+p6v9u37kjmdFbB1eu+UvD8hXO+tWjfvgO69vmEZ4gvXLiosCvumEzScnsObp0tw2r58hWaIV7sAyC7upgpnufl+OVgRzovGHXAi+u/g0fwS6OSkk3KjdH9IN27c9+Hx9GqI8t1/M1Wm2nIM7OaQu4x+BOh4E/8QqmgwNCX/zr+gWxWSEUuOviLpxCBh5pdOFz8Gr4RBZ6Xp8P/vup/dCq3ZIzUzHOch0A7o4FPt2maMKCOy82MNnFpf+ikUa/pPPssLuo7lRhFpOfu3TsaMLmoM5A4N2FH+kVnJxzcf9CrMi9dvqjDJm8qj4caLJmR0CtrOGx9wzoNpCzTltfZWtU22SvbJk3SlaZjx6hsIxQ/BmPjTDLKFdtqQ0Y69b/UjqgBj5Z/4hT9Q/tb1z8cMvrHP/67z6p69dXX3BZQ0+B9yAArjpAXVhVC94DcH/hhy9ZT4O7oP9efFv2PrQO9yzNq1Bg51bbrFqybsidvqS5d1RYbttCWbXOHDx9K+1Tfjh45pHP7tEp02hRdGTxTg5TTU1f3Up9Xsnbter1X6dbCRbYzfWmCZIH21yIBFFXpsLViNXZMhgTfOBMN/jPAQnnMQRLo8eG0Q9z+LPLvAkfpQ3+10B/cjJ+IUuzP8OvYH5ANvvd8imaAdjlG5awcLcl68ydaCSvvSNr8Vc9uaMLXwIms5UKQOl4VgnUkmt3NyBYEc5ymQH0gxPpfa5PbRWzQ9RnCD8AZbSuIKFvUK9yE5TIrsnz8yKfp4buEUXCW8nEQoJGoBbpa4u/OidI4kSI4ao6YkwWdCNOf82rAIGb9ySmzVwd+oT/GoDvhojcNBPTHSIuT/kVWGYcYhBiLHHDJzD0dR68MUZvPNa2MZUTHT/uFdWYHWxM4I4UG7ai29bC1Z/8B3VSgax6PHz/u/G7oNPN7agQxEGdq9m2RljazJ3ypOqxdXcvSIr0XLVzsAZSxY0f7JhJOL+f6QMri2Qh1kllG/SLxH0MYmWcrh8xgu6lLaAQOIuT6u7ffftf4I9PIPzzAoGf2EsO5zHox2GXcXyL5L4ZL4f99HfB4V4NPGEoso7+t83mg4zHJ2cFDhyxz+/fvT6dOn9aAyh3PLCFz7H1m7zIDJKtXrU6rZEwx89St7WVz5migZPIkxdcBsdoSRSeKrWOF/qwgYZsbMMeNjysIH6hDxAAfe+G9ekqDJDwYHXcEl84Zgywxw8nspZjufhc/rFDBEEN/SRI6+u+l1v9C3g1ZX9p/9MaLpP8s5Mi46snzbP9V6wwf3QoJrVNz/WOQhFWB93VQ7ATdnEGEon9hDHHdcVY9pf0r9KdjTP3nzWw4Zyec15lIrGrb/NNPafeuXT5s/fz5c9raelP64471Cysz165bqwHzd9LCBfNTaf+mTJ7qgyjRMZzBEXmHbFDmjv2jxrGf9md0gtmqo1Uk4l/hP/z9j//4D22zmpo2btTAic6lCZ1+3xM41LN7nLvm1YU0yh36P2v5g/7Yp3e00pmz4K5ppegpbbXhFrvz5y+kH77/QXblfk2CHNMq6Jse4GLijVVc69atS+s3bEjvvP22t+PM0xbbyRza7MkLDX6ojtL+MkDiQRB98tD+epIOFiucaMX+wv4lBv/h/0PZn8gf9i+0oN4TxIN41B++Sxihz1v/deCLV9hZcKrGnL7YX/DZT5WucoR/m9/mGGiThqzU4Ud5lL/+96X9D6Fqzr0N+GYQ+hpM+NWKkyZKq5SNYjRcPQrfC26t3uU73vXfeo4lVvg1oDZc9dh2Nyepglu9y3e8yQ8GwmRc5Smx+Laq0S/aIxRSJZBZLh1LSSyn5jRsCyo25ySDQ3AiJ2znwKcD/+nQH9r6FhG9aQSgvf/UUtDpYxaNARMJug5F09kM7AOGv234760/SuMaLj6xYoK+Igr517KqQh1OANB5vHfnbrp4+ZIaPc5IOaxG7/u0U4bkPnVqubWEpbKUjVn6kcpjnLbvrNCS5rlq8Lp15ePaNau9F5WZAmYQuNIVBcYMEA1XgU8hqgGJImSUknJQHBA2VniGVBPNj8KgEfDVZgpzvgJ/UGk8ZFI8yEOxLdx6C/6j5J+BJ+DTiSaXkhMG3YH9B9IPGjj56MMP02Ktwnmc/IcCD/il/rncLotypogA0FPwL55VkHC1W4SxlOmjXv8YKm7Hf2ZeeaAn+MfAmeil/Ew8AMsZuTfDxxABZiv9iWuKK1PwAH/gs0qHpevQCDm9p8OFr2mbEwe1Htc5OwcPHdDNF/u1PHdvOq2lukd1Yj6rnEoHh4GMiZKXLsnT669zneDytHLFCt0GtVzLdpdoAGWK82YwZgQDJYJd4LPHWWAlXyod8qKyxUyYBg9Vh26r88Phx6M9I9bA3zNW0FJ0Il0r/SGcUDMNTAucGf86/CJ/cLnx4G6kglqF/00GEVFyslb4Jag5p47+Rf6eO/2RM3O4/tvgfjA1+F/qX4f/wbUnkX+IXvhPPugf6n/Rf6Fv0XnWsq7X1EPqX9F/DIii/xlsdRVVnugMzku5pnPA6PCd0oHTW7Wy8K9ffpm2aWsm7R+Dq8ymMykxS9sD0f8bNJiy8bXX0vr1G1OXzklheyBbCpjZpt2jE2g9rHLW+W/lBTIuQshJqf/IC/Jj30rOsmy1wR/924o/6SMf0gEk4Pt8NOs6dGUEoVMdi3yIGr1MwyeEP5fF72evf8JuiZWCbMPggFEGpKDlnz77kwfM1+vA78k6h+ahBuXvqb2eMEEdbhUSGwWdb9xqpX7R9W9hZr39KfIPZwv/4WM8zVzz5Jm0pm/JzPw3Z0WzOv9NROfXaP+9fbWWLQOWD8WTe2rnr2kb3BHZi7t1kOvuXTvTjh070l7dcMVZJdhRDJLwTNFK0TWrGXh8K7256S2513jF7lQNggG/3v5iY3rlmNJ5C574zoQID/gjo9h/7fDH/uAP/lP/4pHMMnjqOs9LuZBJrn8WcPLO8h/SHvXvech/B37wJnNoQPoH/gcdecNp+Bk5FleE1H7rVeYR3iVavOu/tURNpR768D1wQoMB4RrEEkIF28pZQ7gW5oj+dgZ12kdKtyx1AvV0DzZ8Gv36YyXq4qP89A+37IigCRjShLvb4V8URmxQJxd99MBfGdiTMLlpoQSz+JKq/nTgZ5qLuv2lf5ldgVn8M5H1E4Zg0D9WmjQatjrtcRf6R2OZy5L5T75WHDQkkoJoRGCrIsB3/dCQ4eQqN27XYbvECc0YMCtH53eHGkhm5jh/4vLlyyRStsB5xStZuPbVK1O6dGPBgoWaWVirfatzU5dWCnCWypSp09M4VqFkKWyVP+ACH/xjlrFIGmUjmSlj0tCoI/8FvkIr/MNYauBv3Iggw/BR8g989qwDBUpgyGFk0xk/dvR4OqRBpQ3r1/tAXeC3PoX+YaxjrAhypr/pqp8C/64MEPAfMQpoUVaQJ55/lDYOJKQcufHXwIGQVOyIwkAFNMBgCFqMFB9EX2VsHAScmRdWZpBv4T/lhJw+4NcHVOiDGDX43q9fQSKAAimdYPJ3V4bqNd10wHJc5IM9ywcP6jBGuQ9pO9gJDZowCxXXAkJ6eDpKy3OnagXT0rRy1cq0UZ2OdRvWpuXLltuYmsCAW6Y/8THkTUOIKPjIfx1/cC4PciAyRHxhx6AICJHUS/v9pTyy/IM/9C+7cIjLfzKwQSl0S0ciwgL/OnzFbnrq/If+rfyH/oX/BX5H/2a61+gffIC0RdKJMzToT4l6exBH81wIwH/kpl7/O/wfuPxDx2J/0JmmqtbrPzzprf6V+i917vrHqs1S/2j/4BXtrHVVrv932NqjM5Vo/7Zu3Zq+/uqbtE8z6Fx3TtsXAyOvaAvrJB2sPssr4tav3+DZ9Fdf25i6u7q1tWSq5TZkAmkQpJr+oaWp138fCG801U6h6wgPTL3CgiqA/ikTK0Xvl7IU/Gn/0WHgDcWB7/bSX8I348+AUYFPHMtsKSywavAHq/4xWHJBA1hHtHLhpla+uuxq/37avMXniC3VYPpETdyAD+diLOvWBI5uuUMWwD94HTw1KkKxoPQi1j/a/tL+0/55IEkIIf+tT+E/9ge8BG9WZjKYx6rKVvyZ+Sj2H3QjUwasLDVKgz7D9mFV86WLl9PuvXt8ztsvGlBkoOSA6sKVK1dsf1CWMTosnwNdV61emd57972oE+vWy2+RVn+OB5wEkUNjx6jdlS0jj3r7G+1ulEOFUPkpAzVVCWv6v9R/YJan8B+ci/2DG/sn/KDZwPVPHT4gQKX1aUf/jv4f3Pon81G8gdNwSQ+ManKGh39rYY7ob0Vu8s+ZIH85H3zaPdQXZO1Fgl9bcdKCUjsiVBTIKPaIQx419KvwypGJW/tuAVt99ohS93hC+FKQiEjTA38FAihwmhUKdJKLTy64Q3FbmRCfdro8+kbNoELd/SAzg8EhHwC4QYr0JZnfHfjDgv6sIrhPw6pOMvvNGUi5eeOGljhfS9d12NdBHfJ15swprSjQ4bOaaaDDzCqD2xxSK2EZqdmfiVrmzGqAsbqyd7K2V3CNr7f7dC9Li7U6ZY4OWZ2vAZZp2pYxXdfEchAtRqAHUDAYavLHlcDIK51jDwxJkbEHntPT6w8SiUGIpBb5p+m18YEP8skACh10uVvl3wMxggv+anN1NaLKIzqAF1dfTlZZWZ6N/KOkVQDnJVc8Wf6dOT7QsA/17yF1So87A3LaNBJuGAkY1PIQHOHrAQHFI18ZyAV/0lH2Ov7Epf6DP5jyNOWhVp1ZO8ZORowOOpIHUbGjrmjlCKfceyb29Kl0Wn90Jthuc1S325w7dyZd1qGuLNVlmxMzs3Q4RmuAZIoOVeQE/MWLdIPT0q4Ug2da6q7Z2Vm6DnuijF7OyuFmorFaYg18YJMPq6uAX+c/+AcOHf3T0b8SUAm1Z1FDMBCOeCQnz7L9K/XP1cl11oJZoDe9bUh12t8+6b/nan9IhnyTTuGexAtNele68bbOa7ijFZn3tC3h4sULOqz6hAeG9+m8sF27ODdsr9s99C/bYMdoqyEr6GbPmZPWrl2XNm5Yn9bovUDbfLqXLfN2E+eOXEj/smpGmq/Svy5Chv8r+lseZSC3rv9i5rxh/1n3K0+2LT5Qez1Gq1+K/keZMsgSD+0EDRcPrYVgC16d/vX2dwyrXxXugljon63+ZTLAN6/oito7wmO88GC1z3fff6cD7TVQsnx5Gq8tWg+1VWSMJmHmqo1h8Kq0f5TzWdZ/tYwBoA/6B7qxQsOTVCpXsTnq9PeqKcnZK1oZCQ/9F1wXL8Vf0QPis13U8iKjhMP94T+3HDIZwQPd4DEDbpYa5xXtf7E/oBF2kPmpNAyo2H6hnCJaacNPa8vNgf370p59+9OhAwc8IXJYk0bnzp01L7AFH2ir3IwZ03SA60oPFi5fvkyHs3dpxegK324HrHGy+yZqwGQU1wVLBvELGS46s738IWbUv0773+l/oaksNhYZ5OfR9Y/60/SQpMmr7hHy174Zz2FkViWpHNmv9t0EtPbRI0rd4/nDrw2cRMGai5dpV/c0bjUPO+ObX54GvRvxiqu8I2b9N0Lq4ZW7cpT4NQ8745tfnsfBryTKLYWUrjksZqhhRElZ6JQJ+Y1gxAwJ5IVSlZt/0tSG44Fd5cNXaWINXz8RVyFyE2ZwVfkcgUyVEkgd+MOF/siLZ/7hKqwVb1mm+VAN/S11ljmX4ob2g1/VIZ5c38v+8DPqXJ/UFcjsH9+rc1MuaUnnOXW+b1674dkLDLoJuoFm/Fidki6jbLQaVfa3cpPBzFkzfKMPW4C4WnaujM8ZOqxvxrRpWqY7xafpqwW2/CGMDCqgKDFq2fPOlbM+BCzLP0YHpiVFZ5MH8RF25N91gxD8hJxnJuTJSgXCqT8Kcf6yXD1b4tkf+WGYkKfzk5srmfkXD8aroJKefDN86p/zw4BQ/gzIRJJIF8Zx+/ona0e5CA5pSUS+/Ak/lreyHWXUSN0uQVYqa5E/11sDVdqW+k+HgJlTlqjfEg+v6zamCxfPawXJqXT0+HFf63lOW2vOi3c3NFB2S9te2Ed+W0YTW7s4dJjDf6doIGnO3NnqEMzwNhsGxmbqViZmAefomuvxkyZo4GxcmjppShorXnOV5Eh1MKA/W7/AAzqynasoFusY+Zky+unoH2RY9BAtOvq3EpOQD34Hsf2LGqpqJvm1fEalr/1GiH8l2652Cm2tfyHTipXlvNP+QtCg7mDaH/DpV+lX9I+1tQYy2NJDJxe9RM1z5zfXv1sMDuvWtRvSn1c123769Bmd0XTSKxLZ5nro4IG0V5MKtH9jpec4j2m8JgTGjxknnTjH2w/ZlrhaHc6uri5PJkynfWOSQXrcq16ArT/ki/M9aMd89a4mD8pMPeWmfE0dBeSNtkopi/2nSJYx5A88yTTOcIqtt0X+2M5EO0b7B/1JTzvjLZXkm/F/VvqnnEkhQN6KfEvnYYEf58c80OrGv3zxV92qMymtWLXa9gIrIygfh/VSbs7cGMVAABwbIvZvaX/hFfYH5W2lPyssvTU4ExY7AGer/HmADfwU38aIaOP8yDDrv8xq859BkNFa1eFVmoqDzaaY5if8v6+2nAORz2sw8PChQ7p1arfOHjvqCbHjeUUxh7wyGfJA9hX859abBZrwelfn/SzSypKFutGOGxlnaNJrorZMMREyUYNaozQZwkAO+HPbEW38KMkX9gL2JPCRactWlvOO/hNznoP+K/V/2NAfIZesUufs1LtyVw7i8NQ87IxvfnlK+nq8kqK8I2b9t5FHSV/FrRwlfs3Dzvjml6ekf5bwVUet6gNYrTwugSoqAsJTBVUOezd+Wv3bfhfP/G5+1YDkbJ8R/FjGJrwEX21F6FMoYHgNlOK7r/iHgU5TadZVZedbD5Yg2ztM6RKnA3840p+VFhhbLPOk0272m/XBf+SP1SgYe+B/V24OoOUKOk5Xp2NO40wHnMETOuEcTItRyXLcSzpXhSvq7qoRp6HlGT9ugm/w8eCKZu5YljvVhuUEX49MAz1tuk5k10ALNx2M1QDMZBlPc+fN0YDMJBtVY3S4KOKJ8TFGnXUfHKb6j4bAUMEg81aOQMkzNYR5Vk7y7VUZwpf0GDIMElDJfLAoBlBN/qkerfXP+3Jz3jZscx3CnWuMBno0ayQY+PFnQ1lvR3CWpW7pjX8OUGy5ZEzLG/rzhv737mtQgwENGZl39Yb+N7RC6Lq201wRHxjoYlkt9L908WK6qG1YXAWIgcqKkfJ3XQNcNzSYwq038J/BDpZEc9sEt9pM03arWbNneNUI59nM1ODIbK0gmTyVWyemmv5suYF/zMAaZxvsQRDo79uWkC0wIUwPtPUj+gcZ2uPf0T+iY03+LOiISEf/D0r7R/0qq0jQAABAAElEQVRzR6Wl/eMzS2zlCItEUp3rP/LdGscyX//pyP+g1n/0D21PnI+hDqZm9D0rL055ZUDFO7Y86gB2DdDfvXdb+kqD/hqop/27pS0l13UzHVeqcyYK2xRP6qyUM2fO+PB1tvecPHXKupazG5gomKKJAM55mKa2jUkCtjhwcOYynRk2b95cXbU+TWEz0jhWjtAOSde3a38oowdb1Oly51qr9axp/YPo0eVFP8RkGp1Y5LGxikXth2iAniZ/TxwonEFtopabyUJyEWDl9gz0T6krdOCY/AA+uFF2Bkn+9KfPfDU052VMnjIRjCiM2k1tQVFbwipUVl+4flJwgsvj7/AYzPoHTsiWB7hEf7eFKpNLnts/ikjbXx7wL+0fxTav9MKP9LIU5Aj7CzkNGZWXvOF0wZ8BN+JjW92QDcC2WuTxjOytMzqMnRWjx3UeCX7cuoithr1w6+ZtD4DMmj1Lq6QWpvm6tW6+BkgWaeXoUp3tQ/s/YyayqcEStfHYX2FLwQPxTvygzAW+8dI38scDSkX+PHACYvWno//Ma2wjcdW0skdQTt8IQvA/qEmc4D/0L/wnafUgSJnOOdcqaRWnOF5w+lPXCukKSlZkfcZfqZpo1+67UDG/m18vFHwPnJhQNSGpCCdHC6r1oMpd4jRiV0GVw3EaEbN/zWMw4QsWkKO8KCZ9ZfjwPsLC5RIiU3IgQx6ktoSEuiUnxyE7XBZAqmGBIe+SPkeMnHN8zyEroAPftGvQJlwmWaGf3kOd/pTXqzAy/92xV0efzjRlLytRyqACVyJbrkJ6POCCB/E5ef2mOug3uYJWgyrXdKPBdd2yclWNOYMouC+p4WYg5Zq2Al3RKpVozK+40Wd2jxkUJJGbe6ZooGTipIk2KmNL0AQdzicjUysbJmhpKPtpx2nWg9k+GniuUsZgxSAbM1r+WgVBHhxyS/yxistWIIxKDxQJTw764+C5sZrxAkeMEJZhY6j5bBDYCq7Iu8ol8yfe5CGXgyGIXLzw47ccAMxyaho6h+jNEmuWwHKbEQesAtNGvP3veWAjBkM0C6QZz9vaEsMgCQMeDHRcuqTVI6KjZ4lYGSJ6Qzf+cLN3nENZWRnk66c1uIKhybXS47WkluXPk0SvGaIjV/9O0aGs0I5Zpuk6EJHDDznojZlR019xTWvR3DfhaN+yB3IYCBHCYA56NmxlxEFbkyPTAxmytUcs/8fYQhc5eRhaNOIKJKof6KiPSKu3PCOM3wb96/BLLGKYD6RXwqFe/6KwHfyHGv+RpyJ/lkn/WLri0wKKZCKrxT/rhCK/HfkbMvXPgyTR47BiwPC2HjMD1U2FZ3K706oPDr5Ed9PpG6M2hXpKHkX/PJCevcNsvlbxMTDNhMFpDZpc0GA1B2eeOnlag9fn0mmt7Dul1SqcA0XHk/aMwea5czRooltjGESZN2+BvmdrJaYGp6WH6cyif2dKR4+V3gY+kxqUkPMVKHfppFN8HkSw4IPes/y2yB8DL0hoNaid9bVeXsUB/n6ghT6K/JNdhPH75Pq3DOrEykpKysCJtkypnfviiy/c/qxYsdLtDytReGJwQraB8Ocsryhgo/0hF0pJCcG/VM/B0P+2J6AXRdAPb9tPegcNY0AsCkcp5Z/LSWToD/6mi/wdqB8PJxExI4P83dOAnu0AyVmZnEL2GLBjoOSithlflTxevqS/K5fS5YuXZAfc0sTSaE98LFi4QHI202fTLZSbwbw5yKIG9th2ywDfZE1UsW0K+4etQ2Wyw/Il+oMBv1H2bAcw+CXPoEVM9HhiSsgSmzC4BYcK/qYEAZ32P1MGouqBJMhTJpQ+TbP4ffL6NxzoD00sVLxrT8GtvGtBlbMR1nBVgdnhkB7BNY+iYFoSlhjl3RLsz0ZYw9UazyE9gmse/YTfGDipQ6rl11Kyeiy5WyO282lJwqeSsTc21HT7cIS9enqCyUE9A3r6VLlUDkbaf9UG1RHaoBpqR0FVbZIDIro+ZU99Zn0UCDonwrKSy8qPTGhKCfGTkze5c7IO/OFHf8TGplSeiUIQkHGMR8tZFk7vs5Z88Y4ZKgRMIufGUjJJ+ryNBQPHM2P/f3tf2nbJUVzZQhubkNiMAM8Y//8/M/42/mY/eGwQIIOEFkDLnBORJ5asrFfSq5boluN2vxUnIk5mVkXlVnmr6rJeYdJJDr/p4kD8l7/4Rf77+Nm6dzCRfAff3PFXfngrtN8l8XvcJcGB/x37BoeLKFwkoO3d99998rcPkQcWEvjtCiXL4YvK+G3dt7/7bUi8S4MLJ1j04EX+t7lIwFtLIfkcLp+VtoUVvjQPsxUe68vgvYEJA9/FwUUV7q+eYWc+NpEGj+3/Rb7hjO2BH0h7JwkATTx+a6fr+PmyNd6+ygUSiwnaMBdK+LjRh/i2h4sjH2JhxN5gz8UTxIiLKvx5TN4dwjt4KHmszOODvwJ/gG+VsChFG/efP+HHBSbeafMSjo134HAhhN9y8lsiHT9vsf3Rj36CO3m+h4n79/EM+fdtovQ6j/u11+ybUR4/H69hnvbNGA5KE22cRTybjYnc6n/sQpEHjWrAOsSNv5uGEItGuOiwO5fW1YjFkHUK+2y3Xns35IstyIdZ2Wfl2TBt4E//883rf3jmn9Xxh3U2PqznD4z/9q3+jL8MEkLGYOHMAkr1OK6GDIe/t4Mdx9d3/tlP2S/OcGfQn3Bxnd/icx+5Z/xwzLNzCcwvDHgcdNq49wSL6p+scZFkJqIfwCDqB+9SYb//pz/iohXj2h/+++0nv/2v3z75zW9/g2/838LjPrg7BXcCcNzjHQC8Y5ML47zb4zX0w3xclXf1/QgL2HwPGF/A/gYWV974wRtPfvgT3PX3xo/t7hX28VyA4RcBXLD2RRQu7HCHeKDYK+7+2kevy/R/ZA72577zHNcIMVYzweLDkpi2p9z/Wn4cfxH/j7AYwOPX+Pcv/+df7Jda+Mt9HJN4JynPySf89h3rVjxHdiEf+wrAk/uMtD9bXFvxx4nxMCOE3EUL5Kr/DDeDzPhz/sTjV/w59vPP7g7Fl1Hv/undJ2//6e0n7/w3HpP+w+/t7t238B6S3731O3svGe9++uM7uNvUXrKLFxljfOcLi/lzzj/98U/xmNg/2KM2v/jFm09+ibjySxO/q/T1J6/iCyVbGOTelP3l/IpzAdr4nhk8g/Tk5W+9bPHnnvM8sObbcWFj7w7CwXCMh8vqH38dj9g/PGAe49+n/Xtwp/xvSvz38dnajqoa65y1Lxkkr46rRdwiQXpo/Lc6Xsu7zfTquFpKuYJPofy+cGIZlhixFXuPpCK7LHvpsBjIlGqSG5rY0AHYwbmg2T8wsOOLmH1F5bO/4U5YWb5b3Ct0Usu2dgfBAUeDooyQSMN95O75ziKtHdnq/My5eMDO80T2BulV8JRvpwGBYhif//hjxPbjwDnn8fAf/3tFZw3Bt0B/xUs8+dwqn8umBYsVHOR90QQpMAG1hRO0O9Y85sjJI+smL8D5jR3z82/4LHBWv+wWXThscQI83urKhYR38Z4NTij9kZQP8aI+3JXyOyyoYCL6HhYb3sfdK7zdlN+ocFHFFyf+6osMuCuDL5rjAgX3kXlyoYKLEXy0iC//4+IEJ2wsn89UcxL2KhYMuOhgEWAI4LfnwLnvOCqbaPLCCTot7LR5tPxpRLYJm7jC5u9aYb3wGHEfWH49fi548NtMf7nf3yxGL+HboJfx4lTuAxd87G4XxJuLEfxlhFdwlwfvnnkR3/581+784N0jeLwJE8sfcIKEiTR/FpAv2Hsdk6Kf4Bsl3jnCybdx4Odt4y+hDJZvJ9n2m8gO2OLFhRL7lgjHz1uiuY88fp5/xsMmglzExbn1WPmFD88/P/4TgjyXqDOIK+PEOmTHj3PMCbDnidSYkHGexnrEGJLDPySx+mGdrmFwp/9BYBBJxMPGIoSK8Z/+f8UE8eDnaY5/qJb2+Tzjv1fYp1u+neip/wjt023/1u+zn0Nfbj9vjnbEcc36HmACLkBYf1fiD2u0P7Y7S0KJTswec0FKviOFLxpnH8r+z3jIg/0q3xv1Fi50+WgPX8LNX5N5CwsqfHcKF1a4mMK7V3g3AR8F+vN779qdKa+++p0nr7/2Ou5C+fGTn+IC+E28D4zvlfoxFsLZt/8Qiy18mSr7+u/j79vfxYIKvzRY4wTv2GBfwTHKeg3sG4/PvzjwfpzHwH8WAztmUnEghp9u/JknM1b5HBN4gc0vBzgWMD4cB19/4wfrzk+8EwvDC+PPsZ53YdhdsZDPYv/HF+1zb+3LBz9UG/fYR3Bc53HzsS8uTPAOUS6OcM7DO3T5DjLercsXs/8GL+T/I+7IfReP1P4Rc6DfvvWbJ7/DoskfcaeJvxfmRRvzX+Ndo1h0450i/AntH/74J/aOHdaRN/H4zS//8Zd2dxPnCBz//bEZdG48t5gDMf4cV2w+RyMrAaT1f1A/wqIJF06oc/7B+NsiD2n42JgESRv7XxRAq4395mcskLjO/5g/5wXcst4xD0D7m/EfYbBTgA3i8rT7P+b5jYi/VRTEB8djh8TDYuCsn6Ny+Fhc3e6wGGiWapIbmtj/AfBcuKDZPzA8L+X7wsnlCPaDKoQCdbyPlspLsmTUTUUrsNC/EOQAwcrhDYoAWvy8qGfViuGZ1hld10meQS/W8lwmx9aNRQVZpRpjyv/mxZ8VyuoNOpu78//hhx/YJNDeH0KWEbFZ9Y91k3eG8FZieykY8uIkx18Ch5hhJsmLZPuzxF7hbKAlZNns7JAP5xyg4uLanxfmaO0X8sgTAzMnfPz34Qd/efL+h+/ZgggXR/hisz/juV37NSDcvcLn0Pk2eD4S8x4mHpyY8I4We8QFkxXuL9PxHSUsizoXQfzuDryHBWXxhbh8UR9nbdbuGCs2K/6jGcfPSUd8cAycHHDC9CJ+OYaLDVxowZKDLYBwIs07Wdof7hRh3PhrAfb4ESa99jwxJr18S/0ruIOGd8nwbpk3MRGi5GM2XDR5+VW/w4QLJnzcyJ4txg5aLFf7t7uDVvy1gMXj5scWtbC/9fzb4fDg1vlg+7c7R8C3025kxGAdP2sN82P86LJbynH8NtKg/1H5dpGyLiTIZ52xhRXuCMuiwB/L4x6tYkynNfzhAHv6P4XG4qPQeLCgTf/vlYp1lZXLK5iFR5tqclzq36qX4ppUkCWX0y7ggJmHfWb8febqn/fbGGPWIi/7SfbX7Ot5/oj5y2VcTLFHNnki7eoRKdlfsV/jGbaKgk3p/32BGG7w+GEy5G79n31bD7rdkedu6zPZ//Jxnz/h7ko+rso/vgfsP/Fz7//27/9u7wv7De5SeQd3EvCF3hzb+LO9fqclekq7+fGFJ7/61a9skfxHeOTnp3jUhz+ZzDsMf4I7DX6Kx39+gJd183EL3k3IRzU5zvAuDnthO8YfPgIr/HX1vxYvxgp/1pQQe36xwV9e+whjNscy/ylmnBPEjvH32EKx/wjwM9T/q/1zPsF5xEcf+0vs/8I7TD/kO8nwpQ0WSngnJhc9+MXP21gA4TvJ+LiN3j9C+Rs8bsN5Ch9r5vzj1VdfxqIYFsY4H0B8+BgNz+2bv/gZfsnuf+GdJP9oCyT8JcOfoA7wbqTvfu87iJ+//82/dED48M/vkEUIEVT7Q11V/bf6zYq7xn97P5lXdl8UYb32io18uIjoX7qxHjMv5oNT4/myqWD84SLRR/jizRbvOLFb9R9e+5Avk+PS/5ovGYasstAB7Rk6/zwY7Rqx5j8WkMeOP5bnHL9qyH7+rWu2YPumxb+eje4oKR4BlZdkyaKbilZgoT8OKi/Jkks3FW1BXzgpCRwWonwHk7nMXpwBA0TY07I1DJURsjKX8WB6bPnWMaPm2MrXyl5NyqWGvNih1XbdK66Xz04THZ+1alhsP9lhpU27Lsl0Ng+nYX2U55TPED6f8beLaJxTTlS+xcq1eiOdd0mecp5/W9mA0KAr/+c5fn/2FSmQEa6hW/1jPh/jpae8cLdbc1keBmIOxvw2j+Wxdqp8G6zJ4b7gz7+9AgKPF/GcHNs3U5gUwgSiv/jVfFhI+Pjjv9qxMh8ufvBxmY8wsbFflLGJDyZA+IaQE2hfsPFvu/jGeuatiTXjZ9ljT/htkk1AsTjChYxX8O0Of56Pzxbbty2Lw191eOUVcPGtGvm844aLJ6/wZ4IfiD+/bbN3ruBw+PHjXq2Yh45/jAc/tihFxSY1yTUn4v8tTvotB8aZGBN7zplkg+7HBcDJLf9gYEx5LqjTz+BSMv525wht+Kj8jxAr3n3Dc+jP6D9BXHkXi7/gz88jcrB8pv85xd/ijIgx/tP/s3b5h3FR+/fanPWfDF+7cJa4lhKB/Dzjn7dZ5GOJsqylNsE6XtsfnV+2fO1zlA8w5z/DXuPzec4/xxIf7/AOK/S/amu8G4T9z2n845hV4x/ftmM3/MLS6wfLX/+RlafQ+eejG2ZZdh9IcbejrVIzPf6hn2b/S1MdX/juCj6G8Z/49ZN/wx0qv+aLPv8D70/BxTXtvMD2BX58EYCx6UMu9nMRG/vNd399HxfRfLE6F935gtqf/exN3H3wJh7T+BEeAcIf32thF9rfw10t30cfzUc/+Y4wLCBx3LX94pcX+MNdIS9hrOLxmc7jwR+PLb4U0THyNHFcMd1jaG2VOtsfL3qRkvvJuDHOf8VjvFw4YbnudbvFjtnBqvGH2dMu3un8WzCx32SJy3RoqtjtHGvsCOw48MUMM8TH5iqoL5TYSas3HH/tfWXLznmBzyf8Cxq+qJ1f3nBBhI8av4PHtd7Gwshb+Lllvt/tbTy2xfeT/TfuHvn973+LR5X/bOfaHo1FbF9GBeDxc/7DX6fjO0e4GMZftOELhf/3r/7pyc/wK3b2K4S4k+QNnDfyeGz7+GtfTqBeM/58JE111Q5u9X+sfzyX9fhV2xVXthlbUEHCOv7YfAA+8uhn+ZpHsAyeQ56r/fzTZ+e/xN9PIvciz4nKl1zJpv9jQNZHddolz9zTHf9q/iyynn/q1X9qf+r/kDK4TPfUzj/7kvjUmrKMB5N5zF6cAQN4lQQ5LR1HsQEqcxkPJvOYvTgDBohy0/Lly8fCCboza/y+gzVzHYdskrQLS4p7lZVRsZioJl93+ei8uSf+8fK1gMvOyW6Rg9OrErsgDg7kyVaPH3yr9Znjyhjp2fi0RKo8aPX0zpvyef4n/l4npv5N+5v+x/vS6X85Qnwzxh+/AOZ59bPqYx+3Pv5p3mZnngMtPnP+PQoz/9DcSXXi8fMvXJ8iMeoX53n440LMe+99gAtw/EIK7qL8L9yl8Dv8/frXv8av/PyHPfrzW1yo//73b+EuFjzWgXdlcDGGNdS+BOBc0i6a/UsKPuLDu1Bew4X4T/GYxy9+6S8L5fuvXvsB34eFF7TjUSDqfPcKXzLOl6l/G48D2fvCsNDCR0S54OKPLOHuA2sP/kgq24QtxNiXBFjIwJcDdhcj9wN/tlCExR5+bBGB70pbXyD4HPbh+Scyw8oBg6RYZ/m02AIoy2F54PmCFXR86cGPLd4jrrYvWGywR4TxaC8foeE72T7EHSNcCHkb8X6b71rDu9k+wB2vH/CRGixe/QZ3Cb0NyV8S5F0zfMzqXSygfIy7TlS+LSRwQQjl8wuE7+EOEi5o8cXAv7BfssFdIz9/88k//9M/P/mnX+G9NnhZO31v4HGll/GSe553LobYFxIIKF/QP/NPb1Mz//wfOv9kA8CH/ZojU20jmySNwpLJ3lFlVCzedfx/Hsovd5ysgzJRD7DaebDVp4P//DJSA9iqWySt5QQrywtTgEj5RYCltkHGy+dAwIGU/6zKOMGxnUG3Y1EX+0uDfxvG2oVk9qE11wFdKw6viTzYVYqt2yMtTVM+gjDxn/o37Q/9A9sChfc5PoI5nv6H/SX70Ol/7dLHu03EYw0vVllWvbFKFI5nYvzxy7VVvbnT2NW78T++bODpXv3C1H/Ga+r/l23//NafCwOcu/GuD7t7ADrfFcUXrvMFsax/fG8G36f1Ef4+wLvC/swXivIOhz/+4cm//t9/tUdFeFHPC/y3sNDCl9X+AYsB9m6ND/hz9u/ZnSNcvLBu3dYiUCjO4cu42Lf3Y1j5fKfLS/bCdV7g/xB/fB8JHyfh40D+y2t4JPVFPGKCBZfv4bGR1/Hz9XzElD9pa3e4YLGGefARDv7cs90YgkLt/RuQNrf9jPkn73LgjnIRiOMPH+P199fwjlPkwJfNwP8R7i7Vu9N4/HxxLx/dfQ/H6/Hio754aT3eL8IX1qf/fcT3L+DgV+qwEMKY2jtW2I9ZmVzA4GIIy+Wjv7izEo9EvYbj/NGPfvzk57iz5+f4md83f/4P9jJWvviXd/zwLh8uivDXAb+Ll9YzJvy5ay6uYHftziHGhb8KyL6ER2GLPzwm+Nmi+OceIDtMP1ZzuMF+yXbaHyM14+/zOP5aZV+1/HT96XV9q/8426tBhCjAvF90s1qWZaP253ksT21/5qh2GiIHT/YFt5Ea4MuUXxZO+h5EATJvBXHgsXEcfueWFIIm2SVp2lQy8zMlw0Uqi3DAUA/0y5TPgZHPKrIMdZsqx44JvrzRzT28lvkEq3KWivsCjh+/H9+3eLscd5AfZswVPBsQDsePzKb8if/UPzYWthmPhGmmTvub/seXolkn+Jn+9/kef+wLAj+V2LK9r7EybAXAPfV/6n+tIU+z/dvCnK0T+GMyvDuCH79b42+ombi0wPyNiyq8O8HunsCjO3/hI6h4zwarL9df+KtCfHH3X7HIwkdjeOeEvZz0nXef/Pr//dp+7Y0/YfsWXk76O9xJwfeu2ONAf8MjQHjE8l28mJ0LLfy5W74PjBf7vMuEu2OPkq7yuW/2qIY9JsL94qM/XCTxCHHv7dEo7jc+Ntt+7PwT6Vm+2p89RsT9gO2v9vJTfwHqx1hI4fEzZrYYBcz5L2PFxQ+mt0dxYee7bvi+Lj5yy7txfvD6a09exULGd7Dg8xp+sY4/I813ydhPR+NF7LwDh+8a4/vJuDDCF7O/9j28jB1357zEZ7B4wDjWl/DCdz6K8yIe430Zv1Kn+be9mJU8Lo5gT+yunHWxwv3lweA+HUimYV62IfA5/cz/4/xbUBgXhGiuf765139+R6jOtreI2v+yifDyVja2m9WkrPWgJbGWeAaCJg/Xv5WbRTakLMIIw7NUfiycXA6vRib2PsGFny4gHbbkckqVDB5TbMsrX2H5/lwmS4wdyfJVM9Dp2qDJ2oF9oZlsfZjW7xbpdveTj/c9sNNelcmyRQaf2oDGSjflT/yn/ql9sC7YhG/an3ch0/9M/8uJ/Tdl/LHjWMNhdHsB0PpX+wfFxoWp/1P/v4L6zwWRff7Hi2m74wLV0X8lBYPQVv/sLgXWTc4FsV/O15jlg5blg7x4l8pf/vI3LCpgoQULLnpJOt/b8REK5wtu+cLTD/Crdlw4+SPuZHn/w/ft7gi+gJ2/BPP+n+nDr8PgV4H4vhW9bJ15WX7263Kf2p0bfEEqX+hOSR8Xc3iVwTs3OI3mJ+efUGJamu2Pd9xw/OWx8cW7fP+KvfSW7xXDogftfMRI72zh3SCvwvcy3i1mixdY8OHCDxc/vsMX534Hd8x8mz5/zwhfxsqX6X4HP/v85j/8zOLM94Xwzht7LAn5kcs8+QtMnD9zEenF9cgSsb0jxY7G+wjayORR8LzofWzYVSwu8U4f9CRwct81V4fDjt/uygFP8Yn+x0/l5fzTvEJpe0C+8qx23z3Wi5n/z/VPNDTUB69Az+r1H3eQbSk+3nBC3UG0l91hOluEWkzNczeL5+35eSr/BawOs5W3Tx6Od0jsePip9pbg5NvIm2rJzXZwVJMGKiaodsugbC6+zVBV5skTayd/ObwDplLKwWFz9dwGTWIMRkzFaFgys1kS09ElmzQCF7Zh8I55leXUtZ3yJ/6oS1P/rI+d9uf9ATuHVSUsLtP/TP/7TRh/OCZy3Kwfq+dR2dMTUxKOrzP+zvwD809+2/hl51/81R27iMZVM++kYNXjAgn/8T8XQ2RnXaWdL7u1essvvHAxTGHtkWnJQDqbH2MfLQ0MtqhS5n/69RSmszsemHKVxxekcrGE7whhvdcdLB/hper85Rj+rDDz5R0oH/0Niy58jAULKfYLa7DxF1m46MEX2fLneLk4A5YtDmj+yfJ5LBxhais0TSaUbZCFgc+7bezRHyygcDGDB/7Ky9+2uzx4pFxc4cLHt/BzRLbYBP6LOD6+c4SPGL2CX6F5Gb9gx8dv7A9c/rIPF1d454ntB2OJvPgyXH04N+cdLIwVD5ztn956/i2O4JHCvPmxRRVbAIOd+SEEXIDhY0Z5/MiSieyzFr/Yv2DfZv4x8w9vAWzVVvWi/ll7R7X5sv2Pap7qn7XJVdjfvf5FuyjHjx39Kq6//fhXORFsWZ+P8uOOE9/t/SiqXvE6yINJh+0ubNnj9RoDG+zlRJ1DRmstoOIvVz5fKsWBgNlz1/zZRSjcL+tC1YmsMo2UPiak1nZvpfRV1jVQ2f1tHBicylhYzixnyregTPwRBmsLqBNT/6b9Tf9jEfAe1vtcGFrbYMcx/e/zNf60ScBnjf/sCu0Mz/m3uE39fyrtn4sMnHbZxbPmXxh7rS9hjeOFt12oc85mQdeQbOXzApwzOL5QVv2P/QIO8rA8YfS7PLiAwK/RuIBBbIVaWmbIxRNOi5mJX5QtZRv/+etSOv9kEPMfE3KhwBYVedGP8vmyU19Q4YthUTZovKv5hU+J1V8yNTDL0fEzN9t/Ah4VEkIwBV86y2P2hQV38zEYIyANeX6cMK0P75Z5CYsmlmbNf1W+LbIgmd2Zg8RalHoBiylWFvKw96ngHPj+I38sjNjeY98Uf8bFHw8Czc6TF26x4TH7b0vbsVqgYbOFlTX/NjbSaVEGhcDEg8YfUq0ouc12hDB95FKzDY5HH0/Hczbzf0aI9Y/1f65/vKqwLt61P2OUOuY1zOscq+BXXf+8Gq/yVKFbBd993Cf8lfrvydwYW3Z0O4fHyf7j8tkzrHrFK+HBpJ1yF7ZfUfm2cFJXlbxg7tg6MNuD2A3bYx1yWrtlHVYKEhnlY7DgQSBzJdjI4H+F5aM8lLj2CeVz1Rn75jZ0fFY0Nqzo5NpvnHtnyCOxw8Bu8px8K/Jio/CBjnY7XBPMh8e0jh8q+5MpP4M08Z/6N+1v+p/pf9knfjPHHxszP+f475NGjpsz/s784+nOvzTP5AW6zcLWBbTdaQIT7+io8z+7pMb8D/dYmJ1z1b/hDhFbQEBau9+BCyp2dcj3fFi1dT8UFsMFGfuJbVyg80Kfyy/MkXeXUK/zH1vsgI352boJ067yvV34/JNpOM98AX5emdoCByeWTMe5Jv5zZ1gOy4/2xzTmZl9DPvfv880/7E4apGJenO2yfN5lYplgy/w+wh00XJTxf4gny+I+wvci7jhhnPnhy195/L6HeGEufLbvfG/KWjCySPEY1vyb++rnzxdd7D0wzIyxh5PvebG4cZe247dFKOwrF3ReJN/2mzT8+5zHTy4/lpTHhXJm/s+oeB2Y65/n+PpvtUtvOFbLra5blWeHgXPstZ+ouJzQLEtJYQmxWW0uHY6+9uv/bQceU77fcWJxsU3GiZkrQltBJ9VioyRNubIvbivaNl9L+bb6jN2yQZEHydHOOmdWDnT51Gm3CiPIYHjX6Wj5ObmDhy/D4ofYelTqXoAb3U3vYsqNFFP+xH/qn7WbaX/T/0z/+80bf3yUtOEvxj8bKzUg1skZr7Zm/J35x1c0/7K7NTDn4kW/jTesgxh//XodkzYuVlgV9EmbL3b4ZM7vdEAq3Alh19+sp6DxIt6y4cUB5oR2ob7mf1wI4N0V/PDC/wUuoFCiEKvqa/5nCwXMF7rKt4UCZGk/JoB8uEf8ksGbCBcQeARWpC9KMC2KsoUa3umCBHZXhh8Kd8HyoPTdIwdp/GerSDY7y4UVLmYAMvK0nz2GnR/tH4/V3V6AebmiwFLW8TMf/o8Pj535MM4WP3LJ8fPhizFgg8I7aZiex6/ymTt/BvkTvngWhfCxIsuJ+cKn4/+U9cf+IwcLGH6i2RZosnyaGc9a/ow/CAoj+RW1P+bOT61/E39W1NUfeeP6WuNva48453bmdYLYmD7npyVpyjWDi5uGMv6r2lljviY/WlqeTbnSL+5HlP8CBhHr5jJ7y2WppYhlrl6SXL9xKlO42Yl5t5apmLpanV5LIOYHZ/CmCDffOD2xpa3lo8u1PG3g0l6x87RBAnlFRJgvP6V88HyV2T135bPOWTYgsCJ8goHhW9YwmJ2nmvJ5JlatmPhP/Zv2533D9D/euVqnTzj9b4x/z+n4Y08M8Dzax8c/dHj4l7MC+WzeGDyCOf/P+/l/luY/vOBmnbILN1wx2FwNFt4ZwcdqeOHND6ck/vH6x9r6MSrnS3QsJ2syE+gRn3h0hZUYd17wvGn+Z2sToOsihfM//sovi+FiSp1/csGAie1ndcGwC39mh3J5l4uVu+S63DGjLdLgOCxP3F1hvzAD8mfF3x9lQWvEbnhkkDlKsceKkJ7l870mWL24zH89YOtYecz4cFvnvx/z4PlLPLC/iF/DIUPzXx53LZ+P9PhHR+l5M1Me/2n+zXAzc1skstx7+afj/5SPNyENXzJr55OZ8GTN+PuF48/zaSeXJx4finr+T/HX+bckbE8T/79b/bOFTDtzqP929qgQ88Nz04RZuXHzjVMsuOv1t1JRnsb/56F81F2uuZaVJqvFOmKXKyzd+AW0lp4KP1GOe23biE7j9sachM9ALT0U62TZQXIfXPEyaFp1hE5GhR01KbwBknyjr/IsuW1WVof9oNuXadxp/TIzmvJX0BBXi/QKCUJDC20T/6l/0/6m/5n+9/kff+xbXXbr3rl730/dDOzpvf/n0Gjfck//P+MfasX0/9P/T////Pf/vA7iZ65/OOI9e9d/vPDlueGu7Z8b80671Vt6KvxEOe61bSM6jdsbcxI+A7X0VPj5kuVj4WQtBbTcPe/YkrJWFLiKGGUG4R5ktqfVpZIuicW44JQ/8Z/6Zz3ItL/pf6b/vQ4Rd5YcVmb8uX67U6KWgSrGBWf8nfF3xt8Zf9FHzPxj5h8z/7gOkXeWHFZn/vFNmn8cHtWpkyXiczPJCnGoMnTqc0y+UkPcdsQ2WWMmxwweXoWa8hX9m/BN/C0wU/+m/aF7OfYw0/+swByjM/3v7cgEx4w/M/4oAsfmM+PvjL+oGDP/mPkHqsG5i0DlMMfRO+PvjL/nesNxZ+YfGn3vGpfXni/R/8ajOlYSA761025amoySuZsLuUNuyaA1Q1EKFLebliajpMgh3SG35OZeavEWKG43LU1GSZFDukNuyc291OItUNxuWpqMkiKHdIfckpt7qcVboLjdtDQZJUUO6Q65JTf3Uou3QHG7aWkySooc0h1yS27upRZvgeJ209JklBQ5pDvkltzcSy3eAsXtpqXJKClySHfILbm5l1q8BYrbTUuTUVLkkO6QW3JzL7V4CxS3m5Ymo6TIId0ht+TmXmrxFihuNy1NRkmRQ7pDbsnNvdTiLVDcblqajJIih3SH3JKbe6nFW6C43bQ0GSVFDukOuSU391KLt0Bxu2lpMkqKHNIdcktu7qUWb4HidtPSZJQUOaQ75Jbc3Est3gLF7aalySgpckh3yC25uZdavAWK201Lk1FS5JDukFtycy+1eAsUt5uWJqOkyCHdIbfk5l5q8RYobjctTUZJkUO6Q27Jzb3U4i1Q3G5amoySIod0h9ySm3upxVuguN20NBklRQ7pDrklN/dSi7dAcbtpaTJKihzSHXJLbu6lFm+B4nbT0mSUFDmkO+SW3NxLLd4Cxe2mpckoKXJId8gtubmXWrwFittNS5NRUuSQ7pBbcnMvtXgLFLebliajpMgh3SG35OZeavEWKG43LU1GSZFDukNuyc291OItUNxuWpqMkiKHdIfckpt7qcVboLjdtDQZJUUO6Q65JTf3Uou3QHG7aWkySooc0h1yS27upRZvgeJ209JklBQ5pDvkltzcSy3eAsXtpqXJKClySHfILbm5l1q8BYrbTUuTUVLkkO6QW3JzL7V4CxS3m5Ymo6TIId0ht+TmXmrxFihuNy0NIu84EUNSKQ/SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsIJ1yT7x6dkvqgQQkD+XK5JR74tWzW1IPJCCpwg7SKffEq2e3pB5IQPJQrkxOuSdePbsl9UACkirsII0S7zgJglJS8sNbUID1TI3cbi03qMghybTXT3gNhFaIslHyM+VP/FEXpv5FU1CjU0vxdiJN0q37NrwGQis02Sj5mfY37Q91YdpfNIVpf94zqKfoWre6L7fhNRBaEuImdPr4mf5n+h/Uhel/oilM/+M9Q+89pEk6Z9+G10BohSYbJT/T/0z/g7ow/U80hel/EAr8qA7e2aLOwruK0/bMkFWyp+R7Z/1njrYX44CueujzpHP6mtuZIatkTYEypvyJP19qh0rWXkyE6jL1z/vBaX+oHtP/MQi989y0mx52pbvxTv87/e/0v2gjM/7M+MtfCVyfmX/M/AuVweqDDZ3n8VPVhfLMkFWypkCaGX9n/J3x11rP0x5/7Fd1WrMLJUBvjQdNTLRTVFQ2WJdBFSEMHTR3KAE6+aCJqXIlgypCGDpo7lACdPJBE1PlSgZVhDB00NyhBOjkgyamypUMqghh6KC5QwnQyQdNTJUrGVQRwtBBc4cSoJMPmpgqVzKoIoShg+YOJUAnHzQxVa5kUEUIQwfNHUqATj5oYqpcyaCKEIYOmjuUAJ180MRUuZJBFSEMHTR3KAE6+aCJqXIlgypCGDpo7lACdPJBE1PlSgZVhDB00NyhBOjkgyamypUMqghh6KC5QwnQyQdNTJUrGVQRwtBBc4cSoJMPmpgqVzKoIoShg+YOJUAnHzQxVa5kUEUIQwfNHUqATj5oYqpcyaCKEIYOmjuUAJ180MRUuZJBFSEMHTR3KAE6+aCJqXIlgypCGDpo7lACdPJBE1PlSgZVhDB00NyhBOjkgyamypUMqghh6KC5QwnQyQdNTJUrGVQRwtBBc4cSoJMPmpgqVzKoIoShg+YOJUAnHzQxVa5kUEUIQwfNHUqATj5oYqpcyaCKEIYOmjuUAJ180MRUuZJBFSEMHTR3KAE6+aCJqXIlgypCGDpo7lACdPJBE1PlSgZVhDB00NyhBOjkgyamypUMqghh6KC5QwnQyQdNTJUrGVQRwtBBc4cSoJMPmpgqVzKoIoShg+YOJUAnHzQxVa5kUEUIQwfNHUqATj5oYqpcyaCKEIYOmjuUAJ180MRUuZJBFSEMHTR3KAE6+aCJqXIlgypCGDpo7lACdPLS8ueIl8HpPVGzXfbqmG8xZl6GsIlv+wtLsJW1jM025W+rUorcnfTo0WsIm4n/Wu0/hKzVteVvtql/U/9sFf9QeY4mrz10GcJm2t+0v/j2easzra9Zvmab/mf6n+l/tlbzkOqthwxD2Ez/O/3v9L/nNtPGmkVpthl/ZvyZ8Qc3qfN+rvh4EzE1YIBuhqY2ZIxOixwdPOgs3MILGMB4VZvyvQ1bTGpgSkQn/ozAg8Ep0Sq8gAGMV7Wpf1P/7A471IyHn/aptaZUtwssvIABpv4hAjUa0/6m/U37W22iNoyH+pWLrxpKJgEDGLFq0/6m/U37m/ZnfULtGGqXcuk1Ls5iKJkEDHDJafqf6X/+Xv3PtnCyTUxRVe9WZsXs1bq0gQ06L9mGUg12NVUchADufZgT5DXhTrahVINYTRUHIYB7H+YEecpHKF4olz0Wt0PwqqnijKSQex/miHutrZbukLiaKs6chNz7MEfcKX+PlumH4FVTxRlJIfc+zBF34r9Hy/RD8Kqp4oykkHsf5og78d+jZfoheNVUcUZSyL0Pc8Sd+O/RMv0QvGqqOCMp5N6HOeJO/PdomX4IXjVVnJEUcu/DHHEn/nu0TD8Er5oqzkgKufdhjrgT/z1aph+CV00VZySF3PswR9yJ/x4t0w/Bq6aKM5JC7n2YI+7Ef4+W6YfgVVPFGUkh9z7MEffpxz8WTu53QC82K4yAAXIPDd3Z6ZRP0pN2zW3y+ItdCiNggJoA+M5OmnySnrRrbpNnyueLzUqEAgaoAQO+s5Mmn6Qn7Zrb5Jn4T/yn/pUWEjBAbTDAd3bS5JP0pF1zmzzT/qb9TfsrLSRggNpggO/spMkn6Um75jZ5pv1N+5v2V1pIwAC1wQDf2UmTT9KTds1t8kz7m/Y37a+0kIABaoMBvrOTJp+kJ+2a2+R5FtsfFk4+wT6X+0riCALUo2i4M7TAsijNCWU9WKrbqzKjRsy4RoCTuaOecsr3Cjbxtwi0yjH1b9of+jj8n/4HYSjdfelwveOIdhNg73ZD74zpf6f/5QR7fVrlgDLj//Q/0//O+IOuYMYfdZKUraMs6mavSRbujBl/Z/yd8ffrmn/EHSelxVqzzGaY6NJ2e8sN926W7rJuIwmAWG7LUhNVtuGeJNy7WbrLuo0kAGK5LUtNVNmGe5Jw72bpLus2kgCI5bYsNVFlG+5Jwr2bpbus20gCIJbbstRElW24Jwn3bpbusm4jCYBYbstSE1W24Z4k3LtZusu6jSQAYrktS01U2YZ7knDvZuku6zaSAIjltiw1UWUb7knCvZulu6zbSAIgltuy1ESVbbgnCfdulu6ybiMJgFhuy1ITVbbhniTcu1m6y7qNJABiuS1LTVTZhnuScO9m6S7rNpIAiOW2LDVRZRvuScK9m6W7rNtIAiCW27LURJVtuCcJ926W7rJuIwmAWG7LUhNVtuGeJNy7WbrLuo0kAGK5LUtNVNmGe5Jw72bpLus2kgCI5bYsNVFlG+5Jwr2bpbus20gCIJbbstRElW24Jwn3bpbusm4jCYBYbstSE1W24Z4k3LtZusu6jSQAYrktS01U2YZ7knDvZuku6zaSAIjltiw1UWUb7knCvZulu6zbSAIgltuy1ESVbbgnCfdulu6ybiMJgFhuy1ITVbbhniTcu1m6y7qNJABiuS1LTVTZhnuScO9m6S7rNpIAiOW2LDVRZRvuScK9m6W7rNtIAiCW27LURJVtuCcJ926W7rJuIwmAWG7LUhNVtuGeJNy7WbrLuo0kAGK5LUtNVNmGe5Jw72bpLus2kgCI5bYsNVFlG+5Jwr2bpbus20gCIJbbstRElW24Jwn3bpbusm4jCYBYbstSE1W24Z4k3LtZusu6jSQAYrktS01U2YZ7knDvZuku6zaSAIjltiw1UWUb7knCvZulu6zbSAIgltuy1ESVbbgnCfdulu6ybiMJgFhuy1ITVbbhniTcu1m6y7qNJABiuS1LTVTZtnCib2EbpeRTizJcfFag6fw6AVnHkg+LgeGhV5ivPZnyfRV+4l/WjEtdcli2xTf1D8GweEz7m/5n+t8Zf9agaoJ9A/qFNiZXv+MZf2f89Zfszfgb39mWOYbDsi2+mX8gGBaPmX/M/APjSRtrWDdm/OkxmfF3j8DzOP8od5xsh9MGB/qqYV3iV1MkL5f/4Q+wsil6pNvAhVINU75NcWpIInwT/5j+RXwCTP2zUJR4RL3ZwIVSDdP+pv3hEqNWiag+0/9M/7MuP6N+BJj+10JR4hHtZgMXSjVM/zv97/S/M/6gz2gLFexDZvyd8XfGX2sWMWQGeGrzj7Jw4pmXInKppBptfC8Gg65zy0+25eQJSTqzbt1T/YEDiF8MBl3nlp8p3+OwaokpipikGCkzhopfcAOIXQwGXeeWH6Wf8jNOQpIeqbp1T/UHDiB+MRh0nVt+Jv4eh6l/WU+EJBWhlO6p/sABxC4Gg65zy8/UP4/D1L+sJ0KSilBK91R/4ABiF4NB17nlZ+qfx2HqX9YTIUlFKKV7qj9wALGLwaDr3PIz9c/jMPUv64mQpCKU0j3VHziA2MVg0HVu+Zn653GY+pf1REhSEUrpnuoPHEDsYjDoOrf8/E+of/6ojg62xMMioHtooIQrgDFys9uPuoxLdlEKWdlO+aiFXg0VuTwRGXpDQVDsILMGr9iK1AMv6yXvif/Ef+pfb17RWFY7k9jtR13GJbuY/k/hiZjCMPVv6h8iEFUjgCrJkrv9qMu4ZBelEOUJwtQ/C4Yilydi4t8iEAFS3YGc+VeGyOKjIC3ZxbQ/hUdRm/n39L8z/lhriKYRQI1kyd1+1GVcsosv1P/kHSelkdZd2oqqrsDiXEsOio+3SVyOYpjyo5PIqGVES6Sq23D6Eu0k81zcxTDxn/ivTrrWHdUQyeoTTl8i+STNc3EXw9S/qX9T/9RcQqqFSIajgPQlKm6D5rm4i2Ha37S/aX97s4l1otJSbjk5W7tQZv6LkFwfrSlRnf5n+p/pfy4dh1qI5IUAQ/oS7TzzXNzFMO3vC7W/XDipkS7xNPOuB/fquFqCnACkT9GLriex0i60Z7Lr4pUqI9MtVQTKKX/iP/Vv2l/7SrB0EHsnsutBvTquliAnAGn6v+n/Z/yrX8ln87gM67eN6uq4Wkq+gtP+pv+Z8X/G/xn/1SN2uXeiux7sq+NqCXICkGb+M/Of53n+0xdOrEJj4UVV/GYVSu46wfEGszUbqSa54ZoFwrXsctf82kuYp/zjKliNl06Wx3KLqFST3Ez8p/5N+5v+Bx0BOnl1D9YxeOeACc30/zP+rRox4++Mv4dvgWt/MfMPj4b3pVuPKtUkNzP/mvnXzL9m/oWOYOZfz/X80xdO1MFb1+6bbipagYX+OKi8JEsu3VS0Agv9cVB5SZZcuqloBRb646Dykiy5dFPRCiz0x0HlJVly6aaiFVjoj4PKS7Lk0k1FK7DQHweVl2TJpZuKVmChPw4qL8mSSzcVrcBCfxxUXpIll24qWoGF/jiovCRLLt1UtAIL/XFQeUmWXLqpaAUW+uOg8pIsuXRT0Qos9MdB5SVZcummohVY6I+Dykuy5NJNRSuw0B8HlZdkyaWbilZgoT8OKi/Jkks3Fa3AQn8cVF6SJZduKlqBhf44qLwkSy7dVLQCC/1xUHlJlly6qWgFFvrjoPKSLLl0U9EKLPTHQeUlWXLppqIVWOiPg8pLsuTSTUUrsNAfB5WXZMmlm4pWYKE/DiovyZJLNxWtwEJ/HFRekiWXbipagYX+OKi8JEsu3VS0Agv9cVB5SZZcuqloBRb646Dykiy5dFPRCiz0x0HlJVly6aaiFVjoj4PKS7Lk0k1FK7DQHweVl2TJpZuKVmChPw4qL8mSSzcVrcBCfxxUXpIll24qWoGF/jiovCRLLt1UtAIL/XFQeUmWXLqpaAv2O04iYSHKdjCZy+zFGTBArCyl5fBtp8rJTJslMunWlVHJOWCASJqWKb/GYg/pMTp3CcxenAEDTPwRYH6LnRHpeOK/R6BGavkOJvOYvTgDBoi4p2XiX2OxR/8YnbsEZi/OgAEm/gjwtP9eq7J2XGtfZy7/XQKzF2fAAFP/pv5N+0MdyBbR8bUFVua0P4vAISRpL86AASLuaZn411hM/dsjcIjOwTT1DxGwuJTgBAzwlbQ/LJx88umnuD9bd2RmcXkyZZOkR1gy2TuqjIrFw++OT/kT/3V//LmG3F94nPiqWS4ro2Kxpv5N+5v+b/r/a2+RPcT0P3cLP6ceVXG7RvTEnv53+t/pf6f/vfYW6kfUa0jSLiwp7lVWRsViTv8z/c/0P9P/eH9w7iGu879yx8lKYqImr3ZmXn3qfD6/jNQA7Xl65WuEYGV5YQrw+QstzEgNMOV7hfDwrMiYiChN/Fu9ZKRqbDxyX2QbqQGm/k39W+uFWa+sgkQt2eysadVH/Yt9IjXA1L+pf1P/1H5WyzARrQTOaie3+pT288tIDTDtb9rftD+1ndrOopVkewtTACX8QjJSA0z7m/Y37U/NZ7UME9FKpv1pvI+QOCgLJwqgy+DJDEPtaOp745xbUgiaxIoubpTMCsoMRVDmV3lhwDDlZ0c38UcsVqXyulJqjKDJqX/T/qb/mf63jjHqIKqt4wsDhhl/ZvxRO5rxd8bfmX94n+l9ZekxBU3O/GvmXzP/0rhRW0yfcXRNTSisMMz849mZf8TCyaV7qzODOHsJLvx0Aem0Sy6nVMngMcXWvUz5OTNpsXXlEq/GUYAlJ/4WAYVDcuofwuJd+qU+Tfub9qcrg9a3uHKpL42jBia5nFIlp/0hMNP+WDsu9Wn6n+l/pv9ZHedVXNpLo6iDlVxOqZLT/yIw0/+ydlzq0/S/0/9O/7s6zqt4ga842W4HKd0pGhTdK4DR317zaWnMvZE3NSkHRzVN+RP/qX8a3DXMXxtgbTPZuJJ38cNltoOjmqb9Tfub9jftjz1J7ReyZ3F08W2GTbVEZjs4qmn6n+l/pv+Z/ocdRu0XvNfJ7cW3GTbVEprt4Kim6X+m/5n+Z/ofdhi1X4g7TrwLqq6duvu2nDyDMDob23p/kTi3q5l7GVWveGV0MOnw3IXtlH+92p7436wm7xWq6hVP/bMIHEIy7c+DEtvpf6b/8XmHRj80EdSO47c5e4OqesXT/0z/gwgcqoSM7sJ2+p/pf6b/yb6XaPrfGX9m/O1twrR9QKl6xSvpwfQ/ZfyxhZO6quoHzsCs3taC4xHilh/1w2ntFmeVrSXE5lhZ2Y/lquaUv0XZgmwbmydN/Kf+9dbmdcNbW8XT/iICDAtbz/Q/EZIKpv+d8Uffqs34a50FmsfMf6yPsCHFNjP/WJ3mjL8eiKwVPSIrTCmsSWEz42/GpKAZf2f8nfG39iFsHM/2+Ot3nFgPWLpB6+hy30sbv4UtSVOuSS5uGl6wjX+JciFc89gtLUlTduZlauiGKZ8nYeLPKHxG/bnWqK1OfUb6i5uGqX9T/6b9Tf+jroCdjOYSxJ/xaX1KU64JL+7pf6b/nfEHDWXmP9EVsNuY/ufaed5YWp/alGuCizuCPvUvQsGwTf27Vp4bS6tTTbkmuLgj6FP/IhQM2wP17wW84oQ3c5aPJV16CfEyVy9Jrt84lSvcn2JgzncrK5fLK4kiR09KHj/Yw5si3Hzj9MSWdsqf+E/9U0tf7QWNqtYKby7yUSPmZ9rf9D+tGlit4MZry6ozteoEw0nT/9aWpkBN+6tR8Sqj2FAj5mf6n+l/WjWwWsGN15ZVZ2rVCYaTpv+pLU2Bmv6nRsWrjGJDjZif6X+m/2nVwGoFN15bVp2pVScYTpr+p7Y0Ber57X9wx8knOIqy0qRrq3LidZjF9IVgS0+FnyjHvbZtRKdxe2NOwmeglp4KP1O+x6E2/Rao5Ya4MSfhM1BLP/H3aE39W7XGa4dtW0XJSnVjTsJnoJaeCj8Tf4/D+17OmwAADm9JREFUtH+rDFZHWkVZ4YG4MSfhM1BLT4WfqX8eh6l/VhmsjrSKssIDcWNOwmeglp4KP1P/PA5T/6wyWB1pFWWFB+LGnITPQC09FX6m/nkcpv5ZZbA60irKCg/EjTkJn4Faeir8TP3zOEz9s8pgdaRVlBUeiJM5Xw578iqtXqYEzuldY6KdZGZ7Wl0qKZJYjAtO+WjoaOkT/6l/qAbR519bysWSzWraX13zfiBQF1e8TG7a37S/aX/T/1x7iFvL9L8ar2b8mfEn7/m9NJhsKBfXjL8Izsz/5/pn5p/PzPzz8KjO6rdssYL4fJn2UD9nSzTq/o7JV+qHKsKUv0J/DOBxFUwhn/hHJG6q79Q/C8y0v/uOePqf6X+s653+t/SmAVcPGnoDdOpzDN/0v9P/omLM+DPjD6rBuYtA5TDH0TvzX/Sv58jAMf2vRp+bIM34Y4GZ/vfR/W88qmM1bdWnrHXeBrOBZoWzVnvge1p3yC0Z+TZDUQoUt5uWJqOkyCHdIbfk5l5q8RYobjctTUZJkUO6Q27Jzb3U4i1Q3G5amoySIod0h9ySm3upxVuguN20NBklRQ7pDrklN/dSi7dAcbtpaTJKihzSHXJLbu6lFm+B4nbT0mSUFDmkO+SW3NxLLd4Cxe2mpckoKXJId8gtubmXWrwFittNS5NRUuSQ7pBbcnMvtXgLFLebliajpMgh3SG35OZeavEWKG43LU1GSZFDukNuyc291OItUNxuWpqMkiKHdIfckpt7qcVboLjdtDQZJUUO6Q65JTf3Uou3QHG7aWkySooc0h1yS27upRZvgeJ209JklBQ5pDvkltzcSy3eAsXtpqXJKClySHfILbm5l1q8BYrbTUuTUVLkkO6QW3JzL7V4CxS3m5Ymo6TIId0ht+TmXmrxFihuNy1NRkmRQ7pDbsnNvdTiLVDcblqajJIih3SH3JKbe6nFW6C43bQ0GSVFDukOuSU391KLt0Bxu2lpMkqKHNIdcktu7qUWb4HidtPSZJQUOaQ75Jbc3Est3gLF7aalySgpckh3yC25uZdavAWK201Lk1FS5JDukFtycy+1eAsUt5uWJqOkyCHdIbfk5l5q8RYobjctTUZJkUO6Q27Jzb3U4i1Q3G5amoySIod0h9ySm3upxVuguN20NBklRQ7pDrklN/dSi7dAcbtpaTJKihzSHXJLbu6lFm+B4nbT0mSUFDmkO+SW3NxLLd4Cxe2mpckoKXJId8gtubmXWrwFittNS5NRUuSQ7pBbcnMvtXgLFLebliajpMgh3SG35OZeavEWKG43LQ0i7zgRQ1IpD9Ip98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgnXJPvHp2S+qBBCQP5crklHvi1bNbUg8kIKnCDtIp98SrZ7ekHkhA8lCuTE65J149uyX1QAKSKuwgjYIf1aEsH6WUmfebAOvlJnK7tdwuJodkybLA8BoI7cCgj58pf+KPujD1L5qCGl1vPdIkvfXs2/AaCK3QZKPkZ9rftD/UhWl/0RSm/XnPoJ6ia93qvtyG10BoSYib8OnjZ/qf6X9QF6b/iaYw/Y/3DL33kCbpnH0bXgOhFZpslPxM/zP9D+rC9D/RFKb/QSjwozp4Z5U6C+8qTtszQ1bJnpJrMi/wpUaYDLUXY4GueujzpHP6mtuZIatkTYEypvyJ/9S/aX/T/0z/i9GeI5F9MFzM+OPzoBl/USNm/sMgeNu42d7MsFa6G+/Mv2b+NfMvtJG5/pnrv5l/xAjzDZh/2a/qtGEvlAA3Q2maxcQ46S9/XjIYIoShg+YOJUAnHzQxp/yJv718HBXCxmvVFVUQ6Zts7lACbOyrKubUv6l/U/+4YD3tb/qf0k+qgyymCps7lACVesRiqt5JBlmEMHTQ3KEE6OSDJqbKlQyqCGHooLlDCdDJB01MlSsZVBHC0EFzhxKgkw+amCpXMqgihKGD5g4lQCcfNDFVrmRQRQhDB80dSoBOPmhiqlzJoIoQhg6aO5QAnXzQxFS5kkEVIQwdNHcoATr5oImpciWDKkIYOmjuUAJ08kETU+VKBlWEMHTQ3KEE6OSDJqbKlQyqCGHooLlDCdDJB01MlSsZVBHC0EFzhxKgkw+amCpXMqgihKGD5g4lQCcfNDFVrmRQRQhDB80dSoBOPmhiqlzJoIoQhg6aO5QAnXzQxFS5kkEVIQwdNHcoATr5oImpciWDKkIYOmjuUAJ08kETU+VKBlWEMHTQ3KEE6OSl5c8RL4PTe6Jmu+zVMd9izLwMYRPf9hWWYCtrGZttyt+uShS5O+nRo9cQNhN/hPAmXK2uLU6zTf2b+teuim8qUpi99lA1hM20v2l/0/9EA2mg9bXL02zT/07/O/1vazMPK956yDGEzYw/M/7M+HNuNW2sWZRmm/Fnxh+MP9vCiVcRqy8BA3QzNNUhY3TaqnISDzpFYo74W006YADjVW3K9zpsMamBKRF1+KCzsAsvYICJPyJQozH1b+of5/BWJ2rFKC3K4YPOwi68gAGMV7Wpf1P/pv5N+7M+oXYMpUeZ/ocReDA4JVqFFzCA8ao2/e/0v9P/Tv9rfULtGEqP4vBBZ2EXXsAA0/8gAjUaf8/+d1s42XYMO3q3MilmPZBSAy7Qeck2lGrwq6niIARw78OcIK+AJ9tQqkGspoqDEMC9D3OCPOUjFC+Uam9xOwSvmirOSAq592GOuNfaaukOiaup4sxJyL0Pc8Sd8vdomX4IXjVVnJEUcu/DHHEn/nu0TD8Er5oqzkgKufdhjrgT/z1aph+CV00VZySF3PswR9yJ/x4t0w/Bq6aKM5JC7n2YI+7Ef4+W6YfgVVPFGUkh9z7MEXfiv0fL9EPwqqnijKSQex/miDvx36Nl+iF41VRxRlLIvQ9zxJ3479Ey/RC8aqo4Iynk3oc54k7892iZfgheNVWckRRy78MccZ9+/GPh5H4H9GKjwggYIPfQ0J2dTvkkPWnX3CaPv1ioMAIGqAmA7+ykySfpSbvmNnmmfL7YqEQoYIAaMOA7O2nySXrSrrlNnon/xH/qX2khAQPUBgN8ZydNPklP2jW3yTPtb9rftL/SQgIGqA0G+M5OmnySnrRrbpNn2t+0v2l/pYUEDFAbDPCdnTT5JD1p19wmz7S/aX/T/koLCRigNhjgOztp8kl60q65TZ5nsf1h4eQT7HO5rySOIEA9ioY7Qwssi9KcUNaDlbq9JjNqxIxrBDiZO+opp3yvYBN/i0CrHFP/pv2hj+NjNagKvL02P62iTP8T4QiQodpQZ0z/O/0vJ9jr0yoHlBn/p/+Z/nfGH3QFM/6qk6RsHWVRN3tNsnBnzPg74++Mv1/X/CPuOCkt1pplNsNEl7bbW264d7N0l3UbSQDEcluWmqiyDfck4d7N0l3WbSQBEMttWWqiyjbck4R7N0t3WbeRBEAst2WpiSrbcE8S7t0s3WXdRhIAsdyWpSaqbMM9Sbh3s3SXdRtJAMRyW5aaqLIN9yTh3s3SXdZtJAEQy21ZaqLKNtyThHs3S3dZt5EEQCy3ZamJKttwTxLu3SzdZd1GEgCx3JalJqpswz1JuHezdJd1G0kAxHJblpqosg33JOHezdJd1m0kARDLbVlqoso23JOEezdLd1m3kQRALLdlqYkq23BPEu7dLN1l3UYSALHclqUmqmzDPUm4d7N0l3UbSQDEcluWmqiyDfck4d7N0l3WbSQBEMttWWqiyjbck4R7N0t3WbeRBEAst2WpiSrbcE8S7t0s3WXdRhIAsdyWpSaqbMM9Sbh3s3SXdRtJAMRyW5aaqLIN9yTh3s3SXdZtJAEQy21ZaqLKNtyThHs3S3dZt5EEQCy3ZamJKttwTxLu3SzdZd1GEgCx3JalJqpswz1JuHezdJd1G0kAxHJblpqosg33JOHezdJd1m0kARDLbVlqoso23JOEezdLd1m3kQRALLdlqYkq23BPEu7dLN1l3UYSALHclqUmqmzDPUm4d7N0l3UbSQDEcluWmqiyDfck4d7N0l3WbSQBEMttWWqiyjbck4R7N0t3WbeRBEAst2WpiSrbcE8S7t0s3WXdRhIAsdyWpSaqbMM9Sbh3s3SXdRtJAMRyW5aaqLIN9yTh3s3SXdZtJAEQy21ZaqLKNtyThHs3S3dZt5EEQCy3ZamJKttwTxLu3SzdZd1GEgCx3JalJqpswz1JuHezdJd1G0kAxHJblpqosg33JOHezdJd1m0kARDLbVlqosq2hRN9C9soJZ9alOHiswJN59cJyDqWfFgMDA+9wnvtyZTvq/AT/7JmXOqSw7Itvql/CIbFY9rf9D/T/874swZVE+wb0C+0Mbn6Hc/4O+Ovv+Ryxt/4zrbMMRyWbfHN/APBsHjM/GPmHxhP2ljDujHjT4/JjL97BJ7H+cf/B5kwmOpoRVvFAAAAAElFTkSuQmCC) Instead of going-----> : **wide -> narrow -> wide** ⏬ in normal bottleneck block, they do the opposite---> **narrow -> wide -> narrow** **reference**:[MobileNetV2: Inverted Residuals and Linear Bottlenecks](https://arxiv.org/abs/1801.04381) ###Code !nvidia-smi import torch import torch.nn as nn import torch.optim as optim import torch.nn.functional as F import torch.backends.cudnn as cudnn import glob import cv2 import shutil from PIL import Image import matplotlib.pyplot as plt import pandas as pd import numpy as np import torchvision import torchvision.transforms as transforms import torchvision.datasets as datasets from torch.utils.data import DataLoader,Dataset from torchvision.models import resnet18, resnet34 ,resnet50 import os import torchvision.models as models device = 'cuda' if torch.cuda.is_available() else 'cpu' best_acc = 0 # best test accuracy start_epoch = 0 # start from epoch 0 or last checkpoint epoch use_gpu = torch.cuda.is_available() use_gpu from typing import Optional from functools import partial class ConvNormAct(nn.Sequential): def __init__( self, in_features: int, out_features: int, kernel_size: int, norm: nn.Module = nn.BatchNorm2d, act: nn.Module = nn.ReLU, **kwargs ): super().__init__( nn.Conv2d( in_features, out_features, kernel_size=kernel_size, padding=kernel_size // 2, ), norm(out_features), act(), ) # -------------------------------------------------------------------- Conv1X1BnReLU = partial(ConvNormAct, kernel_size=1) Conv3X3BnReLU = partial(ConvNormAct, kernel_size=3) # --------------------------------------------------------------------- class ResidualAdd(nn.Module): def __init__(self, block: nn.Module, shortcut: Optional[nn.Module] = None): super().__init__() self.block = block self.shortcut = shortcut def forward(self,x): res = x x = self.block(x) if self.shortcut: res = self.shortcut(res) x += res return x # ---------------------------------------------------------------------- # class BottleNeck(nn.Sequential): # def __init__(self, in_features: int, out_features: int, reduction: int = 4): # reduced_features = out_features // reduction # super().__init__( # nn.Sequential( # ResidualAdd( # nn.Sequential( # # wide -> narrow # Conv1X1BnReLU(in_features, reduced_features), # # narrow -> narrow # Conv3X3BnReLU(reduced_features, reduced_features), # # narrow -> wide # Conv1X1BnReLU(reduced_features, out_features, act=nn.Identity), # ), # shortcut=Conv1X1BnReLU(in_features, out_features) # if in_features != out_features # else None, # ), # nn.ReLU(), # ) # ) # --------------------------------------------------------------------- class InvertedResidual(nn.Sequential): def __init__(self, in_features: int, out_features: int, expansion: int = 4): expanded_features = in_features * expansion super().__init__( nn.Sequential( ResidualAdd( nn.Sequential( # narrow -> wide Conv1X1BnReLU(in_features, expanded_features), # wide -> wide Conv3X3BnReLU(expanded_features, expanded_features), # wide -> narrow Conv1X1BnReLU(expanded_features, out_features, act=nn.Identity), ), shortcut=Conv1X1BnReLU(in_features, out_features) if in_features != out_features else None, ), nn.ReLU(), ) ) img_BNeck = Image.open(str('/content/sample_data/cat.jpeg')) plt.imshow(img_BNeck) transform = transforms.Compose([ transforms.Resize((224, 224)), transforms.ToTensor(), transforms.Normalize(mean=0., std=1.) ]) img_BNeck = transform(img_BNeck) print(f"Image shape before: {img_BNeck.shape}") img_BNeck = img_BNeck.unsqueeze(0) print(f"Image shape after add dim: {img_BNeck.shape}") img_BNeck = img_BNeck.to(device) invertedresidual = InvertedResidual(3,6) invertedresidual invertedresidual.to(device) invertedresidual(img_BNeck).shape ir = invertedresidual(img_BNeck) shortcut_image = ir.squeeze(0) shortcut_image.shape shortcut_image = torch.sum(shortcut_image/6,0)# decrease by sum dim shortcut_image = shortcut_image.data.cpu().numpy() shortcut_image.shape #---> (224, 224) plt.imshow(shortcut_image) ###Output _____no_output_____ ###Markdown In the following class **MobileNetLikeBlock**, if only the input and output are not the same, the residual_block is used ###Code class ResidualAdd(nn.Module): def __init__(self, block: nn.Module, shortcut: Optional[nn.Module] = None): super().__init__() self.block = block self.shortcut = shortcut def forward(self,x): res = x x = self.block(x) if self.shortcut: res = self.shortcut(res) x += res return x class MobileNetLikeBlock(nn.Sequential): def __init__(self, in_features: int, out_features: int, expansion: int = 4): # use ResidualAdd if features match, otherwise a normal Sequential residual = ResidualAdd if in_features == out_features else nn.Sequential expanded_features = in_features * expansion super().__init__( nn.Sequential( residual( nn.Sequential( # narrow -> wide Conv1X1BnReLU(in_features, expanded_features), # wide -> wide Conv3X3BnReLU(expanded_features, expanded_features), # wide -> narrow Conv1X1BnReLU(expanded_features, out_features, act=nn.Identity), ), ), nn.ReLU(), ) ) # import torch # x = torch.randn((1, 32, 56, 56)) # Conv1X1BnReLU(32, 64)(x).shape MLB=MobileNetLikeBlock(3, 6) MLB.to(device) MLB # MLB(img_BNeck).shape # MobileNetLikeBlock(32, 32)(x).shape MLB = MLB(img_BNeck) shortcut_image = MLB.squeeze(0) shortcut_image.shape shortcut_image = torch.sum(shortcut_image/6,0)# decrease by sum dim shortcut_image = shortcut_image.data.cpu().numpy() shortcut_image.shape #---> (224, 224) plt.imshow(shortcut_image) MLB1=MobileNetLikeBlock(3, 3) MLB1 MLB1.to(device) MLB1 = MLB1(img_BNeck) shortcut_image = MLB1.squeeze(0) shortcut_image.shape shortcut_image = torch.sum(shortcut_image/3,0)# decrease by sum dim shortcut_image = shortcut_image.data.cpu().numpy() shortcut_image.shape #---> (224, 224) plt.imshow(shortcut_image) ###Output _____no_output_____
c4_convolutional_neural_networks/week_10/Convolution_model_Step_by_Step_v2a.ipynb
###Markdown Convolutional Neural Networks: Step by StepWelcome to Course 4's first assignment! In this assignment, you will implement convolutional (CONV) and pooling (POOL) layers in numpy, including both forward propagation and (optionally) backward propagation. **Notation**:- Superscript $[l]$ denotes an object of the $l^{th}$ layer. - Example: $a^{[4]}$ is the $4^{th}$ layer activation. $W^{[5]}$ and $b^{[5]}$ are the $5^{th}$ layer parameters.- Superscript $(i)$ denotes an object from the $i^{th}$ example. - Example: $x^{(i)}$ is the $i^{th}$ training example input. - Subscript $i$ denotes the $i^{th}$ entry of a vector. - Example: $a^{[l]}_i$ denotes the $i^{th}$ entry of the activations in layer $l$, assuming this is a fully connected (FC) layer. - $n_H$, $n_W$ and $n_C$ denote respectively the height, width and number of channels of a given layer. If you want to reference a specific layer $l$, you can also write $n_H^{[l]}$, $n_W^{[l]}$, $n_C^{[l]}$. - $n_{H_{prev}}$, $n_{W_{prev}}$ and $n_{C_{prev}}$ denote respectively the height, width and number of channels of the previous layer. If referencing a specific layer $l$, this could also be denoted $n_H^{[l-1]}$, $n_W^{[l-1]}$, $n_C^{[l-1]}$. We assume that you are already familiar with `numpy` and/or have completed the previous courses of the specialization. Let's get started! Updates If you were working on the notebook before this update...* The current notebook is version "v2a".* You can find your original work saved in the notebook with the previous version name ("v2") * To view the file directory, go to the menu "File->Open", and this will open a new tab that shows the file directory. List of updates* clarified example used for padding function. Updated starter code for padding function.* `conv_forward` has additional hints to help students if they're stuck.* `conv_forward` places code for `vert_start` and `vert_end` within the `for h in range(...)` loop; to avoid redundant calculations. Similarly updated `horiz_start` and `horiz_end`. **Thanks to our mentor Kevin Brown for pointing this out.*** `conv_forward` breaks down the `Z[i, h, w, c]` single line calculation into 3 lines, for clarity.* `conv_forward` test case checks that students don't accidentally use n_H_prev instead of n_H, use n_W_prev instead of n_W, and don't accidentally swap n_H with n_W* `pool_forward` properly nests calculations of `vert_start`, `vert_end`, `horiz_start`, and `horiz_end` to avoid redundant calculations.* `pool_forward' has two new test cases that check for a correct implementation of stride (the height and width of the previous layer's activations should be large enough relative to the filter dimensions so that a stride can take place). * `conv_backward`: initialize `Z` and `cache` variables within unit test, to make it independent of unit testing that occurs in the `conv_forward` section of the assignment.* **Many thanks to our course mentor, Paul Mielke, for proposing these test cases.** 1 - PackagesLet's first import all the packages that you will need during this assignment. - [numpy](www.numpy.org) is the fundamental package for scientific computing with Python.- [matplotlib](http://matplotlib.org) is a library to plot graphs in Python.- np.random.seed(1) is used to keep all the random function calls consistent. It will help us grade your work. ###Code import numpy as np import h5py import matplotlib.pyplot as plt %matplotlib inline plt.rcParams['figure.figsize'] = (5.0, 4.0) # set default size of plots plt.rcParams['image.interpolation'] = 'nearest' plt.rcParams['image.cmap'] = 'gray' %load_ext autoreload %autoreload 2 np.random.seed(1) ###Output _____no_output_____ ###Markdown 2 - Outline of the AssignmentYou will be implementing the building blocks of a convolutional neural network! Each function you will implement will have detailed instructions that will walk you through the steps needed:- Convolution functions, including: - Zero Padding - Convolve window - Convolution forward - Convolution backward (optional)- Pooling functions, including: - Pooling forward - Create mask - Distribute value - Pooling backward (optional) This notebook will ask you to implement these functions from scratch in `numpy`. In the next notebook, you will use the TensorFlow equivalents of these functions to build the following model:**Note** that for every forward function, there is its corresponding backward equivalent. Hence, at every step of your forward module you will store some parameters in a cache. These parameters are used to compute gradients during backpropagation. 3 - Convolutional Neural NetworksAlthough programming frameworks make convolutions easy to use, they remain one of the hardest concepts to understand in Deep Learning. A convolution layer transforms an input volume into an output volume of different size, as shown below. In this part, you will build every step of the convolution layer. You will first implement two helper functions: one for zero padding and the other for computing the convolution function itself. 3.1 - Zero-PaddingZero-padding adds zeros around the border of an image: **Figure 1** : **Zero-Padding** Image (3 channels, RGB) with a padding of 2. The main benefits of padding are the following:- It allows you to use a CONV layer without necessarily shrinking the height and width of the volumes. This is important for building deeper networks, since otherwise the height/width would shrink as you go to deeper layers. An important special case is the "same" convolution, in which the height/width is exactly preserved after one layer. - It helps us keep more of the information at the border of an image. Without padding, very few values at the next layer would be affected by pixels as the edges of an image.**Exercise**: Implement the following function, which pads all the images of a batch of examples X with zeros. [Use np.pad](https://docs.scipy.org/doc/numpy/reference/generated/numpy.pad.html). Note if you want to pad the array "a" of shape $(5,5,5,5,5)$ with `pad = 1` for the 2nd dimension, `pad = 3` for the 4th dimension and `pad = 0` for the rest, you would do:```pythona = np.pad(a, ((0,0), (1,1), (0,0), (3,3), (0,0)), mode='constant', constant_values = (0,0))``` ###Code # GRADED FUNCTION: zero_pad def zero_pad(X, pad): """ Pad with zeros all images of the dataset X. The padding is applied to the height and width of an image, as illustrated in Figure 1. Argument: X -- python numpy array of shape (m, n_H, n_W, n_C) representing a batch of m images pad -- integer, amount of padding around each image on vertical and horizontal dimensions Returns: X_pad -- padded image of shape (m, n_H + 2*pad, n_W + 2*pad, n_C) """ ### START CODE HERE ### (≈ 1 line) X_pad = np.pad(X, ((0,0), (pad, pad), (pad, pad), (0,0)), mode='constant', constant_values = (0,0)) ### END CODE HERE ### return X_pad np.random.seed(1) x = np.random.randn(4, 3, 3, 2) x_pad = zero_pad(x, 2) print ("x.shape =\n", x.shape) print ("x_pad.shape =\n", x_pad.shape) print ("x[1,1] =\n", x[1,1]) print ("x_pad[1,1] =\n", x_pad[1,1]) fig, axarr = plt.subplots(1, 2) axarr[0].set_title('x') axarr[0].imshow(x[0,:,:,0]) axarr[1].set_title('x_pad') axarr[1].imshow(x_pad[0,:,:,0]) ###Output x.shape = (4, 3, 3, 2) x_pad.shape = (4, 7, 7, 2) x[1,1] = [[ 0.90085595 -0.68372786] [-0.12289023 -0.93576943] [-0.26788808 0.53035547]] x_pad[1,1] = [[0. 0.] [0. 0.] [0. 0.] [0. 0.] [0. 0.] [0. 0.] [0. 0.]] ###Markdown **Expected Output**:```x.shape = (4, 3, 3, 2)x_pad.shape = (4, 7, 7, 2)x[1,1] = [[ 0.90085595 -0.68372786] [-0.12289023 -0.93576943] [-0.26788808 0.53035547]]x_pad[1,1] = [[ 0. 0.] [ 0. 0.] [ 0. 0.] [ 0. 0.] [ 0. 0.] [ 0. 0.] [ 0. 0.]]``` 3.2 - Single step of convolution In this part, implement a single step of convolution, in which you apply the filter to a single position of the input. This will be used to build a convolutional unit, which: - Takes an input volume - Applies a filter at every position of the input- Outputs another volume (usually of different size) **Figure 2** : **Convolution operation** with a filter of 3x3 and a stride of 1 (stride = amount you move the window each time you slide) In a computer vision application, each value in the matrix on the left corresponds to a single pixel value, and we convolve a 3x3 filter with the image by multiplying its values element-wise with the original matrix, then summing them up and adding a bias. In this first step of the exercise, you will implement a single step of convolution, corresponding to applying a filter to just one of the positions to get a single real-valued output. Later in this notebook, you'll apply this function to multiple positions of the input to implement the full convolutional operation. **Exercise**: Implement conv_single_step(). [Hint](https://docs.scipy.org/doc/numpy-1.13.0/reference/generated/numpy.sum.html). **Note**: The variable b will be passed in as a numpy array. If we add a scalar (a float or integer) to a numpy array, the result is a numpy array. In the special case when a numpy array contains a single value, we can cast it as a float to convert it to a scalar. ###Code # GRADED FUNCTION: conv_single_step def conv_single_step(a_slice_prev, W, b): """ Apply one filter defined by parameters W on a single slice (a_slice_prev) of the output activation of the previous layer. Arguments: a_slice_prev -- slice of input data of shape (f, f, n_C_prev) W -- Weight parameters contained in a window - matrix of shape (f, f, n_C_prev) b -- Bias parameters contained in a window - matrix of shape (1, 1, 1) Returns: Z -- a scalar value, the result of convolving the sliding window (W, b) on a slice x of the input data """ ### START CODE HERE ### (≈ 2 lines of code) # Element-wise product between a_slice_prev and W. Do not add the bias yet. s = np.multiply(a_slice_prev, W) # Sum over all entries of the volume s. Z = np.sum(s) # Add bias b to Z. Cast b to a float() so that Z results in a scalar value. Z = Z + float(b) ### END CODE HERE ### return Z np.random.seed(1) a_slice_prev = np.random.randn(4, 4, 3) W = np.random.randn(4, 4, 3) b = np.random.randn(1, 1, 1) Z = conv_single_step(a_slice_prev, W, b) print("Z =", Z) ###Output Z = -6.999089450680221 ###Markdown **Expected Output**: **Z** -6.99908945068 3.3 - Convolutional Neural Networks - Forward passIn the forward pass, you will take many filters and convolve them on the input. Each 'convolution' gives you a 2D matrix output. You will then stack these outputs to get a 3D volume: **Exercise**: Implement the function below to convolve the filters `W` on an input activation `A_prev`. This function takes the following inputs:* `A_prev`, the activations output by the previous layer (for a batch of m inputs); * Weights are denoted by `W`. The filter window size is `f` by `f`.* The bias vector is `b`, where each filter has its own (single) bias. Finally you also have access to the hyperparameters dictionary which contains the stride and the padding. **Hint**: 1. To select a 2x2 slice at the upper left corner of a matrix "a_prev" (shape (5,5,3)), you would do:```pythona_slice_prev = a_prev[0:2,0:2,:]```Notice how this gives a 3D slice that has height 2, width 2, and depth 3. Depth is the number of channels. This will be useful when you will define `a_slice_prev` below, using the `start/end` indexes you will define.2. To define a_slice you will need to first define its corners `vert_start`, `vert_end`, `horiz_start` and `horiz_end`. This figure may be helpful for you to find out how each of the corner can be defined using h, w, f and s in the code below. **Figure 3** : **Definition of a slice using vertical and horizontal start/end (with a 2x2 filter)** This figure shows only a single channel. **Reminder**:The formulas relating the output shape of the convolution to the input shape is:$$ n_H = \lfloor \frac{n_{H_{prev}} - f + 2 \times pad}{stride} \rfloor +1 $$$$ n_W = \lfloor \frac{n_{W_{prev}} - f + 2 \times pad}{stride} \rfloor +1 $$$$ n_C = \text{number of filters used in the convolution}$$For this exercise, we won't worry about vectorization, and will just implement everything with for-loops. Additional Hints if you're stuck* You will want to use array slicing (e.g.`varname[0:1,:,3:5]`) for the following variables: `a_prev_pad` ,`W`, `b` Copy the starter code of the function and run it outside of the defined function, in separate cells. Check that the subset of each array is the size and dimension that you're expecting. * To decide how to get the vert_start, vert_end; horiz_start, horiz_end, remember that these are indices of the previous layer. Draw an example of a previous padded layer (8 x 8, for instance), and the current (output layer) (2 x 2, for instance). The output layer's indices are denoted by `h` and `w`. * Make sure that `a_slice_prev` has a height, width and depth.* Remember that `a_prev_pad` is a subset of `A_prev_pad`. Think about which one should be used within the for loops. ###Code # GRADED FUNCTION: conv_forward def conv_forward(A_prev, W, b, hparameters): """ Implements the forward propagation for a convolution function Arguments: A_prev -- output activations of the previous layer, numpy array of shape (m, n_H_prev, n_W_prev, n_C_prev) W -- Weights, numpy array of shape (f, f, n_C_prev, n_C) b -- Biases, numpy array of shape (1, 1, 1, n_C) hparameters -- python dictionary containing "stride" and "pad" Returns: Z -- conv output, numpy array of shape (m, n_H, n_W, n_C) cache -- cache of values needed for the conv_backward() function """ ### START CODE HERE ### # Retrieve dimensions from A_prev's shape (≈1 line) (m, n_H_prev, n_W_prev, n_C_prev) = A_prev.shape # Retrieve dimensions from W's shape (≈1 line) (f, f, n_C_prev, n_C) = W.shape # Retrieve information from "hparameters" (≈2 lines) stride = hparameters['stride'] pad = hparameters['pad'] # Compute the dimensions of the CONV output volume using the formula given above. # Hint: use int() to apply the 'floor' operation. (≈2 lines) n_H = int((n_H_prev - f + 2*pad)/stride) + 1 n_W = int((n_W_prev - f + 2*pad)/stride) + 1 # Initialize the output volume Z with zeros. (≈1 line) Z = np.zeros((m, n_H, n_W, n_C)) # Create A_prev_pad by padding A_prev A_prev_pad = zero_pad(A_prev, pad) for i in range(m): # loop over the batch of training examples a_prev_pad = A_prev_pad[i] # Select ith training example's padded activation for h in range(n_H): # loop over vertical axis of the output volume # Find the vertical start and end of the current "slice" (≈2 lines) vert_start = h * stride vert_end = vert_start + f for w in range(n_W): # loop over horizontal axis of the output volume # Find the horizontal start and end of the current "slice" (≈2 lines) horiz_start = w * stride horiz_end = horiz_start + f for c in range(n_C): # loop over channels (= #filters) of the output volume # Use the corners to define the (3D) slice of a_prev_pad (See Hint above the cell). (≈1 line) a_slice_prev = a_prev_pad[vert_start:vert_end, horiz_start:horiz_end, :] # Convolve the (3D) slice with the correct filter W and bias b, to get back one output neuron. (≈3 line) weights = W[:, :, :, c] biases = b[:, :, :, c] Z[i, h, w, c] = conv_single_step(a_slice_prev, weights, biases) ### END CODE HERE ### # Making sure your output shape is correct assert(Z.shape == (m, n_H, n_W, n_C)) # Save information in "cache" for the backprop cache = (A_prev, W, b, hparameters) return Z, cache np.random.seed(1) A_prev = np.random.randn(10,5,7,4) W = np.random.randn(3,3,4,8) b = np.random.randn(1,1,1,8) hparameters = {"pad" : 1, "stride": 2} Z, cache_conv = conv_forward(A_prev, W, b, hparameters) print("Z's mean =\n", np.mean(Z)) print("Z[3,2,1] =\n", Z[3,2,1]) print("cache_conv[0][1][2][3] =\n", cache_conv[0][1][2][3]) ###Output Z's mean = 0.6923608807576933 Z[3,2,1] = [-1.28912231 2.27650251 6.61941931 0.95527176 8.25132576 2.31329639 13.00689405 2.34576051] cache_conv[0][1][2][3] = [-1.1191154 1.9560789 -0.3264995 -1.34267579] ###Markdown **Expected Output**:```Z's mean = 0.692360880758Z[3,2,1] = [ -1.28912231 2.27650251 6.61941931 0.95527176 8.25132576 2.31329639 13.00689405 2.34576051]cache_conv[0][1][2][3] = [-1.1191154 1.9560789 -0.3264995 -1.34267579]``` Finally, CONV layer should also contain an activation, in which case we would add the following line of code:```python Convolve the window to get back one output neuronZ[i, h, w, c] = ... Apply activationA[i, h, w, c] = activation(Z[i, h, w, c])```You don't need to do it here. 4 - Pooling layer The pooling (POOL) layer reduces the height and width of the input. It helps reduce computation, as well as helps make feature detectors more invariant to its position in the input. The two types of pooling layers are: - Max-pooling layer: slides an ($f, f$) window over the input and stores the max value of the window in the output.- Average-pooling layer: slides an ($f, f$) window over the input and stores the average value of the window in the output.These pooling layers have no parameters for backpropagation to train. However, they have hyperparameters such as the window size $f$. This specifies the height and width of the $f \times f$ window you would compute a *max* or *average* over. 4.1 - Forward PoolingNow, you are going to implement MAX-POOL and AVG-POOL, in the same function. **Exercise**: Implement the forward pass of the pooling layer. Follow the hints in the comments below.**Reminder**:As there's no padding, the formulas binding the output shape of the pooling to the input shape is:$$ n_H = \lfloor \frac{n_{H_{prev}} - f}{stride} \rfloor +1 $$$$ n_W = \lfloor \frac{n_{W_{prev}} - f}{stride} \rfloor +1 $$$$ n_C = n_{C_{prev}}$$ ###Code # GRADED FUNCTION: pool_forward def pool_forward(A_prev, hparameters, mode = "max"): """ Implements the forward pass of the pooling layer Arguments: A_prev -- Input data, numpy array of shape (m, n_H_prev, n_W_prev, n_C_prev) hparameters -- python dictionary containing "f" and "stride" mode -- the pooling mode you would like to use, defined as a string ("max" or "average") Returns: A -- output of the pool layer, a numpy array of shape (m, n_H, n_W, n_C) cache -- cache used in the backward pass of the pooling layer, contains the input and hparameters """ # Retrieve dimensions from the input shape (m, n_H_prev, n_W_prev, n_C_prev) = A_prev.shape # Retrieve hyperparameters from "hparameters" f = hparameters["f"] stride = hparameters["stride"] # Define the dimensions of the output n_H = int(1 + (n_H_prev - f) / stride) n_W = int(1 + (n_W_prev - f) / stride) n_C = n_C_prev # Initialize output matrix A A = np.zeros((m, n_H, n_W, n_C)) ### START CODE HERE ### for i in range(m): # loop over the training examples for h in range(n_H): # loop on the vertical axis of the output volume # Find the vertical start and end of the current "slice" (≈2 lines) vert_start = h * stride vert_end = vert_start + f for w in range(n_W): # loop on the horizontal axis of the output volume # Find the vertical start and end of the current "slice" (≈2 lines) horiz_start = w * stride horiz_end = horiz_start + f for c in range (n_C): # loop over the channels of the output volume # Use the corners to define the current slice on the ith training example of A_prev, channel c. (≈1 line) a_prev_slice = A_prev[i, vert_start:vert_end, horiz_start:horiz_end , c] # Compute the pooling operation on the slice. # Use an if statement to differentiate the modes. # Use np.max and np.mean. if mode == "max": A[i, h, w, c] = np.max(a_prev_slice) elif mode == "average": A[i, h, w, c] = np.mean(a_prev_slice) ### END CODE HERE ### # Store the input and hparameters in "cache" for pool_backward() cache = (A_prev, hparameters) # Making sure your output shape is correct assert(A.shape == (m, n_H, n_W, n_C)) return A, cache # Case 1: stride of 1 np.random.seed(1) A_prev = np.random.randn(2, 5, 5, 3) hparameters = {"stride" : 1, "f": 3} A, cache = pool_forward(A_prev, hparameters) print("mode = max") print("A.shape = " + str(A.shape)) print("A =\n", A) print() A, cache = pool_forward(A_prev, hparameters, mode = "average") print("mode = average") print("A.shape = " + str(A.shape)) print("A =\n", A) ###Output mode = max A.shape = (2, 3, 3, 3) A = [[[[1.74481176 0.90159072 1.65980218] [1.74481176 1.46210794 1.65980218] [1.74481176 1.6924546 1.65980218]] [[1.14472371 0.90159072 2.10025514] [1.14472371 0.90159072 1.65980218] [1.14472371 1.6924546 1.65980218]] [[1.13162939 1.51981682 2.18557541] [1.13162939 1.51981682 2.18557541] [1.13162939 1.6924546 2.18557541]]] [[[1.19891788 0.84616065 0.82797464] [0.69803203 0.84616065 1.2245077 ] [0.69803203 1.12141771 1.2245077 ]] [[1.96710175 0.84616065 1.27375593] [1.96710175 0.84616065 1.23616403] [1.62765075 1.12141771 1.2245077 ]] [[1.96710175 0.86888616 1.27375593] [1.96710175 0.86888616 1.23616403] [1.62765075 1.12141771 0.79280687]]]] mode = average A.shape = (2, 3, 3, 3) A = [[[[-3.01046719e-02 -3.24021315e-03 -3.36298859e-01] [ 1.43310483e-01 1.93146751e-01 -4.44905196e-01] [ 1.28934436e-01 2.22428468e-01 1.25067597e-01]] [[-3.81801899e-01 1.59993515e-02 1.70562706e-01] [ 4.73707165e-02 2.59244658e-02 9.20338402e-02] [ 3.97048605e-02 1.57189094e-01 3.45302489e-01]] [[-3.82680519e-01 2.32579951e-01 6.25997903e-01] [-2.47157416e-01 -3.48524998e-04 3.50539717e-01] [-9.52551510e-02 2.68511000e-01 4.66056368e-01]]] [[[-1.73134159e-01 3.23771981e-01 -3.43175716e-01] [ 3.80634669e-02 7.26706274e-02 -2.30268958e-01] [ 2.03009393e-02 1.41414785e-01 -1.23158476e-02]] [[ 4.44976963e-01 -2.61694592e-03 -3.10403073e-01] [ 5.08114737e-01 -2.34937338e-01 -2.39611830e-01] [ 1.18726772e-01 1.72552294e-01 -2.21121966e-01]] [[ 4.29449255e-01 8.44699612e-02 -2.72909051e-01] [ 6.76351685e-01 -1.20138225e-01 -2.44076712e-01] [ 1.50774518e-01 2.89111751e-01 1.23238536e-03]]]] ###Markdown ** Expected Output**```mode = maxA.shape = (2, 3, 3, 3)A = [[[[ 1.74481176 0.90159072 1.65980218] [ 1.74481176 1.46210794 1.65980218] [ 1.74481176 1.6924546 1.65980218]] [[ 1.14472371 0.90159072 2.10025514] [ 1.14472371 0.90159072 1.65980218] [ 1.14472371 1.6924546 1.65980218]] [[ 1.13162939 1.51981682 2.18557541] [ 1.13162939 1.51981682 2.18557541] [ 1.13162939 1.6924546 2.18557541]]] [[[ 1.19891788 0.84616065 0.82797464] [ 0.69803203 0.84616065 1.2245077 ] [ 0.69803203 1.12141771 1.2245077 ]] [[ 1.96710175 0.84616065 1.27375593] [ 1.96710175 0.84616065 1.23616403] [ 1.62765075 1.12141771 1.2245077 ]] [[ 1.96710175 0.86888616 1.27375593] [ 1.96710175 0.86888616 1.23616403] [ 1.62765075 1.12141771 0.79280687]]]]mode = averageA.shape = (2, 3, 3, 3)A = [[[[ -3.01046719e-02 -3.24021315e-03 -3.36298859e-01] [ 1.43310483e-01 1.93146751e-01 -4.44905196e-01] [ 1.28934436e-01 2.22428468e-01 1.25067597e-01]] [[ -3.81801899e-01 1.59993515e-02 1.70562706e-01] [ 4.73707165e-02 2.59244658e-02 9.20338402e-02] [ 3.97048605e-02 1.57189094e-01 3.45302489e-01]] [[ -3.82680519e-01 2.32579951e-01 6.25997903e-01] [ -2.47157416e-01 -3.48524998e-04 3.50539717e-01] [ -9.52551510e-02 2.68511000e-01 4.66056368e-01]]] [[[ -1.73134159e-01 3.23771981e-01 -3.43175716e-01] [ 3.80634669e-02 7.26706274e-02 -2.30268958e-01] [ 2.03009393e-02 1.41414785e-01 -1.23158476e-02]] [[ 4.44976963e-01 -2.61694592e-03 -3.10403073e-01] [ 5.08114737e-01 -2.34937338e-01 -2.39611830e-01] [ 1.18726772e-01 1.72552294e-01 -2.21121966e-01]] [[ 4.29449255e-01 8.44699612e-02 -2.72909051e-01] [ 6.76351685e-01 -1.20138225e-01 -2.44076712e-01] [ 1.50774518e-01 2.89111751e-01 1.23238536e-03]]]]``` ###Code # Case 2: stride of 2 np.random.seed(1) A_prev = np.random.randn(2, 5, 5, 3) hparameters = {"stride" : 2, "f": 3} A, cache = pool_forward(A_prev, hparameters) print("mode = max") print("A.shape = " + str(A.shape)) print("A =\n", A) print() A, cache = pool_forward(A_prev, hparameters, mode = "average") print("mode = average") print("A.shape = " + str(A.shape)) print("A =\n", A) ###Output mode = max A.shape = (2, 2, 2, 3) A = [[[[1.74481176 0.90159072 1.65980218] [1.74481176 1.6924546 1.65980218]] [[1.13162939 1.51981682 2.18557541] [1.13162939 1.6924546 2.18557541]]] [[[1.19891788 0.84616065 0.82797464] [0.69803203 1.12141771 1.2245077 ]] [[1.96710175 0.86888616 1.27375593] [1.62765075 1.12141771 0.79280687]]]] mode = average A.shape = (2, 2, 2, 3) A = [[[[-0.03010467 -0.00324021 -0.33629886] [ 0.12893444 0.22242847 0.1250676 ]] [[-0.38268052 0.23257995 0.6259979 ] [-0.09525515 0.268511 0.46605637]]] [[[-0.17313416 0.32377198 -0.34317572] [ 0.02030094 0.14141479 -0.01231585]] [[ 0.42944926 0.08446996 -0.27290905] [ 0.15077452 0.28911175 0.00123239]]]] ###Markdown **Expected Output:** ```mode = maxA.shape = (2, 2, 2, 3)A = [[[[ 1.74481176 0.90159072 1.65980218] [ 1.74481176 1.6924546 1.65980218]] [[ 1.13162939 1.51981682 2.18557541] [ 1.13162939 1.6924546 2.18557541]]] [[[ 1.19891788 0.84616065 0.82797464] [ 0.69803203 1.12141771 1.2245077 ]] [[ 1.96710175 0.86888616 1.27375593] [ 1.62765075 1.12141771 0.79280687]]]]mode = averageA.shape = (2, 2, 2, 3)A = [[[[-0.03010467 -0.00324021 -0.33629886] [ 0.12893444 0.22242847 0.1250676 ]] [[-0.38268052 0.23257995 0.6259979 ] [-0.09525515 0.268511 0.46605637]]] [[[-0.17313416 0.32377198 -0.34317572] [ 0.02030094 0.14141479 -0.01231585]] [[ 0.42944926 0.08446996 -0.27290905] [ 0.15077452 0.28911175 0.00123239]]]]``` Congratulations! You have now implemented the forward passes of all the layers of a convolutional network. The remainder of this notebook is optional, and will not be graded. 5 - Backpropagation in convolutional neural networks (OPTIONAL / UNGRADED)In modern deep learning frameworks, you only have to implement the forward pass, and the framework takes care of the backward pass, so most deep learning engineers don't need to bother with the details of the backward pass. The backward pass for convolutional networks is complicated. If you wish, you can work through this optional portion of the notebook to get a sense of what backprop in a convolutional network looks like. When in an earlier course you implemented a simple (fully connected) neural network, you used backpropagation to compute the derivatives with respect to the cost to update the parameters. Similarly, in convolutional neural networks you can calculate the derivatives with respect to the cost in order to update the parameters. The backprop equations are not trivial and we did not derive them in lecture, but we will briefly present them below. 5.1 - Convolutional layer backward pass Let's start by implementing the backward pass for a CONV layer. 5.1.1 - Computing dA:This is the formula for computing $dA$ with respect to the cost for a certain filter $W_c$ and a given training example:$$ dA += \sum _{h=0} ^{n_H} \sum_{w=0} ^{n_W} W_c \times dZ_{hw} \tag{1}$$Where $W_c$ is a filter and $dZ_{hw}$ is a scalar corresponding to the gradient of the cost with respect to the output of the conv layer Z at the hth row and wth column (corresponding to the dot product taken at the ith stride left and jth stride down). Note that at each time, we multiply the the same filter $W_c$ by a different dZ when updating dA. We do so mainly because when computing the forward propagation, each filter is dotted and summed by a different a_slice. Therefore when computing the backprop for dA, we are just adding the gradients of all the a_slices. In code, inside the appropriate for-loops, this formula translates into:```pythonda_prev_pad[vert_start:vert_end, horiz_start:horiz_end, :] += W[:,:,:,c] * dZ[i, h, w, c]``` 5.1.2 - Computing dW:This is the formula for computing $dW_c$ ($dW_c$ is the derivative of one filter) with respect to the loss:$$ dW_c += \sum _{h=0} ^{n_H} \sum_{w=0} ^ {n_W} a_{slice} \times dZ_{hw} \tag{2}$$Where $a_{slice}$ corresponds to the slice which was used to generate the activation $Z_{ij}$. Hence, this ends up giving us the gradient for $W$ with respect to that slice. Since it is the same $W$, we will just add up all such gradients to get $dW$. In code, inside the appropriate for-loops, this formula translates into:```pythondW[:,:,:,c] += a_slice * dZ[i, h, w, c]``` 5.1.3 - Computing db:This is the formula for computing $db$ with respect to the cost for a certain filter $W_c$:$$ db = \sum_h \sum_w dZ_{hw} \tag{3}$$As you have previously seen in basic neural networks, db is computed by summing $dZ$. In this case, you are just summing over all the gradients of the conv output (Z) with respect to the cost. In code, inside the appropriate for-loops, this formula translates into:```pythondb[:,:,:,c] += dZ[i, h, w, c]```**Exercise**: Implement the `conv_backward` function below. You should sum over all the training examples, filters, heights, and widths. You should then compute the derivatives using formulas 1, 2 and 3 above. ###Code def conv_backward(dZ, cache): """ Implement the backward propagation for a convolution function Arguments: dZ -- gradient of the cost with respect to the output of the conv layer (Z), numpy array of shape (m, n_H, n_W, n_C) cache -- cache of values needed for the conv_backward(), output of conv_forward() Returns: dA_prev -- gradient of the cost with respect to the input of the conv layer (A_prev), numpy array of shape (m, n_H_prev, n_W_prev, n_C_prev) dW -- gradient of the cost with respect to the weights of the conv layer (W) numpy array of shape (f, f, n_C_prev, n_C) db -- gradient of the cost with respect to the biases of the conv layer (b) numpy array of shape (1, 1, 1, n_C) """ ### START CODE HERE ### # Retrieve information from "cache" (A_prev, W, b, hparameters) = cache # Retrieve dimensions from A_prev's shape (m, n_H_prev, n_W_prev, n_C_prev) = A_prev.shape # Retrieve dimensions from W's shape (f, f, n_C_prev, n_C) = W.shape # Retrieve information from "hparameters" stride = hparameters['stride'] pad = hparameters['pad'] # Retrieve dimensions from dZ's shape (m, n_H, n_W, n_C) = dZ.shape # Initialize dA_prev, dW, db with the correct shapes dA_prev = np.zeros((m, n_H_prev, n_W_prev, n_C_prev)) dW = np.zeros((f, f, n_C_prev, n_C)) db = np.zeros((1, 1, 1, n_C)) # Pad A_prev and dA_prev A_prev_pad = zero_pad(A_prev, pad) dA_prev_pad = zero_pad(dA_prev, pad) for i in range(m): # loop over the training examples # select ith training example from A_prev_pad and dA_prev_pad a_prev_pad = A_prev_pad[i] da_prev_pad = dA_prev_pad[i] for h in range(n_H): # loop over vertical axis of the output volume for w in range(n_W): # loop over horizontal axis of the output volume for c in range(n_C): # loop over the channels of the output volume # Find the corners of the current "slice" vert_start = h * stride vert_end = vert_start + f horiz_start = w * stride horiz_end = horiz_start + f # Use the corners to define the slice from a_prev_pad a_slice = a_prev_pad[vert_start:vert_end, horiz_start:horiz_end, :] # Update gradients for the window and the filter's parameters using the code formulas given above da_prev_pad[vert_start:vert_end, horiz_start:horiz_end, :] += W[:,:,:,c] * dZ[i, h, w, c] dW[:,:,:,c] += a_slice * dZ[i, h, w, c] db[:,:,:,c] += dZ[i, h, w, c] # Set the ith training example's dA_prev to the unpadded da_prev_pad (Hint: use X[pad:-pad, pad:-pad, :]) dA_prev[i, :, :, :] = da_prev_pad[pad:-pad, pad:-pad, :] ### END CODE HERE ### # Making sure your output shape is correct assert(dA_prev.shape == (m, n_H_prev, n_W_prev, n_C_prev)) return dA_prev, dW, db # We'll run conv_forward to initialize the 'Z' and 'cache_conv", # which we'll use to test the conv_backward function np.random.seed(1) A_prev = np.random.randn(10,4,4,3) W = np.random.randn(2,2,3,8) b = np.random.randn(1,1,1,8) hparameters = {"pad" : 2, "stride": 2} Z, cache_conv = conv_forward(A_prev, W, b, hparameters) # Test conv_backward dA, dW, db = conv_backward(Z, cache_conv) print("dA_mean =", np.mean(dA)) print("dW_mean =", np.mean(dW)) print("db_mean =", np.mean(db)) ###Output dA_mean = 1.4524377775388075 dW_mean = 1.7269914583139097 db_mean = 7.839232564616838 ###Markdown ** Expected Output: ** **dA_mean** 1.45243777754 **dW_mean** 1.72699145831 **db_mean** 7.83923256462 5.2 Pooling layer - backward passNext, let's implement the backward pass for the pooling layer, starting with the MAX-POOL layer. Even though a pooling layer has no parameters for backprop to update, you still need to backpropagation the gradient through the pooling layer in order to compute gradients for layers that came before the pooling layer. 5.2.1 Max pooling - backward pass Before jumping into the backpropagation of the pooling layer, you are going to build a helper function called `create_mask_from_window()` which does the following: $$ X = \begin{bmatrix}1 && 3 \\4 && 2\end{bmatrix} \quad \rightarrow \quad M =\begin{bmatrix}0 && 0 \\1 && 0\end{bmatrix}\tag{4}$$As you can see, this function creates a "mask" matrix which keeps track of where the maximum of the matrix is. True (1) indicates the position of the maximum in X, the other entries are False (0). You'll see later that the backward pass for average pooling will be similar to this but using a different mask. **Exercise**: Implement `create_mask_from_window()`. This function will be helpful for pooling backward. Hints:- [np.max()]() may be helpful. It computes the maximum of an array.- If you have a matrix X and a scalar x: `A = (X == x)` will return a matrix A of the same size as X such that:```A[i,j] = True if X[i,j] = xA[i,j] = False if X[i,j] != x```- Here, you don't need to consider cases where there are several maxima in a matrix. ###Code def create_mask_from_window(x): """ Creates a mask from an input matrix x, to identify the max entry of x. Arguments: x -- Array of shape (f, f) Returns: mask -- Array of the same shape as window, contains a True at the position corresponding to the max entry of x. """ ### START CODE HERE ### (≈1 line) mask = (x == np.max(x)) ### END CODE HERE ### return mask np.random.seed(1) x = np.random.randn(2,3) mask = create_mask_from_window(x) print('x = ', x) print("mask = ", mask) ###Output x = [[ 1.62434536 -0.61175641 -0.52817175] [-1.07296862 0.86540763 -2.3015387 ]] mask = [[ True False False] [False False False]] ###Markdown **Expected Output:** **x =**[[ 1.62434536 -0.61175641 -0.52817175] [-1.07296862 0.86540763 -2.3015387 ]] **mask =**[[ True False False] [False False False]] Why do we keep track of the position of the max? It's because this is the input value that ultimately influenced the output, and therefore the cost. Backprop is computing gradients with respect to the cost, so anything that influences the ultimate cost should have a non-zero gradient. So, backprop will "propagate" the gradient back to this particular input value that had influenced the cost. 5.2.2 - Average pooling - backward pass In max pooling, for each input window, all the "influence" on the output came from a single input value--the max. In average pooling, every element of the input window has equal influence on the output. So to implement backprop, you will now implement a helper function that reflects this.For example if we did average pooling in the forward pass using a 2x2 filter, then the mask you'll use for the backward pass will look like: $$ dZ = 1 \quad \rightarrow \quad dZ =\begin{bmatrix}1/4 && 1/4 \\1/4 && 1/4\end{bmatrix}\tag{5}$$This implies that each position in the $dZ$ matrix contributes equally to output because in the forward pass, we took an average. **Exercise**: Implement the function below to equally distribute a value dz through a matrix of dimension shape. [Hint](https://docs.scipy.org/doc/numpy-1.13.0/reference/generated/numpy.ones.html) ###Code def distribute_value(dz, shape): """ Distributes the input value in the matrix of dimension shape Arguments: dz -- input scalar shape -- the shape (n_H, n_W) of the output matrix for which we want to distribute the value of dz Returns: a -- Array of size (n_H, n_W) for which we distributed the value of dz """ ### START CODE HERE ### # Retrieve dimensions from shape (≈1 line) (n_H, n_W) = shape # Compute the value to distribute on the matrix (≈1 line) average = np.ones((n_H, n_W)) # Create a matrix where every entry is the "average" value (≈1 line) a = average/dz ### END CODE HERE ### return a a = distribute_value(2, (2,2)) print('distributed value =', a) ###Output distributed value = [[0.5 0.5] [0.5 0.5]] ###Markdown **Expected Output**: distributed_value =[[ 0.5 0.5] [ 0.5 0.5]] 5.2.3 Putting it together: Pooling backward You now have everything you need to compute backward propagation on a pooling layer.**Exercise**: Implement the `pool_backward` function in both modes (`"max"` and `"average"`). You will once again use 4 for-loops (iterating over training examples, height, width, and channels). You should use an `if/elif` statement to see if the mode is equal to `'max'` or `'average'`. If it is equal to 'average' you should use the `distribute_value()` function you implemented above to create a matrix of the same shape as `a_slice`. Otherwise, the mode is equal to '`max`', and you will create a mask with `create_mask_from_window()` and multiply it by the corresponding value of dA. ###Code def pool_backward(dA, cache, mode = "max"): """ Implements the backward pass of the pooling layer Arguments: dA -- gradient of cost with respect to the output of the pooling layer, same shape as A cache -- cache output from the forward pass of the pooling layer, contains the layer's input and hparameters mode -- the pooling mode you would like to use, defined as a string ("max" or "average") Returns: dA_prev -- gradient of cost with respect to the input of the pooling layer, same shape as A_prev """ ### START CODE HERE ### # Retrieve information from cache (≈1 line) (A_prev, hparameters) = cache # Retrieve hyperparameters from "hparameters" (≈2 lines) stride = hparameters['stride'] f = hparameters['f'] # Retrieve dimensions from A_prev's shape and dA's shape (≈2 lines) m, n_H_prev, n_W_prev, n_C_prev = A_prev.shape m, n_H, n_W, n_C = dA.shape # Initialize dA_prev with zeros (≈1 line) dA_prev = np.zeros(A_prev.shape) for i in range(m): # loop over the training examples # select training example from A_prev (≈1 line) a_prev = A_prev[i] for h in range(n_H): # loop on the vertical axis for w in range(n_W): # loop on the horizontal axis for c in range(n_C): # loop over the channels (depth) # Find the corners of the current "slice" (≈4 lines) vert_start = h * stride vert_end = vert_start + f horiz_start = w * stride horiz_end = horiz_start + f # Compute the backward propagation in both modes. if mode == "max": # Use the corners and "c" to define the current slice from a_prev (≈1 line) a_prev_slice = a_prev[vert_start:vert_end, horiz_start:horiz_end, c] # Create the mask from a_prev_slice (≈1 line) mask = create_mask_from_window(a_prev_slice) # Set dA_prev to be dA_prev + (the mask multiplied by the correct entry of dA) (≈1 line) dA_prev[i, vert_start: vert_end, horiz_start: horiz_end, c] += np.multiply(dA[i, h, w, c], mask) elif mode == "average": # Get the value a from dA (≈1 line) da = dA[i, h, w, c] # Define the shape of the filter as fxf (≈1 line) shape = (f, f) # Distribute it to get the correct slice of dA_prev. i.e. Add the distributed value of da. (≈1 line) dA_prev[i, vert_start: vert_end, horiz_start: horiz_end, c] += distribute_value(da, shape) ### END CODE ### # Making sure your output shape is correct assert(dA_prev.shape == A_prev.shape) return dA_prev np.random.seed(1) A_prev = np.random.randn(5, 5, 3, 2) hparameters = {"stride" : 1, "f": 2} A, cache = pool_forward(A_prev, hparameters) dA = np.random.randn(5, 4, 2, 2) dA_prev = pool_backward(dA, cache, mode = "max") print("mode = max") print('mean of dA = ', np.mean(dA)) print('dA_prev[1,1] = ', dA_prev[1,1]) print() dA_prev = pool_backward(dA, cache, mode = "average") print("mode = average") print('mean of dA = ', np.mean(dA)) print('dA_prev[1,1] = ', dA_prev[1,1]) ###Output mode = max mean of dA = 0.14571390272918056 dA_prev[1,1] = [[ 0. 0. ] [ 5.05844394 -1.68282702] [ 0. 0. ]] mode = average mean of dA = 0.14571390272918056 dA_prev[1,1] = [[-0.53193997 5.7923756 ] [ 0.32005384 1.24308632] [ 0.85199382 -4.54928928]]
notebooks/text_analytics.ipynb
###Markdown PUBLICAÇÕES DE ATOS JUDICIAIS 1. Entendimento do projeto**Objetivo:** coletar, tratar, classificar e analisar publicações de atos judiciais.**Dados:** Publicações de atos judiciais obtidas da plataforma do Diário da Justiça Eletrônico Nacional (DJEN), mantida pelo Conselho Nacional de Justiça (CNJ).**Unidades judiciais:** Varas Cíveis do Termo Judiciário de São Luis/MA (1ª a 16ª).**Período:** 01/01/2021 a 09/08/2021 2. Bibliotecas e funções ###Code import pandas as pd import numpy as np import matplotlib.pyplot as plt import seaborn as sns from wordcloud import WordCloud import warnings warnings.filterwarnings ('ignore') import matplotlib.image as mpimg import nltk from nltk.corpus import stopwords from nltk.tokenize import word_tokenize from nltk.stem.snowball import SnowballStemmer from nltk.stem.wordnet import WordNetLemmatizer nltk.download('stopwords') nltk.download('punkt') nltk.download('wordnet') from sklearn.cluster import MiniBatchKMeans from sklearn.metrics.pairwise import cosine_similarity from sklearn.feature_extraction.text import TfidfVectorizer, CountVectorizer from sklearn.model_selection import train_test_split from sklearn.feature_extraction import _stop_words from sklearn.linear_model import LogisticRegression from sklearn.neighbors import KNeighborsClassifier from sklearn.tree import DecisionTreeClassifier from sklearn.ensemble import RandomForestClassifier from sklearn.metrics import accuracy_score, confusion_matrix, classification_report, precision_score, recall_score import xgboost as xgb # Função de tratamento dos textos def trata_texto(text): text = text.replace(regex=r'[!/,.-]', value='').apply(lambda x: x.lower()) stop_words = stopwords.words('portuguese') text = text.apply(lambda x: ' '.join([word for word in x.split() if word not in (stop_words)])) text = text.map(lambda x: word_tokenize(x)) snowball = SnowballStemmer(language = 'portuguese') text = text.map(lambda x: [snowball.stem(y) for y in x]) text = text.apply(lambda x: ' '.join(x)) return text ###Output _____no_output_____ ###Markdown 3. Coleta dos dadosOs dados abaixo foram obtidos por meio do endereço da API do Conselho Nacional de Justiça (https://comunicaapi.pje.jus.br/), com uso das bibliotecas 'requests' e 'json' em python. ###Code df = pd.read_csv('../data/dataset.csv') df ###Output _____no_output_____ ###Markdown 4. Tratamento dos textos 4.1. PreprocessamentoChamamento da função criada para tratamento dos textos.A função inclui as seguintes etapas: (1) remoção de caracteres especiais (2) conversão em letras minúsculas (3) remoção de stop words (4) tokenização (5) stemização ###Code df['texto_processado'] = trata_texto(df['texto']) ###Output _____no_output_____ ###Markdown 4.2. Vetorização ###Code tfidf = TfidfVectorizer(max_df=0.90, min_df=50, max_features=1000) vectorized = tfidf.fit_transform(df['texto_processado']) vectorized ###Output _____no_output_____ ###Markdown 5. Classificação dos textosAs publicações judicias não estão classificadas quanto ao tipo do ato judicial. Deste modo, realizou-se a rotulação dos textos dentre as seguintes categorias: [ATO ORDINATÓRIO] [DESPACHO/DECISÃO] [SENTENÇA] [EDITAL] Para tanto seguiu-se as seguintes etapas: 1º) Clusterização dos textos para identificação das melhores amostras para rotulação manual 2º) Classificação de todo coleção de textos a partir das amostras rotuladas 5.1. Clusterização - K-means ###Code # Método do cotovelo para encontrar número ideeal de clusters iters = [2, 20, 40, 60, 80, 100, 125, 150, 175, 200, 250, 300, 350, 400, 500, 600, 800, 1000] sse = [] models = [] for k in iters: model = MiniBatchKMeans(n_clusters=k, init_size=256, batch_size=512, random_state=42).fit(vectorized) sse.append(model.inertia_) plt.plot(iters, sse) plt.savefig(f'../img/elbow_method.png') plt.show() ###Output _____no_output_____ ###Markdown Utilizando o método do cotovelo não se identificou o número exato de clusters para o modelo, mas percebe-se pela representação gráfica que a partir de 100 clusters a variação da inércia (soma das distâncias quadradas das amostras ao centro do cluster) reduz significativamente.Assim, promove-se a predição com o número de clusters definido em 100 e, após, classifica-se cada agrupamento em uma das categorias desejados. ###Code # Predição do modelo com número de clusters k = 100 k = 100 model = MiniBatchKMeans(n_clusters=k, init_size=256, batch_size=512, random_state=42).fit(vectorized) df['cluster'] = model.predict(vectorized) df['cluster'].hist(bins=100) ###Output _____no_output_____ ###Markdown 5.2. Identificando melhores amostras para rotulação manual - similaridade dos cossenos ###Code # Melhores amostras de cada cluster por similaridade de cossenos topn_indices = [] for i in range(0, 100): similarities = [] centroid = model.cluster_centers_[i] for v in vectorized: similarities.append(cosine_similarity([centroid], v)) indexes = np.array([s[0][0] for s in similarities]) indexes = np.argsort(indexes)[::-1] topn_indices.append(indexes) # Geração de nuvem de palavras sobre as 50 melhores amostras de cada cluster topn = 50 stopwordcloud = ['var', 'cível', 'estad', 'maranhã', 'juíz', 'juiz', 'direit', 'term', 'luís', 'process', 'açã', 'secret', 'judicial', 'comarc', 'únic', 'digital', 'únic digital', 'unic', 'únic digital'] wordcloud = WordCloud(width=800, height=400, background_color='white', max_words=50, stopwords=stopwordcloud, random_state=42) for i in range(0, 100): text_wordcloud = ' '.join(df['texto_processado'].iloc[topn_indices[i][0:topn]]) nuvem = wordcloud.generate(text_wordcloud) plt.figure(figsize=(10,10)) plt.axis("off") plt.title(f'CLUSTER nº {i}') #plt.imshow(nuvem) #plt.savefig(f'../img/wordcloud_amostras_cluster_{i}.png') # Exemplo de uma das nuvens de palavras geradas plt.figure(figsize=(10,10)) plt.axis("off") plt.imshow(nuvem) # Exportando indices das melhores amostras np.savetxt('../data/topn_indices.csv', topn_indices, delimiter =",") ###Output _____no_output_____ ###Markdown 5.3. Rotulação manual ###Code # Rotulação manual das melhores amostras de cada cluster ato_ord = [1,2,5,6,8,10,15,17,19,21,23,25,28,34,44,45,48,50,54,55,66,78,86,95,96] desp_dec = [0,11,12,14,16,18,20,22,24,26,29,30,31,32,35,36,43,46,49,51,52,53,56,58,59,60,61,62,64,65,68,69,70,71,73,74,75,77,79,82,83,84,85,88,94,97,99] sentenca = [3,4,9,27,33,37,39,40,41,42,47,63,67,72,76,80,87,91,92,98] edital = [7] # Geração de dataset com amostras rotuladas indices_amostras = [] for i in range(0, 100): for j in range(0, 50): indices_amostras.append(topn_indices[i][j]) df_amostras = df.loc[indices_amostras] df_amostras['categoria'] = df_amostras['cluster'].apply(lambda x: 'ato_ord' if x in ato_ord else 'desp_dec' if x in desp_dec else 'sentenca' if x in sentenca else 'edital' if x in edital else 'indefinido') df_amostras = df_amostras[['texto', 'categoria']] df_amostras = df_amostras[df_amostras['categoria']!='indefinido'] df_amostras # Frequência das categorias sns.countplot(df_amostras['categoria']) # Exportando dataset df_amostras.to_csv('../data/dataset_amostras_rotulado.csv') ###Output _____no_output_____ ###Markdown 5.4. Treinamento e avaliação dos algoritmos de classificação ###Code # Dados de treino X_train = df_amostras['texto'] y_train = df_amostras['categoria'] # Dados de teste - classificados manualmente pelo Stakeholder df_teste = pd.read_csv('../data/dataset_validacao.csv') X_test = df_teste['texto'] y_test = df_teste['categoria'] # Preprocessamento X_train = trata_texto(X_train) X_train = tfidf.transform(X_train) X_test = trata_texto(X_test) X_test = tfidf.transform(X_test) X_train X_test df_teste['categoria'].value_counts() ###Output _____no_output_____ ###Markdown Regressão logística ###Code # Treino e Predição lr = LogisticRegression() lr.fit(X_train,y_train) y_pred = lr.predict(X_test) # Avaliação print(classification_report(y_test, y_pred)) cfm = confusion_matrix(y_test, y_pred) sns.heatmap(cfm, annot=True, fmt='.0f').set_title('Logistic_Regression') ###Output precision recall f1-score support ato_ord 1.00 0.93 0.97 15 desp_dec 0.86 1.00 0.92 30 edital 1.00 0.20 0.33 5 sentenca 1.00 1.00 1.00 9 accuracy 0.92 59 macro avg 0.96 0.78 0.81 59 weighted avg 0.93 0.92 0.90 59 ###Markdown Árvore de decisão ###Code # Treino e Predição dtree = DecisionTreeClassifier() dtree.fit(X_train,y_train) y_pred = dtree.predict(X_test) # Avaliação print(classification_report(y_test, y_pred)) cfm = confusion_matrix(y_test, y_pred) sns.heatmap(cfm, annot=True, fmt='.0f').set_title('Decision_Tree') ###Output precision recall f1-score support ato_ord 1.00 1.00 1.00 15 desp_dec 0.88 0.93 0.90 30 edital 1.00 0.60 0.75 5 sentenca 0.78 0.78 0.78 9 accuracy 0.90 59 macro avg 0.91 0.83 0.86 59 weighted avg 0.90 0.90 0.90 59 ###Markdown Floresta aleatória ###Code # Treino e Predição rf = RandomForestClassifier() rf.fit(X_train, y_train) y_pred = rf.predict(X_test) # Avaliação print(classification_report(y_test, y_pred)) cfm = confusion_matrix(y_test, y_pred) sns.heatmap(cfm, annot=True, fmt='.0f').set_title('Random_Forest') ###Output precision recall f1-score support ato_ord 1.00 1.00 1.00 15 desp_dec 0.88 1.00 0.94 30 edital 1.00 0.20 0.33 5 sentenca 1.00 1.00 1.00 9 accuracy 0.93 59 macro avg 0.97 0.80 0.82 59 weighted avg 0.94 0.93 0.91 59 ###Markdown Impulso do gradiante extremo ###Code # Treino e Predição xgbmodel = xgb.XGBClassifier(random_state=0) xgbmodel.fit(X_train,y_train) y_pred=xgbmodel.predict(X_test) # Avaliação print(classification_report(y_test,y_pred)) cfm = confusion_matrix(y_test, y_pred) sns.heatmap(cfm, annot=True, fmt='.0f').set_title('XGBoost') ###Output [09:51:20] WARNING: C:/Users/Administrator/workspace/xgboost-win64_release_1.3.0/src/learner.cc:1061: Starting in XGBoost 1.3.0, the default evaluation metric used with the objective 'multi:softprob' was changed from 'merror' to 'mlogloss'. Explicitly set eval_metric if you'd like to restore the old behavior. precision recall f1-score support ato_ord 1.00 1.00 1.00 15 desp_dec 0.88 1.00 0.94 30 edital 1.00 0.20 0.33 5 sentenca 1.00 1.00 1.00 9 accuracy 0.93 59 macro avg 0.97 0.80 0.82 59 weighted avg 0.94 0.93 0.91 59 ###Markdown **Conclusão sobre os algoritmos classificadores:*** Considerando a acurácia e precisão dos classificadores treinados, observa-se que o algoritmo FLORESTA ALEATÓRIA (Random Forest) apresentou melhor performance. 5.5. Classificação final do conjunto de dados ###Code # Conjunto de dados para classificação final df['categoria'] = rf.predict(vectorized) df # Exportando dataset final para análise no Power Bi df.to_csv('../data/dataset_final.csv') ###Output _____no_output_____ ###Markdown 6. Análise dos dados finais no Power BI ###Code img = mpimg.imread('../img/imagem_dashboard.png') plt.figure(figsize=(15,10)) plt.imshow(img) plt.show() ###Output _____no_output_____
L1000/0C.preprocessing/0.data-splitsONLYPHASE2.ipynb
###Markdown Split the L1000 Data into Training/Testing/Validation SetsSplit the data 80% training, 10% testing, 10% validation, balanced by platemap. ###Code import sys import pathlib import numpy as np import pandas as pd from sklearn.model_selection import train_test_split from pycytominer import feature_select from pycytominer.cyto_utils import infer_cp_features sys.path.insert(0, "../../scripts") from utils import transform, infer_L1000_features # %load_ext nb_black seed = 9876 test_split = 0.2 output_dir = pathlib.Path("data") output_dir.mkdir(exist_ok=True) # Load data phase2_L1000_df = pd.read_csv("../0B.process-data/data/L1000_phase2.tsv.gz", sep="\t") print(phase2_L1000_df.shape) phase2_L1000_df.head(2) features = infer_L1000_features(phase2_L1000_df) meta_features = infer_L1000_features(phase2_L1000_df, metadata=True) # Zero One Normalize Data phase2_L1000_df = transform( phase2_L1000_df, features=features, meta_features=meta_features, operation = "-1+1" ) # Split data into 80% train, 20% test train_df, test_df = train_test_split( phase2_L1000_df, test_size=test_split, random_state=seed, stratify=phase2_L1000_df.cell_id, ) # Split test data into 50% validation, 50% test test_df, valid_df = train_test_split( test_df, test_size=0.5, random_state=seed, stratify=test_df.cell_id, ) print(train_df.shape) print(test_df.shape) print(valid_df.shape) # Output data splits train_file = pathlib.Path(output_dir, "L1000PHASE2-1+1_train.tsv.gz") test_file = pathlib.Path(output_dir, "L1000PHASE2-1+1_test.tsv.gz") valid_file = pathlib.Path(output_dir, "L1000PHASE2-1+1_valid.tsv.gz") complete_file = pathlib.Path(output_dir, "L1000PHASE2-1+1_complete.tsv.gz") # train_df.to_csv(train_file, sep="\t", index=False, float_format="%.5g") # test_df.to_csv(test_file, sep="\t", index=False, float_format="%.5g") # valid_df.to_csv(valid_file, sep="\t", index=False, float_format="%.5g") phase2_L1000_df.to_csv(complete_file, sep="\t", index=False, float_format="%.5g") ###Output _____no_output_____
docs/source/examples/usage/NetcdfStream.ipynb
###Markdown Setup the Config ###Code from ioos_qc.config import Config config = """ streams: variable1: qartod: aggregate: gross_range_test: suspect_span: [20, 30] fail_span: [10, 40] """ c = Config(config) c.config ###Output _____no_output_____ ###Markdown Setup the sample data ###Code import os import numpy as np import xarray as xr import pandas as pd import netCDF4 as nc4 rows = 50 data_inputs = { 'time': pd.date_range(start='01/01/2020', periods=rows, freq='D'), 'z': 2.0, 'lat': 36.1, 'lon': -76.5, 'variable1': np.arange(0, rows), } df = pd.DataFrame(data_inputs) ncfile = 'tmp.nc' if os.path.exists(ncfile): os.remove(ncfile) ds = xr.Dataset.from_dataframe(df).to_netcdf(ncfile, 'w') ###Output _____no_output_____ ###Markdown Setup the NetcdfStream ###Code from ioos_qc.streams import NetcdfStream ns = NetcdfStream(ncfile) ns ###Output _____no_output_____ ###Markdown Run the NetcdfStream through the Config ###Code results = ns.run(c) results ###Output _____no_output_____
Coding_Interview_exercises/TestDome/03_merge_names.ipynb
###Markdown Instruction - To be completed in **10 min** - Implement the unique_names method. When passed two arrays of names, it will return an array containing the names that appear in either or both arrays.- The returned array should have no duplicates.- For example, `calling unique_names(['Ava', 'Emma', 'Olivia'], ['Olivia', 'Sophia', 'Emma'])` should return an array containing Ava, Emma, Olivia, and Sophia in any order. Start-up code ###Code def unique_names(names1, names2): return None names1 = ["Ava", "Emma", "Olivia"] names2 = ["Olivia", "Sophia", "Emma"] print(unique_names(names1, names2)) # should print Ava, Emma, Olivia, Sophia ###Output _____no_output_____ ###Markdown Dirty solution ###Code def unique_names(names1, names2): return sorted(list(set(names1+names2))) names1 = ["Ava", "Emma", "Olivia"] names2 = ["Olivia", "Sophia", "Emma"] print(unique_names(names1, names2)) # should print Ava, Emma, Olivia, Sophia ###Output ['Ava', 'Emma', 'Olivia', 'Sophia'] ###Markdown Elegant solution - `itertools ` is a module implements a number of iterator building blocks inspired by constructs from APL, Haskell, and SML. Each has been recast in a form suitable for Python.- The module standardizes a core set of fast, memory efficient tools that are useful by themselves or in combination. Together, they form an “iterator algebra” making it possible to construct specialized tools succinctly and efficiently in pure Python. ###Code from itertools import chain def unique_names(names1, names2): return list(set(chain(names1, names2))) names1 = ["Ava", "Emma", "Olivia"] names2 = ["Olivia", "Sophia", "Emma"] print(unique_names(names1, names2)) # should print Ava, Emma, Olivia, Sophia ###Output ['Emma', 'Ava', 'Olivia', 'Sophia']
jupyter/docker_enterprise/docker_notebooks/Spark_NLP/Healthcare/24.Improved_Entity_Resolvers_in_SparkNLP_with_sBert.ipynb
###Markdown ![JohnSnowLabs](https://nlp.johnsnowlabs.com/assets/images/logo.png) [![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/JohnSnowLabs/spark-nlp-workshop/blob/master/tutorials/Certification_Trainings/Healthcare/24.Improved_Entity_Resolvers_in_SparkNLP_with_sBert.ipynb) 24. Improved Entity Resolvers in Spark NLP with sBert ###Code import os jsl_secret = os.getenv('SECRET') import sparknlp sparknlp_version = sparknlp.version() import sparknlp_jsl jsl_version = sparknlp_jsl.version() print (jsl_secret) import json import os from pyspark.ml import Pipeline, PipelineModel from pyspark.sql import SparkSession from sparknlp.annotator import * from sparknlp_jsl.annotator import * from sparknlp.base import * from sparknlp.util import * import sparknlp_jsl import sparknlp from sparknlp.pretrained import ResourceDownloader from pyspark.sql import functions as F params = {"spark.driver.memory":"16G", "spark.kryoserializer.buffer.max":"2000M", "spark.driver.maxResultSize":"2000M"} spark = sparknlp_jsl.start(jsl_secret,params=params) print ("Spark NLP Version :", sparknlp.version()) print ("Spark NLP_JSL Version :", sparknlp_jsl.version()) spark ###Output Spark NLP Version : 3.2.1 Spark NLP_JSL Version : 3.2.0 ###Markdown !!! Warning !!!**If you get an error related to Java port not found 55, it is probably because that the Colab memory cannot handle the model and the Spark session died. In that case, try on a larger machine or restart the kernel at the top and then come back here and rerun.** ICD10CM pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") icd10_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_icd10cm_augmented","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("icd10cm_code")\ .setDistanceFunction("EUCLIDEAN") icd_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, icd10_resolver]) icd_lp = LightPipeline(icd_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_icd10cm_augmented download started this may take some time. Approximate size to download 1.2 GB [OK!] ###Markdown ICD10CM-HCC pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") hcc_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_icd10cm_augmented_billable_hcc","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("icd10cm_hcc_code")\ .setDistanceFunction("EUCLIDEAN") hcc_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, hcc_resolver]) hcc_lp = LightPipeline(hcc_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_icd10cm_augmented_billable_hcc download started this may take some time. Approximate size to download 1.4 GB [OK!] ###Markdown RxNorm pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") rxnorm_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_rxnorm","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("rxnorm_code")\ .setDistanceFunction("EUCLIDEAN") rxnorm_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, rxnorm_resolver]) rxnorm_lp = LightPipeline(rxnorm_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_rxnorm download started this may take some time. Approximate size to download 802.6 MB [OK!] ###Markdown CPT pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") cpt_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_cpt_procedures_augmented","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("cpt_code")\ .setDistanceFunction("EUCLIDEAN") cpt_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, cpt_resolver]) cpt_lp = LightPipeline(cpt_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_cpt_procedures_augmented download started this may take some time. Approximate size to download 78.3 MB [OK!] ###Markdown SNOMED pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") snomed_ct_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_snomed_findings_aux_concepts","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("snomed_code")\ .setDistanceFunction("EUCLIDEAN") snomed_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, snomed_ct_resolver]) snomed_lp = LightPipeline(snomed_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_snomed_findings_aux_concepts download started this may take some time. Approximate size to download 4.3 GB [OK!] ###Markdown LOINC Pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") loinc_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_loinc", "en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("loinc_code")\ .setDistanceFunction("EUCLIDEAN") loinc_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, loinc_resolver]) loinc_lp = LightPipeline(loinc_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_loinc download started this may take some time. Approximate size to download 212.6 MB [OK!] ###Markdown UMLS Pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") umls_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_umls_major_concepts", "en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("umls_code")\ .setDistanceFunction("EUCLIDEAN") umls_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, umls_resolver]) umls_lp = LightPipeline(umls_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_umls_major_concepts download started this may take some time. Approximate size to download 817.3 MB [OK!] ###Markdown HPO Pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") hpo_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_HPO", "en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("umls_code")\ .setDistanceFunction("EUCLIDEAN") hpo_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, hpo_resolver]) hpo_lp = LightPipeline(hpo_pipelineModel) ###Output sbiobert_base_cased_mli download started this may take some time. Approximate size to download 384.3 MB [OK!] sbiobertresolve_HPO download started this may take some time. Approximate size to download 97.9 MB [OK!] ###Markdown All the resolvers in the same pipeline (just to show how it is done.. will not be used in this notebook) ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("ner_chunk") sbert_embedder = BertSentenceEmbeddings\ .pretrained('sbiobert_base_cased_mli', 'en','clinical/models')\ .setInputCols(["ner_chunk"])\ .setOutputCol("sbert_embeddings") snomed_ct_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_snomed_findings","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("snomed_code")\ .setDistanceFunction("EUCLIDEAN") icd10_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_icd10cm_augmented","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("icd10cm_code")\ .setDistanceFunction("EUCLIDEAN") rxnorm_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_rxnorm","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("rxnorm_code")\ .setDistanceFunction("EUCLIDEAN") cpt_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_cpt_procedures_augmented","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("cpt_code")\ .setDistanceFunction("EUCLIDEAN") hcc_resolver = SentenceEntityResolverModel.pretrained("sbert_biobertresolve_icd10cm_augmented_billable_hcc","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("icd10cm_hcc_code")\ .setDistanceFunction("EUCLIDEAN") loinc_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_loinc", "en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("loinc_code")\ .setDistanceFunction("EUCLIDEAN") umls_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_umls_major_concepts", "en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("umls_code")\ .setDistanceFunction("EUCLIDEAN") hpo_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_HPO", "en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("umls_code")\ .setDistanceFunction("EUCLIDEAN") resolver_pipelineModel = PipelineModel( stages = [ documentAssembler, sbert_embedder, snomed_ct_resolver, icd10_resolver, rxnorm_resolver, cpt_resolver, hcc_resolver, loinc_resolver, umls_resolver, hpo_resolver]) resolver_lp = LightPipeline(resolver_pipelineModel) ###Output _____no_output_____ ###Markdown Utility functions ###Code import pandas as pd pd.set_option('display.max_colwidth', 0) def get_codes (lp, text, vocab='icd10cm_code', hcc=False, aux_label=False): full_light_result = lp.fullAnnotate(text) chunks = [] codes = [] begin = [] end = [] resolutions=[] all_distances =[] all_codes=[] all_cosines = [] all_k_aux_labels=[] for chunk, code in zip(full_light_result[0]['ner_chunk'], full_light_result[0][vocab]): begin.append(chunk.begin) end.append(chunk.end) chunks.append(chunk.result) codes.append(code.result) all_codes.append(code.metadata['all_k_results'].split(':::')) resolutions.append(code.metadata['all_k_resolutions'].split(':::')) all_distances.append(code.metadata['all_k_distances'].split(':::')) all_cosines.append(code.metadata['all_k_cosine_distances'].split(':::')) if hcc: try: all_k_aux_labels.append(code.metadata['all_k_aux_labels'].split(':::')) except: all_k_aux_labels.append([]) elif aux_label: try: all_k_aux_labels.append(code.metadata['all_k_aux_labels'].split(':::')) except: all_k_aux_labels.append([]) else: all_k_aux_labels.append([]) df = pd.DataFrame({'chunks':chunks, 'begin': begin, 'end':end, 'code':codes,'all_codes':all_codes, 'resolutions':resolutions, 'all_k_aux_labels':all_k_aux_labels,'all_distances':all_cosines}) if hcc: df['billable'] = df['all_k_aux_labels'].apply(lambda x: [i.split('||')[0] for i in x]) df['hcc_status'] = df['all_k_aux_labels'].apply(lambda x: [i.split('||')[1] for i in x]) df['hcc_score'] = df['all_k_aux_labels'].apply(lambda x: [i.split('||')[2] for i in x]) elif aux_label: df['gt'] = df['all_k_aux_labels'].apply(lambda x: [i.split('|')[0] for i in x]) df['concept'] = df['all_k_aux_labels'].apply(lambda x: [i.split('|')[1] for i in x]) df['aux'] = df['all_k_aux_labels'].apply(lambda x: [i.split('|')[2] for i in x]) df = df.drop(['all_k_aux_labels'], axis=1) return df ###Output _____no_output_____ ###Markdown Getting some predictions from resolvers ###Code text = 'bladder cancer' %time get_codes (icd_lp, text, vocab='icd10cm_code') text = 'severe stomach pain' %time get_codes (icd_lp, text, vocab='icd10cm_code') text = 'bladder cancer' %time get_codes (hcc_lp, text, vocab='icd10cm_hcc_code', hcc=True) text = 'severe stomach pain' %time get_codes (hcc_lp, text, vocab='icd10cm_hcc_code', hcc=True) text = 'bladder cancer' %time get_codes (snomed_lp, text, vocab='snomed_code', aux_label=True) text = 'schizophrenia' %time get_codes (snomed_lp, text, vocab='snomed_code', aux_label=True) text = 'metformin 100 mg' %time get_codes (rxnorm_lp, text, vocab='rxnorm_code') text = 'Advil Allergy Sinus' %time get_codes (rxnorm_lp, text, vocab='rxnorm_code') text = 'heart surgery' %time get_codes (cpt_lp, text, vocab='cpt_code') text = 'ct abdomen' %time get_codes (cpt_lp, text, vocab='cpt_code') text = 'Left heart cath' %time get_codes (cpt_lp, text, vocab='cpt_code') text = 'FLT3 gene mutation analysis' %time get_codes (loinc_lp, text, vocab='loinc_code') text = 'Hematocrit' %time get_codes (loinc_lp, text, vocab='loinc_code') text = 'urine test' %time get_codes (loinc_lp, text, vocab='loinc_code') # medical device text = 'X-Ray' %time get_codes (umls_lp, text, vocab='umls_code') # Injuries & poisoning text = 'out-of-date food poisoning' %time get_codes (umls_lp, text, vocab='umls_code') # clinical findings text = 'type two diabetes mellitus' %time get_codes (umls_lp, text, vocab='umls_code') text = 'bladder cancer' %time get_codes (hpo_lp, text, vocab='umls_code') text = 'bipolar disorder' %time get_codes (hpo_lp, text, vocab='umls_code') text = 'schizophrenia ' %time get_codes (hpo_lp, text, vocab='umls_code') icd_chunks = ['advanced liver disease', 'advanced lung disease', 'basal cell carcinoma of skin', 'acute maxillary sinusitis', 'chronic kidney disease stage', 'diabetes mellitus type 2', 'lymph nodes of multiple sites', 'other chronic pain', 'severe abdominal pain', 'squamous cell carcinoma of skin', 'type 2 diabetes mellitus'] snomed_chunks= ['down syndrome', 'adenocarcinoma', 'aortic valve stenosis', 'atherosclerosis', 'atrial fibrillation', 'hypertension', 'lung cancer', 'seizure', 'squamous cell carcinoma', 'stage IIIB', 'mediastinal lymph nodes'] from IPython.display import display for chunk in icd_chunks: print ('>> ',chunk) display(get_codes (icd_lp, chunk, vocab='icd10cm_code')) for chunk in snomed_chunks: print ('>> ',chunk) display(get_codes (snomed_lp, chunk, vocab='snomed_code', aux_label=True)) clinical_chunks = ['bladder cancer', 'anemia in chronic kidney disease', 'castleman disease', 'congestive heart failure', 'diabetes mellitus type 2', 'lymph nodes of multiple sites', 'malignant melanoma of skin', 'malignant neoplasm of lower lobe, bronchus', 'metastatic lung cancer', 'secondary malignant neoplasm of bone', 'type 2 diabetes mellitus', 'type 2 diabetes mellitus/insulin', 'unsp malignant neoplasm of lymph node'] for chunk in clinical_chunks: print ('>> ',chunk) print ('icd10cm_code') display(get_codes (hcc_lp, chunk, vocab='icd10cm_hcc_code', hcc=True)) print ('snomed_code') display(get_codes (snomed_lp, chunk, vocab='snomed_code', aux_label=True)) ###Output >> bladder cancer icd10cm_code ###Markdown How to integrate resolvers with NER models in the same pipeline ###Code documentAssembler = DocumentAssembler()\ .setInputCol("text")\ .setOutputCol("document") sentenceDetector = SentenceDetectorDLModel.pretrained()\ .setInputCols(["document"])\ .setOutputCol("sentence") tokenizer = Tokenizer()\ .setInputCols(["sentence"])\ .setOutputCol("token")\ word_embeddings = WordEmbeddingsModel.pretrained("embeddings_clinical", "en", "clinical/models")\ .setInputCols(["sentence", "token"])\ .setOutputCol("embeddings") clinical_ner = MedicalNerModel.pretrained("ner_clinical", "en", "clinical/models") \ .setInputCols(["sentence", "token", "embeddings"]) \ .setOutputCol("ner") ner_converter = NerConverter() \ .setInputCols(["sentence", "token", "ner"]) \ .setOutputCol("ner_chunk")\ .setWhiteList(['PROBLEM']) c2doc = Chunk2Doc()\ .setInputCols("ner_chunk")\ .setOutputCol("ner_chunk_doc") sbert_embedder = BertSentenceEmbeddings\ .pretrained("sbiobert_base_cased_mli",'en','clinical/models')\ .setInputCols(["ner_chunk_doc"])\ .setOutputCol("sbert_embeddings") icd10_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_icd10cm_augmented","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("icd10cm_code")\ .setDistanceFunction("EUCLIDEAN") sbert_resolver_pipeline = Pipeline( stages = [ documentAssembler, sentenceDetector, tokenizer, word_embeddings, clinical_ner, ner_converter, c2doc, sbert_embedder, icd10_resolver]) data_ner = spark.createDataFrame([[""]]).toDF("text") sbert_models = sbert_resolver_pipeline.fit(data_ner) clinical_note = 'A 28-year-old female with a history of gestational diabetes mellitus diagnosed eight years prior to presentation and subsequent type two diabetes mellitus (T2DM), one prior episode of HTG-induced pancreatitis three years prior to presentation, associated with an acute hepatitis, and obesity with a body mass index (BMI) of 33.5 kg/m2, presented with a one-week history of polyuria, polydipsia, poor appetite, and vomiting. Two weeks prior to presentation, she was treated with a five-day course of amoxicillin for a respiratory tract infection. She was on metformin, glipizide, and dapagliflozin for T2DM and atorvastatin and gemfibrozil for HTG. She had been on dapagliflozin for six months at the time of presentation. Physical examination on presentation was significant for dry oral mucosa; significantly, her abdominal examination was benign with no tenderness, guarding, or rigidity. Pertinent laboratory findings on admission were: serum glucose 111 mg/dl, bicarbonate 18 mmol/l, anion gap 20, creatinine 0.4 mg/dL, triglycerides 508 mg/dL, total cholesterol 122 mg/dL, glycated hemoglobin (HbA1c) 10%, and venous pH 7.27. Serum lipase was normal at 43 U/L. Serum acetone levels could not be assessed as blood samples kept hemolyzing due to significant lipemia. The patient was initially admitted for starvation ketosis, as she reported poor oral intake for three days prior to admission. However, serum chemistry obtained six hours after presentation revealed her glucose was 186 mg/dL, the anion gap was still elevated at 21, serum bicarbonate was 16 mmol/L, triglyceride level peaked at 2050 mg/dL, and lipase was 52 U/L. The β-hydroxybutyrate level was obtained and found to be elevated at 5.29 mmol/L - the original sample was centrifuged and the chylomicron layer removed prior to analysis due to interference from turbidity caused by lipemia again. The patient was treated with an insulin drip for euDKA and HTG with a reduction in the anion gap to 13 and triglycerides to 1400 mg/dL, within 24 hours. Her euDKA was thought to be precipitated by her respiratory tract infection in the setting of SGLT2 inhibitor use. The patient was seen by the endocrinology service and she was discharged on 40 units of insulin glargine at night, 12 units of insulin lispro with meals, and metformin 1000 mg two times a day. It was determined that all SGLT2 inhibitors should be discontinued indefinitely. She had close follow-up with endocrinology post discharge.' print (clinical_note) clinical_note_df = spark.createDataFrame([[clinical_note]]).toDF("text") icd10_result = sbert_models.transform(clinical_note_df) pd.set_option("display.max_colwidth",0) import pandas as pd pd.set_option('display.max_colwidth', 0) def get_icd_codes(icd10_res): icd10_df = icd10_res.select(F.explode(F.arrays_zip('ner_chunk.result', 'ner_chunk.metadata', 'icd10cm_code.result', 'icd10cm_code.metadata')).alias("cols")) \ .select(F.expr("cols['1']['sentence']").alias("sent_id"), F.expr("cols['0']").alias("ner_chunk"), F.expr("cols['1']['entity']").alias("entity"), F.expr("cols['2']").alias("icd10_code"), F.expr("cols['3']['all_k_results']").alias("all_codes"), F.expr("cols['3']['all_k_resolutions']").alias("resolutions")).toPandas() codes = [] resolutions = [] for code, resolution in zip(icd10_df['all_codes'], icd10_df['resolutions']): codes.append(code.split(':::')) resolutions.append(resolution.split(':::')) icd10_df['all_codes'] = codes icd10_df['resolutions'] = resolutions return icd10_df %%time res_pd = get_icd_codes(icd10_result) res_pd.head(15) ###Output CPU times: user 57.6 ms, sys: 12 ms, total: 69.6 ms Wall time: 3min 3s ###Markdown Lets apply some HTML formating by using `sparknlp_display` library to see the results of the pipeline in a nicer layout: ###Code from sparknlp_display import EntityResolverVisualizer # with light pipeline light_model = LightPipeline(sbert_models) vis = EntityResolverVisualizer() # Change color of an entity label vis.set_label_colors({'PROBLEM':'#008080'}) light_data_icd = light_model.fullAnnotate(clinical_note) vis.display(light_data_icd[0], 'ner_chunk', 'icd10cm_code') ###Output _____no_output_____ ###Markdown BertSentenceChunkEmbeddingsThis annotator let users to aggregate sentence embeddings and ner chunk embeddings to get more specific and accurate resolution codes. It works by averaging context and chunk embeddings to get contextual information. Input to this annotator is the context (sentence) and ner chunks, while the output is embedding for each chunk that can be fed to the resolver model. The `setChunkWeight` parameter can be used to control the influence of surrounding context. For more information and examples of `BertSentenceChunkEmbeddings` annotator, you can check here: [24.1.Improved_Entity_Resolution_with_SentenceChunkEmbeddings.ipynb](https://github.com/JohnSnowLabs/spark-nlp-workshop/blob/master/tutorials/Certification_Trainings/Healthcare/24.1.Improved_Entity_Resolution_with_SentenceChunkEmbeddings.ipynb) ICD10CM with BertSentenceChunkEmbeddingsLets do the same process by using `BertSentenceEmbeddings` annotator and compare the results. We will create a new pipeline by using this annotator with SentenceEntityResolverModel. ###Code #Get average sentence-chunk Bert embeddings sentence_chunk_embeddings = BertSentenceChunkEmbeddings.pretrained("sbiobert_base_cased_mli", "en", "clinical/models")\ .setInputCols(["sentence", "ner_chunk"])\ .setOutputCol("sbert_embeddings")\ .setChunkWeight(0.5) #default : 0.5 icd10_resolver = SentenceEntityResolverModel.pretrained("sbiobertresolve_icd10cm_augmented","en", "clinical/models") \ .setInputCols(["ner_chunk", "sbert_embeddings"]) \ .setOutputCol("icd10cm_code")\ .setDistanceFunction("EUCLIDEAN") resolver_pipeline_SCE = Pipeline( stages = [ documentAssembler, sentenceDetector, tokenizer, word_embeddings, clinical_ner, ner_converter, sentence_chunk_embeddings, icd10_resolver]) empty_data = spark.createDataFrame([['']]).toDF("text") model_SCE = resolver_pipeline_SCE.fit(empty_data) icd10_result_SCE = model_SCE.transform(clinical_note_df) %%time res_SCE_pd = get_icd_codes(icd10_result_SCE) res_SCE_pd.head(15) icd10_SCE_lp = LightPipeline(model_SCE) light_result = icd10_SCE_lp.fullAnnotate(clinical_note) visualiser = EntityResolverVisualizer() # Change color of an entity label visualiser.set_label_colors({'PROBLEM':'#008080'}) visualiser.display(light_result[0], 'ner_chunk', 'icd10cm_code') ###Output _____no_output_____ ###Markdown **Lets compare the results that we got from these two methods.** ###Code sentence_df = icd10_result.select(F.explode(F.arrays_zip('sentence.metadata', 'sentence.result')).alias("cols")) \ .select( F.expr("cols['0']['sentence']").alias("sent_id"), F.expr("cols['1']").alias("sentence_all")).toPandas() comparison_df = pd.merge(res_pd.loc[:,'sent_id':'resolutions'],res_SCE_pd.loc[:,'sent_id':'resolutions'], on=['sent_id',"ner_chunk", "entity"], how='inner') comparison_df.columns=['sent_id','ner_chunk', 'entity', 'icd10_code', 'all_codes', 'resolutions', 'icd10_code_SCE', 'all_codes_SCE', 'resolutions_SCE'] comparison_df = pd.merge(sentence_df, comparison_df,on="sent_id").drop('sent_id', axis=1) comparison_df.head(15) ###Output _____no_output_____
src/notebooks/010_Parse_Data.ipynb
###Markdown Some Notes* Final impression counts in state and demo totals are the same, but slightly _above_ the summary "Campaign Performance sheet* Final impressinon counts from the ZIP code tables are 60% lower* The data does not contain details on the individual combinations that were put together in each location. However, they do have a description in final sheet (not described here) of "Low, Good, Best" for the individual asset combinations Appending Census Data By ZIP Code (ZCTA?) ###Code c = Census(os.environ["CENSUS_API_KEY"]) zcta_df = pd.DataFrame( c.acs5.zipcode(["B01001_001E"], zcta="*", state_fips="*", year=2019) ) zcta_df = zcta_df.rename( columns={ "B01001_001E": "total_population", "state": "state_fips", "zip code tabulation area": "zipcode", } ) merged_zip_df = final_zip_df.merge(zcta_df, how="left", on="zipcode") merged_zip_df.to_parquet(OUTPUT_DATA_DIR / "data_by_zipcode_merged_census.parquet") merged_zip_df[merged_zip_df["state_fips"].isna()] ###Output _____no_output_____ ###Markdown More notesSo that first ZIP code (06338) is a [P.O. Box ZIP Code](https://www.zip-codes.com/zip-code/06338/zip-code-06338.asp), which strikes me as odd. What exactly is going on with this ZIP code data. (Note that that website is usually accurate, though you should really double check with SmartyStreets, which has the raw USPS data.) ###Code table = final_state_df.groupby("language")["clicks", "impressions"].sum() print(table["clicks"] / table["impressions"]) table["impressions"] = table["impressions"] - table["clicks"] st.contingency.chi2_contingency(table.values.T) table = ( final_state_df[final_state_df["language"] == "en"] .groupby(["campaign"])[["clicks", "impressions"]] .sum() ) print(table["clicks"] / table["impressions"]) table = ( final_state_df[final_state_df["language"] == "sp"] .groupby(["campaign"])[["clicks", "impressions"]] .sum() ) print(table["clicks"] / table["impressions"]) ###Output campaign Helping Community SP 0.016799 Helping Others SP 0.012042 Personal Responsibility SP 0.011959 Self-Oriented SP 0.013431 dtype: float64 ###Markdown Explode dataHansoo requests the data be broken into one line per impression. This is done below. ###Code for filename in OUTPUT_DATA_DIR.glob("data*.parquet"): print(f"Exploding {filename.name}...") df = pd.read_parquet(filename) df["num_in_group"] = [ np.arange(num_impressions, dtype=int) for num_impressions in df["impressions"].values ] exploded_df = df.explode("num_in_group") exploded_df["was_clicked"] = exploded_df["num_in_group"] < exploded_df["clicks"] exploded_df = exploded_df.drop( columns=["clicks", "impressions", "ctr", "cost", "average_cpc", "num_in_group"] ) exploded_df.to_parquet(EXPLODED_DATA_DIR / f"exploded_{filename.name}") exploded_df.to_csv( EXPLODED_DATA_DIR / f"exploded_{filename.with_suffix('').name}.csv.gz", index=False, ) print("Done") ###Output Exploding data_by_state.parquet... Exploding data_by_demographics.parquet... Exploding data_by_zipcode_merged_census.parquet... Exploding data_by_zipcode.parquet... Done
2-Data-Analysis.ipynb
###Markdown Data Analysis 1. Loading the data ###Code import re from nltk import sent_tokenize cg_sents = [] smg_sents = [] def remove_duplicate_punctuation(s): # sent_tokenize() gets confused when there's duplicate punctuation return(re.sub(r'(\.|\?|!|;)\1+', r'\1 ', s)) # In a previous version of this I used r'\1' instead of r'\1 '. However it caused a problem for sentences like "this is great...I don't know what to do." It seems that some people do not use spacing with ellipses which caused sent_tokenize() to not identify a new sentence. with open('./Data/cg_twitter.txt', 'r', encoding='utf-8') as in_file: text = remove_duplicate_punctuation(in_file.read()) text = re.sub(r'([a-zα-ωίϊΐόάέύϋΰήώ])(\.|\?|;|!)([A-ZΑ-ΩΆΈΊΌΎΏΉ])', r'\1\2 \3', text) # https://stackoverflow.com/questions/44858741/nltk-tokenizer-and-stanford-corenlp-tokenizer-cannot-distinct-2-sentences-withou lines = [p for p in text.split('\n') if p] # sent_tokenize() doesn't consider a new line a new sentence so this is required. for line in lines: cg_sents += sent_tokenize(line) with open('./Data/cg_fb.txt', 'r', encoding='utf-8') as in_file: text = remove_duplicate_punctuation(in_file.read()) text = re.sub(r'([a-zα-ωίϊΐόάέύϋΰήώ])(\.|\?|;|!)([A-ZΑ-ΩΆΈΊΌΎΏΉ])', r'\1\2 \3', text) lines = [p for p in text.split('\n') if p] for line in lines: cg_sents += sent_tokenize(line) with open('./Data/cg_other.txt', 'r', encoding='utf-8') as in_file: text = remove_duplicate_punctuation(in_file.read()) text = re.sub(r'([a-zα-ωίϊΐόάέύϋΰήώ])(\.|\?|;|!)([A-ZΑ-ΩΆΈΊΌΎΏΉ])', r'\1\2 \3', text) lines = [p for p in text.split('\n') if p] for line in lines: cg_sents += sent_tokenize(line) with open('./Data/smg_twitter.txt', 'r', encoding='utf-8') as in_file: text = remove_duplicate_punctuation(in_file.read()) text = re.sub(r'([a-zα-ωίϊΐόάέύϋΰήώ])(\.|\?|;|!)([A-ZΑ-ΩΆΈΊΌΎΏΉ])', r'\1\2 \3', text) lines = [p for p in text.split('\n') if p] for line in lines: smg_sents += sent_tokenize(line) with open('./Data/smg_fb.txt', 'r', encoding='utf-8') as in_file: text = remove_duplicate_punctuation(in_file.read()) text = re.sub(r'([a-zα-ωίϊΐόάέύϋΰήώ])(\.|\?|;|!)([A-ZΑ-ΩΆΈΊΌΎΏΉ])', r'\1\2 \3', text) lines = [p for p in text.split('\n') if p] for line in lines: smg_sents += sent_tokenize(line) with open('./Data/smg_other.txt', 'r', encoding='utf-8') as in_file: text = remove_duplicate_punctuation(in_file.read()) text = re.sub(r'([a-zα-ωίϊΐόάέύϋΰήώ])(\.|\?|;|!)([A-ZΑ-ΩΆΈΊΌΎΏΉ])', r'\1\2 \3', text) lines = [p for p in text.split('\n') if p] for line in lines: smg_sents += sent_tokenize(line) cg_sents[:3] ###Output _____no_output_____ ###Markdown 2. Cleaning the data ###Code import unicodedata from string import punctuation from nltk.tokenize import WhitespaceTokenizer punctuation += '´΄’…“”–—―»«' # string.punctuation misses these. def strip_accents(s): return ''.join(c for c in unicodedata.normalize('NFD', s) if unicodedata.category(c) != 'Mn') def get_clean_sent_el(sentence): sentence = re.sub(r'^RT', '', sentence) sentence = re.sub(r'\&\w*;', '', sentence) sentence = re.sub(r'\@\w*', '', sentence) sentence = re.sub(r'\$\w*', '', sentence) sentence = re.sub(r'https?:\/\/.*\/\w*', '', sentence) sentence = ''.join(c for c in sentence if c <= '\uFFFF') sentence = strip_accents(sentence) sentence = re.sub(r'#\w*', '', sentence) sentence = sentence.lower() tokens = WhitespaceTokenizer().tokenize(sentence) new_tokens = [] for token in tokens: if token == 'ο,τι' or token == 'ό,τι' or token == 'o,ti' or token == 'ó,ti': new_tokens.append(token) else: token = re.sub(r'(?<=[.,!\?;\'΄´’])(?=[^\s])', r' ', token) # If there is punctuation not followed by a space, we add it. I also added a space after apostrophes because I want something like σαγαπώ to not be considered as one word. new_token = token.translate(str.maketrans({key: None for key in punctuation})) if (new_token != ''): # This might happen if a user surrounds commas with spaces , like so. new_tokens.append(new_token) sentence =' '.join(new_tokens) sentence = re.sub('\ufeff', '', sentence) # \ufeff might appear when dealing with unicode-encoded files sentence = sentence.strip(' ') # performs lstrip() and rstrip() sentence = re.sub(' ', ' ', sentence) # Adding a space after the apostrophe can lead to the appearance of double spaces if apostrophes are used along with spaces in the original text. return sentence cg_sents_clean = [] smg_sents_clean = [] for sent in cg_sents: cg_sents_clean.append(get_clean_sent_el(sent)) for sent in smg_sents: smg_sents_clean.append(get_clean_sent_el(sent)) # Remove empty strings left due to sentences ending up being only URLs then getting deleted on cleaning: cg_sents_clean = list(filter(None, cg_sents_clean)) smg_sents_clean = list(filter(None, smg_sents_clean)) cg_sents_clean[:3] ###Output _____no_output_____ ###Markdown 3. Tokenization and setting up to use `nltk.text` 3.1 Tokenization ###Code cg_sents_tokens = [] smg_sents_tokens = [] for sent in cg_sents_clean: cg_sents_tokens.append(WhitespaceTokenizer().tokenize(sent)) for sent in smg_sents_clean: smg_sents_tokens.append(WhitespaceTokenizer().tokenize(sent)) cg_sents_tokens[:3] ###Output _____no_output_____ ###Markdown 3.2 Setting up to use `nltk.text` 3.2.1 Words `Text` objects ###Code from nltk.text import Text cg_words_flat = [word for sent_tokens in cg_sents_tokens for word in sent_tokens] smg_words_flat = [word for sent_tokens in smg_sents_tokens for word in sent_tokens] cg_Text = Text(cg_words_flat) smg_Text = Text(smg_words_flat) cg_words_flat = [re.sub(r'ς', 'σ', word) for word in cg_words_flat] # The reason I replace ς with σ for char n-grams is because the Greek final sigma form is arbitrary. No other letter has a final form. When counting the number of sigmas I would like that to include the final forms as well. The σ_ bigram will represent the ς feature so it can be removed. smg_words_flat = [re.sub(r'ς', 'σ', word) for word in smg_words_flat] cg_Text ###Output _____no_output_____ ###Markdown 3.2.2 Word n-grams 3.2.2.1 Bigrams `Text` objects ###Code from nltk import ngrams cg_word_bigrams = [] smg_word_bigrams = [] for sent in cg_sents_tokens: cg_word_bigrams.append(list(ngrams(sent, 2))) for sent in smg_sents_tokens: smg_word_bigrams.append(list(ngrams(sent, 2))) print(cg_word_bigrams) cg_word_bigrams_flat_tuples = [bigram for bigram_list in cg_word_bigrams for bigram in bigram_list] smg_word_bigrams_flat_tuples = [bigram for bigram_list in smg_word_bigrams for bigram in bigram_list] cg_word_bigrams_flat = ['%s %s' % bigram_tuple for bigram_tuple in cg_word_bigrams_flat_tuples] smg_word_bigrams_flat = ['%s %s' % bigram_tuple for bigram_tuple in smg_word_bigrams_flat_tuples] cg_word_bigrams_Text = Text(cg_word_bigrams_flat) smg_word_bigrams_Text = Text(smg_word_bigrams_flat) cg_word_bigrams_Text ###Output [[('πρασινο', 'αυκουι'), ('αυκουι', 'μες'), ('μες', 'το'), ('το', 'πασχαλινο'), ('πασχαλινο', 'ποτηρι'), ('ποτηρι', 'που'), ('που', 'επιασε'), ('επιασε', 'ο'), ('ο', 'μιτσης')], [('καμνουν', 'πολλα'), ('πολλα', 'ανακαινιση'), ('ανακαινιση', 'στα'), ('στα', 'περβολια')], [('ελα', 'συγγενη'), ('συγγενη', 'τζιαι'), ('τζιαι', 'εχουμε'), ('εχουμε', 'νεοτερα'), ('νεοτερα', 'π'), ('π', 'το'), ('το', 'νικολη')], [('η', 'αληθκεια'), ('αληθκεια', 'με'), ('με', 'τους'), ('τους', 'τερματοφυλακες'), ('τερματοφυλακες', 'και'), ('και', 'οχι'), ('οχι', 'μονο'), ('μονο', 'εχουμε'), ('εχουμε', 'λιο'), ('λιο', 'θεμα')], [('μεχρι', 'και'), ('και', 'αυγουστη'), ('αυγουστη', 'και'), ('και', 'πανο'), ('πανο', 'κωνσταντινου'), ('κωνσταντινου', 'εφεραμε'), ('εφεραμε', 'ρε')], [('ωσπου', 'να'), ('να', 'φαεις'), ('φαεις', 'το'), ('το', 'κραμπι'), ('κραμπι', 'σου'), ('σου', 'εν'), ('εν', 'να'), ('να', 'τελειωσω'), ('τελειωσω', 'μεν'), ('μεν', 'φοασαι')], [('πρεπει', 'να'), ('να', 'εβρω'), ('εβρω', 'αλλη'), ('αλλη', 'λεξη'), ('λεξη', 'για'), ('για', 'το'), ('το', 'τι'), ('τι', 'εγινε'), ('εγινε', 'σημερα'), ('σημερα', 'το'), ('το', 'σοκ'), ('σοκ', 'εν'), ('εν', 'πολλα'), ('πολλα', 'λαιτ')], [('θεωρω', 'οτι'), ('οτι', 'η'), ('η', 'παντελινα'), ('παντελινα', 'εν'), ('εν', 'δυνατη'), ('δυνατη', 'ανετα'), ('ανετα', 'λαλω'), ('λαλω', 'εγω')], [('εν', 'τουτη'), ('τουτη', 'η'), ('η', 'νεα'), ('νεα', 'ομονοια'), ('ομονοια', 'που'), ('που', 'μας'), ('μας', 'εταξαν')], [('ηντα', 'ωραια'), ('ωραια', 'που'), ('που', 'εφυετε'), ('εφυετε', 'ουλλοι'), ('ουλλοι', 'τζιαι'), ('τζιαι', 'εμεινε'), ('εμεινε', 'μας'), ('μας', 'εμας'), ('εμας', 'η'), ('η', 'θαλασσα')], [('ο', 'ενας'), ('ενας', 'παππους'), ('παππους', 'αποστολος'), ('αποστολος', 'ο'), ('ο', 'αλλος'), ('αλλος', 'παππους'), ('παππους', 'αντρεας'), ('αντρεας', 'μαντεψε'), ('μαντεψε', 'τι'), ('τι', 'εφκαλαν'), ('εφκαλαν', 'το'), ('το', 'μωρο')], [('η', 'κορη'), ('κορη', 'μου'), ('μου', 'η'), ('η', 'μιτσια'), ('μιτσια', 'εν'), ('εν', 'να'), ('να', 'με'), ('με', 'γεροκομησει')], [('η', 'μεγαλη'), ('μεγαλη', 'απλα'), ('απλα', 'εν'), ('εν', 'να'), ('να', 'με'), ('με', 'ποφκαλει'), ('ποφκαλει', 'πριν'), ('πριν', 'την'), ('την', 'ωρα'), ('ωρα', 'μου')], [('οι', 'εδιωξα'), ('εδιωξα', 'τον'), ('τον', 'ουτε'), ('ουτε', 'βιλλο'), ('βιλλο', 'εν'), ('εν', 'εφαγα')], [('τουτον', 'το'), ('το', 'καλο'), ('καλο', 'μηνα'), ('μηνα', 'γιατι'), ('γιατι', 'το'), ('το', 'αντιγραφετε'), ('αντιγραφετε', 'τζιαι'), ('τζιαι', 'τουτο'), ('τουτο', 'που'), ('που', 'τους'), ('τους', 'πρηχτηες'), ('πρηχτηες', 'τους'), ('τους', 'καλαμαραες'), ('καλαμαραες', 'τζιαι'), ('τζιαι', 'καμνετε'), ('καμνετε', 'μας'), ('μας', 'τα'), ('τα', 'σατσιην'), ('σατσιην', 'καθε'), ('καθε', 'μηνα')], [('ατε', 'τζιαι'), ('τζιαι', 'εφκαλαν'), ('εφκαλαν', 'τα'), ('τα', 'πανο'), ('πανο', 'τα'), ('τα', 'οχι'), ('οχι', 'στην'), ('στην', 'ισοπαλια'), ('ισοπαλια', 'γιε'), ('γιε', 'μου')], [('θκιαβαζω', 'οτι'), ('οτι', 'εγεμωσεν'), ('εγεμωσεν', 'γκραφιτι'), ('γκραφιτι', 'η'), ('η', 'παφος'), ('παφος', 'τζιαι'), ('τζιαι', 'σιερουμαι'), ('σιερουμαι', 'απιστευτα'), ('απιστευτα', 'πολλα')], [('ασσιημο', 'πραμα'), ('πραμα', 'να'), ('να', 'μεν'), ('μεν', 'γαμας')], [('πιανει', 'με'), ('με', 'τηλεφωνο'), ('τηλεφωνο', 'ο'), ('ο', 'γερος'), ('γερος', 'μου'), ('μου', 'κλαμενος'), ('κλαμενος', 'θα'), ('θα', 'αυξηθει'), ('αυξηθει', 'η'), ('η', 'συνταξη'), ('συνταξη', 'του'), ('του', 'λαλει'), ('λαλει', 'εν'), ('εν', 'ξερει')], [('να', 'τον'), ('τον', 'τρωεις'), ('τρωεις', 'ουλλη'), ('ουλλη', 'μερα')], [('εκαμνα', 'της'), ('της', 'το'), ('το', 'ανετα'), ('ανετα', 'αλλα'), ('αλλα', 'επηρεν'), ('επηρεν', 'τα'), ('τα', 'μαζιν'), ('μαζιν', 'του')], [('πομπα', 'ηβρα'), ('ηβρα', 'την'), ('την', 'επιτελους'), ('επιτελους', 'εχω'), ('εχω', 'καμποσους'), ('καμποσους', 'ποντους'), ('ποντους', 'για'), ('για', 'εξαργυρωση')], [('μα', 'ο'), ('ο', 'μαρκασας'), ('μαρκασας', 'ηντα'), ('ηντα', 'ενεφκε'), ('ενεφκε', 'μολις'), ('μολις', 'εβαλε'), ('εβαλε', 'το'), ('το', 'γκολ'), ('γκολ', 'επηεννε'), ('επηεννε', 'για'), ('για', 'την'), ('την', 'μεγαλη'), ('μεγαλη', 'αντεπιθεση')], [('δηλαδη', 'εσας'), ('εσας', 'τζιαι'), ('τζιαι', 'βιλλο'), ('βιλλο', 'να'), ('να', 'σας'), ('σας', 'προσφερουν'), ('προσφερουν', 'για'), ('για', 'λυση'), ('λυση', 'παλι'), ('παλι', 'ναι'), ('ναι', 'θα'), ('θα', 'πειτε'), ('πειτε', 'ατε')], [('για', 'τα'), ('τα', 'πραματα'), ('πραματα', 'που'), ('που', 'εννεν'), ('εννεν', 'ασπρομαυρο'), ('ασπρομαυρο', 'εν'), ('εν', 'η'), ('η', 'κοτα'), ('κοτα', 'τζιαι'), ('τζιαι', 'το'), ('το', 'αφκο')], [('εν', 'να'), ('να', 'φορησω'), ('φορησω', 'τα'), ('τα', 'τζινουρκα'), ('τζινουρκα', 'μου'), ('μου', 'τα'), ('τα', 'ρουχα'), ('ρουχα', 'αυριο'), ('αυριο', 'στο'), ('στο', 'μνημοσυνο')], [('ουτε', 'ενα'), ('ενα', 'μαντορινι'), ('μαντορινι', 'εν'), ('εν', 'τους'), ('τους', 'αφηννε'), ('αφηννε', 'να'), ('να', 'κοψουν'), ('κοψουν', 'ρε'), ('ρε', 'ο'), ('ο', 'ππιντης')], [('εννα', 'μας'), ('μας', 'φκαλετε'), ('φκαλετε', 'την'), ('την', 'ψυσιη'), ('ψυσιη', 'μας'), ('μας', 'για'), ('για', 'ενα'), ('ενα', 'λουρι'), ('λουρι', 'ελεος')], [('ειμαστε', 'πλουσιοι'), ('πλουσιοι', 'τζιαι'), ('τζιαι', 'χωρις'), ('χωρις', 'την'), ('την', 'κοπιαν'), ('κοπιαν', 'ολαν')], [('ελεος', 'οι'), ('οι', 'τζιαι'), ('τζιαι', 'με'), ('με', 'τον'), ('τον', 'μισιελ'), ('μισιελ', 'ολαν')], [('εν', 'με'), ('με', 'αφηνει'), ('αφηνει', 'να'), ('να', 'την'), ('την', 'φκαλω'), ('φκαλω', 'βιντεο'), ('βιντεο', 'την'), ('την', 'ωρα'), ('ωρα', 'που'), ('που', 'το'), ('το', 'καμνει'), ('καμνει', 'να'), ('να', 'γινω'), ('γινω', 'εκατομμυριουχος')], [('δευτερα', 'φκαινουν'), ('φκαινουν', 'μονον'), ('μονον', 'οι'), ('οι', 'γεροι')], [('φανταζομαι', 'τον'), ('τον', 'ερτογαν'), ('ερτογαν', 'να'), ('να', 'λαλει'), ('λαλει', 'βρωμοσιυλλε'), ('βρωμοσιυλλε', 'τουρκικε'), ('τουρκικε', 'λαε'), ('λαε', 'γνωριμη'), ('γνωριμη', 'η'), ('η', 'φωνη'), ('φωνη', 'που'), ('που', 'ακουεις')], [('κοπιασε', 'σειρηνα'), ('σειρηνα', 'μωρατσε'), ('μωρατσε', 'μου'), ('μου', 'να'), ('να', 'δροσιστουμε'), ('δροσιστουμε', 'να'), ('να', 'μας'), ('μας', 'καμει'), ('καμει', 'τζιαι'), ('τζιαι', 'μοχιτο'), ('μοχιτο', 'η'), ('η', 'θκια'), ('θκια', 'σου'), ('σου', 'η'), ('η', 'ελενα'), ('ελενα', 'μιλουμεν'), ('μιλουμεν', 'εν'), ('εν', 'εσιει'), ('εσιει', 'λαθος')], [('μια', 'χαρα'), ('χαρα', 'ετζοιμηθηκα'), ('ετζοιμηθηκα', 'παντως'), ('παντως', 'εγω'), ('εγω', 'οπως'), ('οπως', 'το'), ('το', 'πουλλουι')], [('εγω', 'εν'), ('εν', 'τζιαιν'), ('τζιαιν', 'κορες'), ('κορες', 'που'), ('που', 'εκαμα'), ('εκαμα', 'μανα'), ('μανα', 'μου'), ('μου', 'εν'), ('εν', 'τες'), ('τες', 'αρφαες'), ('αρφαες', 'της'), ('της', 'μανταμ'), ('μανταμ', 'σουσους')], [('ακουσαμεν', 'εναν'), ('εναν', 'σουσμαν'), ('σουσμαν', 'εναν'), ('εναν', 'κακον')], [('ατε', 'ρε'), ('ρε', 'καλαμαρουθκια'), ('καλαμαρουθκια', 'πιαστε'), ('πιαστε', 'της'), ('της', 'μαμμας'), ('μαμμας', 'σας'), ('σας', 'που'), ('που', 'ενα')], [('γιατι', 'καλαμαριζει'), ('καλαμαριζει', 'τουτη'), ('τουτη', 'η'), ('η', 'τυπισσα'), ('τυπισσα', 'τζεινη'), ('τζεινη', 'η'), ('η', 'λυσσιασμενη'), ('λυσσιασμενη', 'του'), ('του', 'βιλλου')], [('σημερα', 'ειπαν'), ('ειπαν', 'μου'), ('μου', 'τρια'), ('τρια', 'πλασματα'), ('πλασματα', 'οτι'), ('οτι', 'εμαυρισα')], [('τι', 'διλημμα'), ('διλημμα', 'εν'), ('εν', 'τουτο'), ('τουτο', 'παλε'), ('παλε', 'που'), ('που', 'εβρεθηκεν'), ('εβρεθηκεν', 'ομπρος'), ('ομπρος', 'μου')], [('να', 'δουμε'), ('δουμε', 'τι'), ('τι', 'δικαιολογια'), ('δικαιολογια', 'θα'), ('θα', 'εβρω'), ('εβρω', 'παλε'), ('παλε', 'για'), ('για', 'να'), ('να', 'μεν'), ('μεν', 'παω'), ('παω', 'ποψε')], [('θελω', 'να'), ('να', 'γινω'), ('γινω', 'ιατρικος'), ('ιατρικος', 'επισκεπτης'), ('επισκεπτης', 'ολαν'), ('ολαν', 'λαλω'), ('λαλω', 'σας')], [('μα', 'για'), ('για', 'ετσι'), ('ετσι', 'μοτορο'), ('μοτορο', 'εκαμνα'), ('εκαμνα', 'τζιαι'), ('τζιαι', 'τον'), ('τον', 'πελλο'), ('πελλο', 'λαιβ'), ('λαιβ', 'ολαν')], [('τζιαι', 'αριαν'), ('αριαν', 'τραουδω'), ('τραουδω', 'σου'), ('σου', 'τζιαι'), ('τζιαι', 'οπεραν'), ('οπεραν', 'τζιαι'), ('τζιαι', 'εντεχνον'), ('εντεχνον', 'τζιαι'), ('τζιαι', 'ρεμπετικον'), ('ρεμπετικον', 'τζιαι'), ('τζιαι', 'οτι'), ('οτι', 'τραβα'), ('τραβα', 'η'), ('η', 'ψυσιη'), ('ψυσιη', 'σου'), ('σου', 'τακκαρε'), ('τακκαρε', 'μου')], [('ουτε', 'να'), ('να', 'σιεσουμεν'), ('σιεσουμεν', 'πιον')], [('μεν', 'παρακαλατε'), ('παρακαλατε', 'κανενα'), ('κανενα', 'να'), ('να', 'ερτει'), ('ερτει', 'να'), ('να', 'δει'), ('δει', 'το'), ('το', 'αποελ')], [('ρε', 'χοχο'), ('χοχο', 'εμεις'), ('εμεις', 'εν'), ('εν', 'τζε'), ('τζε', 'αλλασουμεν'), ('αλλασουμεν', 'ομαδα'), ('ομαδα', 'καθε'), ('καθε', 'εφτομα'), ('εφτομα', 'τζε'), ('τζε', 'βιλλο'), ('βιλλο', 'μας'), ('μας', 'με'), ('με', 'ποιον'), ('ποιον', 'παιζει'), ('παιζει', 'η'), ('η', 'χοχα')], [('χοχοι', 'τουρκοσποροι'), ('τουρκοσποροι', 'πουσταραες'), ('πουσταραες', 'τον'), ('τον', 'βιλλο'), ('βιλλο', 'μας'), ('μας', 'θα'), ('θα', 'παιρνετε'), ('παιρνετε', 'ρε'), ('ρε', 'σκαταες')], [('εφκηκεν', 'η'), ('η', 'ψυσχιη'), ('ψυσχιη', 'μου'), ('μου', 'ωσπου'), ('ωσπου', 'να'), ('να', 'ρτω'), ('ρτω', 'εσσω'), ('εσσω', 'με'), ('με', 'τουντον'), ('τουντον', 'λαλλαρο'), ('λαλλαρο', 'που'), ('που', 'εχουμε'), ('εχουμε', 'σημερα')], [('εστειλα', 'τον'), ('τον', 'αντρα'), ('αντρα', 'μου'), ('μου', 'να'), ('να', 'δει'), ('δει', 'μαππα'), ('μαππα', 'να'), ('να', 'βρω'), ('βρω', 'την'), ('την', 'ησυχια'), ('ησυχια', 'μου'), ('μου', 'οι'), ('οι', 'να'), ('να', 'τες'), ('τες', 'φατε'), ('φατε', 'να'), ('να', 'ρτει'), ('ρτει', 'εσσω'), ('εσσω', 'γλιορα'), ('γλιορα', 'τωρα')], [('εβαρεθηκαμεν', 'σας'), ('σας', 'εθνικια'), ('εθνικια', 'που'), ('που', 'την'), ('την', 'μια'), ('μια', 'εν'), ('εν', 'ελληνικη'), ('ελληνικη', 'που'), ('που', 'την'), ('την', 'αλλη'), ('αλλη', 'εν'), ('εν', 'τουρτζικη'), ('τουρτζικη', 'κανει'), ('κανει', 'πκιον'), ('πκιον', 'εσσω'), ('εσσω', 'σας'), ('σας', 'ουλλοι')], [('ατε', 'πιεννε'), ('πιεννε', 'εσσω'), ('εσσω', 'σου'), ('σου', 'να'), ('να', 'πνασουμε')], [('εσχει', 'ετσι'), ('ετσι', 'πραμα'), ('πραμα', 'μονο'), ('μονο', 'σε'), ('σε', 'τζεινο'), ('τζεινο', 'το'), ('το', 'τμημα')], [('εν', 'ενοιωσα'), ('ενοιωσα', 'ακομα'), ('ακομα', 'τζεινο'), ('τζεινο', 'το'), ('το', 'πραμα'), ('πραμα', 'να'), ('να', 'πω'), ('πω', 'της'), ('της', 'αλλης'), ('αλλης', 'αγαπω'), ('αγαπω', 'σε')], [('ε', 'καλαν'), ('καλαν', 'ολαν'), ('ολαν', 'τζεινης'), ('τζεινης', 'γιατι'), ('γιατι', 'εν'), ('εν', 'της'), ('της', 'το'), ('το', 'λαλεις'), ('λαλεις', 'δηλ'), ('δηλ', 'τζιαι'), ('τζιαι', 'λαλεις'), ('λαλεις', 'μας'), ('μας', 'το'), ('το', 'εμας')], [('ινταλως', 'εγερασεν'), ('εγερασεν', 'ετσι'), ('ετσι', 'τζεινη')], [('ατε', 'ελα'), ('ελα', 'λαρνακα'), ('λαρνακα', 'τζαι'), ('τζαι', 'καρτερω'), ('καρτερω', 'σε')], [('ππεσε', 'τζοιμηθου'), ('τζοιμηθου', 'ξερεις'), ('ξερεις', 'ποσα'), ('ποσα', 'πραματα'), ('πραματα', 'εκαμα'), ('εκαμα', 'εγω'), ('εγω', 'που'), ('που', 'τζεινη'), ('τζεινη', 'την'), ('την', 'ωρα')], [('εσιει', 'σκονη'), ('σκονη', 'οξα'), ('οξα', 'σκουπιζει'), ('σκουπιζει', 'η'), ('η', 'γειτονισσα'), ('γειτονισσα', 'παλε')], [('φανταζουμαι', 'εκοπηκε'), ('εκοπηκε', 'η'), ('η', 'ορεξη'), ('ορεξη', 'σου'), ('σου', 'τζεινη'), ('τζεινη', 'την'), ('την', 'νυχτα'), ('νυχτα', 'εννε')], [('ελα', 'ρε'), ('ρε', 'πελλε'), ('πελλε', 'ηντα'), ('ηντα', 'μπου'), ('μπου', 'λαλουσιν'), ('λαλουσιν', 'οι'), ('οι', 'κορουες'), ('κορουες', 'τζιαμε')], [('καρτερω', 'τουντες'), ('τουντες', 'καλυτερες'), ('καλυτερες', 'μερες'), ('μερες', 'που'), ('που', 'λαλουσιν'), ('λαλουσιν', 'ουλλοι'), ('ουλλοι', 'οτι'), ('οτι', 'εν'), ('εν', 'να'), ('να', 'ρτουν')], [('αμπα', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'ειμαι'), ('ειμαι', 'εγιω')], [('αμπα', 'τζαι'), ('τζαι', 'δειτε'), ('δειτε', 'πλασμα'), ('πλασμα', 'να'), ('να', 'υποστηριζει'), ('υποστηριζει', 'τες'), ('τες', 'γεναιτζιες'), ('γεναιτζιες', 'εσεις'), ('εσεις', 'οι'), ('οι', 'θκυο')], [('που', 'την'), ('την', 'ωρα'), ('ωρα', 'που'), ('που', 'το'), ('το', 'εθκιαβασα'), ('εθκιαβασα', 'πονω'), ('πονω', 'το'), ('το', 'στομαχι'), ('στομαχι', 'μου')], [('φατε', 'τζιαι'), ('τζιαι', 'ποψε'), ('ποψε', 'διοτι'), ('διοτι', 'που'), ('που', 'δευτερας'), ('δευτερας', 'εν'), ('εν', 'εσσιει'), ('εσσιει', 'τιποτες'), ('τιποτες', 'πιον')], [('οι', 'εν'), ('εν', 'εσσιει'), ('εσσιει', 'προβλημα'), ('προβλημα', 'ο'), ('ο', 'φιλος'), ('φιλος', 'μου'), ('μου', 'εξαναειδε'), ('εξαναειδε', 'τα'), ('τα', 'ουλα')], [('αληθκεια', 'εσσιει'), ('εσσιει', 'κανενα'), ('κανενα', 'που'), ('που', 'τα'), ('τα', 'γραφει'), ('γραφει', 'τουτα')], [('που', 'εισαστεν'), ('εισαστεν', 'οι'), ('οι', 'εξυπνοι'), ('εξυπνοι', 'τζιαι'), ('τζιαι', 'εσσιετε'), ('εσσιετε', 'τις'), ('τις', 'πιο'), ('πιο', 'ωραιες'), ('ωραιες', 'ιδεες'), ('ιδεες', 'να'), ('να', 'εβρετε'), ('εβρετε', 'λυσεις')], [('εν', 'τζαι'), ('τζαι', 'βαλω'), ('βαλω', 'το'), ('το', 'σιερι'), ('σιερι', 'μου'), ('μου', 'στη'), ('στη', 'φωθκια')], [('η', 'μοναδικη'), ('μοναδικη', 'κυπραια'), ('κυπραια', 'που'), ('που', 'της'), ('της', 'αρεσκει'), ('αρεσκει', 'να'), ('να', 'καμνει'), ('καμνει', 'τες'), ('τες', 'δουλειες'), ('δουλειες', 'του'), ('του', 'σπιθκιου')], [('ειναι', 'για'), ('για', 'να'), ('να', 'φακκας'), ('φακκας', 'τη'), ('τη', 'κκελλε'), ('κκελλε', 'σ'), ('σ', 'δεν'), ('δεν', 'τ'), ('τ', 'αρεσκει'), ('αρεσκει', 'να'), ('να', 'παιζει')], [('αρεσκει', 'μου'), ('μου', 'που'), ('που', 'εν'), ('εν', 'προσπαθει'), ('προσπαθει', 'να'), ('να', 'καλαμαρισει'), ('καλαμαρισει', 'ο'), ('ο', 'παμπος')], [('αρεσκει', 'μου'), ('μου', 'πολλα'), ('πολλα', 'τουτος'), ('τουτος', 'εθθα'), ('εθθα', 'το'), ('το', 'ελεα'), ('ελεα', 'δαμαι')], [('ο', 'κυριος'), ('κυριος', 'που'), ('που', 'καλαμαριζει'), ('καλαμαριζει', 'ακομα'), ('ακομα', 'εν'), ('εν', 'εφκηκε')], [('μια', 'λευκωσιατισσα'), ('λευκωσιατισσα', 'που'), ('που', 'καλαμαριζει'), ('καλαμαριζει', 'οτι'), ('οτι', 'χειροτερο'), ('χειροτερο', 'χειροτερη'), ('χειροτερη', 'απο'), ('απο', 'μια'), ('μια', 'παφιτισσα'), ('παφιτισσα', 'που'), ('που', 'καλαμαριζει')], [('ιντα', 'καλαμαριζει'), ('καλαμαριζει', 'τουτος')], [('δεν', 'το'), ('το', 'κανουμε'), ('κανουμε', 'τοσο'), ('τοσο', 'πολλα'), ('πολλα', 'η'), ('η', 'κυπραια'), ('κυπραια', 'που'), ('που', 'προσπαθει'), ('προσπαθει', 'να'), ('να', 'καλαμαρισει'), ('καλαμαρισει', 'τζε'), ('τζε', 'φεφκει'), ('φεφκει', 'της'), ('της', 'το'), ('το', 'κυπριακο')], [('καμνει', 'λλια'), ('λλια', 'τσιαλιμουθκια'), ('τσιαλιμουθκια', 'παραπανω'), ('παραπανω', 'που'), ('που', 'το'), ('το', 'κανονικο'), ('κανονικο', 'αλλα'), ('αλλα', 'εν'), ('εν', 'μωρο'), ('μωρο', 'πιστευκω'), ('πιστευκω', 'θα'), ('θα', 'σασει')], [('θελω', 'να'), ('να', 'πιστευκω'), ('πιστευκω', 'οτι'), ('οτι', 'τουτες'), ('τουτες', 'εν'), ('εν', 'οι'), ('οι', 'πιο'), ('πιο', 'φτανες'), ('φτανες', 'του'), ('του', 'νησιου'), ('νησιου', 'μας'), ('μας', 'οχι'), ('οχι', 'οι'), ('οι', 'πιο'), ('πιο', 'ομορφες')], [('εσιει', 'χαζιν'), ('χαζιν', 'ο'), ('ο', 'τυπος'), ('τυπος', 'λαλει'), ('λαλει', 'τζιαι'), ('τζιαι', 'καμποσες'), ('καμποσες', 'αληθκειες')], [('στην', 'κυπρο'), ('κυπρο', 'αμα'), ('αμα', 'πεις'), ('πεις', 'οτι'), ('οτι', 'καμνεις'), ('καμνεις', 'εθελοντισμο'), ('εθελοντισμο', 'ρωτουν'), ('ρωτουν', 'σε'), ('σε', 'αν'), ('αν', 'εισαι'), ('εισαι', 'ανεργος'), ('ανεργος', 'η'), ('η', 'παρα'), ('παρα', 'πολλα'), ('πολλα', 'αθκειασερος')], [('εσχιει', 'κανενα'), ('κανενα', 'που'), ('που', 'συγγενεις'), ('συγγενεις', 'του'), ('του', 'επηε'), ('επηε', 'οξα'), ('οξα', 'καμνεις'), ('καμνεις', 'οτι'), ('οτι', 'δεν'), ('δεν', 'τους'), ('τους', 'ξερεις'), ('ξερεις', 'μετα'), ('μετα', 'που'), ('που', 'τουτο')], [('το', 'θεμα'), ('θεμα', 'που'), ('που', 'βαλλεις'), ('βαλλεις', 'εσυζητηθηκε'), ('εσυζητηθηκε', 'δαμαι')], [('μανα', 'μου'), ('μου', 'διουν'), ('διουν', 'μου'), ('μου', 'τες'), ('τες', 'ερωτησεις'), ('ερωτησεις', 'ξερεις'), ('ξερεις', 'ποσα'), ('ποσα', 'πιανω'), ('πιανω', 'για'), ('για', 'να'), ('να', 'δουλεφκω'), ('δουλεφκω', 'δαμαι')], [('εγνωρισα', 'μιαν'), ('μιαν', 'εχτες'), ('εχτες', 'κουκλαρα'), ('κουκλαρα', 'αλλα'), ('αλλα', 'προβληματιζει'), ('προβληματιζει', 'με'), ('με', 'το'), ('το', 'υψος'), ('υψος', 'της'), ('της', 'εν'), ('εν', 'τοσο'), ('τοσο', 'κοντη'), ('κοντη', 'που'), ('που', 'αν'), ('αν', 'ππεσει'), ('ππεσει', 'αφκο'), ('αφκο', 'που'), ('που', 'τον'), ('τον', 'κωλο'), ('κωλο', 'της'), ('της', 'εθθα'), ('εθθα', 'σπασει')], [('εθθα', 'πιστεψετε'), ('πιστεψετε', 'ποια'), ('ποια', 'χωρκανη'), ('χωρκανη', 'εκαμεν'), ('εκαμεν', 'αιτηση')], [('βαρκουμαι', 'ο'), ('ο', 'αδρωπος'), ('αδρωπος', 'εθθα'), ('εθθα', 'τη'), ('τη', 'φκαλει'), ('φκαλει', 'σκαρτη')], [('πε', 'μας'), ('μας', 'τζαι'), ('τζαι', 'βαρκουμαι'), ('βαρκουμαι', 'να'), ('να', 'θκιεβασω')], [('περκει', 'τους'), ('τους', 'κουντησουν'), ('κουντησουν', 'παλε'), ('παλε', 'δεν'), ('δεν', 'θα'), ('θα', 'προλαβουν'), ('προλαβουν', 'τζιαμε'), ('τζιαμε', 'που'), ('που', 'θα'), ('θα', 'τους'), ('τους', 'κοψουμε')], [('επηεν', 'τζιαμαι'), ('τζιαμαι', 'εκατσεν'), ('εκατσεν', 'εναν'), ('εναν', 'τεταρτο')], [('ναι', 'εψες'), ('εψες', 'τα'), ('τα', 'μεσανυχτα'), ('μεσανυχτα', 'αρα'), ('αρα', 'σημερα'), ('σημερα', 'αρκεψαμεν'), ('αρκεψαμεν', 'ιδιαιτερα'), ('ιδιαιτερα', 'περιμενω'), ('περιμενω', 'δαμε'), ('δαμε', 'να'), ('να', 'τελειωσει')], [('γιαννο', 'μα'), ('μα', 'εσιει'), ('εσιει', 'σουβλες'), ('σουβλες', 'δεν'), ('δεν', 'νηστευκω'), ('νηστευκω', 'πλεον')], [('τζε', 'ιντα'), ('ιντα', 'που'), ('που', 'να'), ('να', 'παει'), ('παει', 'να'), ('να', 'κανει'), ('κανει', 'δαμε')], [('ε', 'οι'), ('οι', 'κορη'), ('κορη', 'ιντα'), ('ιντα', 'που'), ('που', 'παθες')], [('αρεσκει', 'μου'), ('μου', 'που'), ('που', 'την'), ('την', 'αρωτα'), ('αρωτα', 'ιντα'), ('ιντα', 'που'), ('που', 'εκαμνε'), ('εκαμνε', 'και'), ('και', 'εκλεισε'), ('εκλεισε', 'ο'), ('ο', 'λαιμος'), ('λαιμος', 'της')], [('οι', 'ρε'), ('ρε', 'τζαι'), ('τζαι', 'εσουνη'), ('εσουνη', 'ετσι'), ('ετσι', 'πραγμα'), ('πραγμα', 'μα'), ('μα', 'ιντα'), ('ιντα', 'που'), ('που', 'μας'), ('μας', 'λαλεις'), ('λαλεις', 'ρε'), ('ρε', 'κουμπαρε'), ('κουμπαρε', 'μου')], [('ιντα', 'που'), ('που', 'καμνεις'), ('καμνεις', 'φιλουι'), ('φιλουι', 'μου')], [('τελικα', 'ηντα'), ('ηντα', 'που'), ('που', 'εν'), ('εν', 'γινει'), ('γινει', 'ολαν'), ('ολαν', 'εννα'), ('εννα', 'φαμεν'), ('φαμεν', 'οξα'), ('οξα', 'εννα'), ('εννα', 'μεινουμε'), ('μεινουμε', 'νηστιτζιοι')], [('μα', 'ηντα'), ('ηντα', 'που'), ('που', 'καμνουν'), ('καμνουν', 'σιορ'), ('σιορ', 'στο'), ('στο', 'πανεπιστημιον'), ('πανεπιστημιον', 'εν'), ('εν', 'μες'), ('μες', 'το'), ('το', 'φακελλουιν'), ('φακελλουιν', 'που'), ('που', 'πκιανουν'), ('πκιανουν', 'τα'), ('τα', 'δοκτορατα')], [('η', 'αληθκεια'), ('αληθκεια', 'ενει'), ('ενει', 'αμαν'), ('αμαν', 'στη'), ('στη', 'κυπρο'), ('κυπρο', 'εν'), ('εν', 'φαντασμαγορικα'), ('φαντασμαγορικα', 'στο'), ('στο', 'μανχατταν'), ('μανχατταν', 'τζιαι'), ('τζιαι', 'στες'), ('στες', 'αγορες'), ('αγορες', 'τις'), ('τις', 'ευρωπαικες'), ('ευρωπαικες', 'ηντα'), ('ηντα', 'που'), ('που', 'ενει')], [('εν', 'μου'), ('μου', 'αρεσκουν'), ('αρεσκουν', 'οι'), ('οι', 'σσιεφταλιες')], [('τωρα', 'πλεον'), ('πλεον', 'εν'), ('εν', 'βαζεις'), ('βαζεις', 'για'), ('για', 'κανενα'), ('κανενα', 'το'), ('το', 'σσιερι'), ('σσιερι', 'στη'), ('στη', 'φωθκια')], [('δαμε', 'ηβρε'), ('ηβρε', 'μας'), ('μας', 'φαση'), ('φαση', 'που'), ('που', 'καμνει'), ('καμνει', 'σσιερι'), ('σσιερι', 'ο'), ('ο', 'παικτης'), ('παικτης', 'τους'), ('τους', 'τζαι'), ('τζαι', 'ζητα'), ('ζητα', 'τη'), ('τη', 'γνωμη'), ('γνωμη', 'μας'), ('μας', 'αν'), ('αν', 'ηταν'), ('ηταν', 'πεναλτυ')], [('σικκιμε', 'ολαν'), ('ολαν', 'διω'), ('διω', 'τζαι'), ('τζαι', 'γω')], [('τωρα', 'εν'), ('εν', 'τους'), ('τους', 'ταιρκαζει'), ('ταιρκαζει', 'ενα'), ('ενα', 'σκασε'), ('σκασε', 'σου'), ('σου', 'ρα')], [('δαμε', 'ταιρκαζει'), ('ταιρκαζει', 'η'), ('η', 'κουβεντα'), ('κουβεντα', 'μονον'), ('μονον', 'τους'), ('τους', 'πελλους'), ('πελλους', 'δεν'), ('δεν', 'φωρει'), ('φωρει', 'ο'), ('ο', 'τοπος')], [('κοπελια', 'η'), ('η', 'αριθμητικη'), ('αριθμητικη', 'μ'), ('μ', 'εν'), ('εν', 'μια'), ('μια', 'χαρα'), ('χαρα', 'η'), ('η', 'μακαριτισσα'), ('μακαριτισσα', 'η'), ('η', 'στετε'), ('στετε', 'μ'), ('μ', 'ηταν'), ('ηταν', 'δασκαλα'), ('δασκαλα', 'ολαν')], [('οποιος', 'εν'), ('εν', 'εσιει'), ('εσιει', 'νουν'), ('νουν', 'εσιει'), ('εσιει', 'ποθκια'), ('ποθκια', 'ελαλεν'), ('ελαλεν', 'η'), ('η', 'στετε'), ('στετε', 'μου')], [('εμεναν', 'η'), ('η', 'στετε'), ('στετε', 'μου'), ('μου', 'εβουραν'), ('εβουραν', 'με'), ('με', 'μες'), ('μες', 'τες'), ('τες', 'αυλαες'), ('αυλαες', 'να'), ('να', 'φαω'), ('φαω', 'αυκα'), ('αυκα', 'που'), ('που', 'τες'), ('τες', 'ορνιθες'), ('ορνιθες', 'μας')], [('ηνταλως', 'καμνεις'), ('καμνεις', 'για'), ('για', 'λλιην'), ('λλιην', 'σκονη')], [('δαμε', 'λαλεις'), ('λαλεις', 'τους'), ('τους', 'ινταλως'), ('ινταλως', 'τους'), ('τους', 'εφκαλλαν'), ('εφκαλλαν', 'που'), ('που', 'τα'), ('τα', 'σπιτια'), ('σπιτια', 'τους'), ('τους', 'οι'), ('οι', 'δικοι'), ('δικοι', 'μας'), ('μας', 'τζιαι'), ('τζιαι', 'λαλουν'), ('λαλουν', 'σου'), ('σου', 'εν'), ('εν', 'μονοι'), ('μονοι', 'τους'), ('τους', 'που'), ('που', 'εφυαν')], [('ως', 'τζιαι'), ('τζιαι', 'του'), ('του', 'μιτση'), ('μιτση', 'φορεις'), ('φορεις', 'του'), ('του', 'μασκα')], [('τουτην', 'την'), ('την', 'κουβεντα'), ('κουβεντα', 'ακουα'), ('ακουα', 'πολλα'), ('πολλα', 'που'), ('που', 'ημουν'), ('ημουν', 'μιτσια')], [('κορη', 'μιτσια'), ('μιτσια', 'ηνταλως'), ('ηνταλως', 'εβρεθηκε')], [('αφου', 'εν'), ('εν', 'τζοιματε'), ('τζοιματε', 'ρε'), ('ρε', 'κοπελια'), ('κοπελια', 'καθε'), ('καθε', 'δκυο'), ('δκυο', 'ωρες'), ('ωρες', 'η'), ('η', 'μιτσια'), ('μιτσια', 'εγερτηριο')], [('εισχιεν', 'τα'), ('τα', 'ουλλα'), ('ουλλα', 'η'), ('η', 'μαρουλου')], [('α', 'μανα'), ('μανα', 'μου'), ('μου', 'ιντα'), ('ιντα', 'αππαρα')], [('ελεος', 'πκιον'), ('πκιον', 'το'), ('το', 'παναυρι'), ('παναυρι', 'του'), ('του', 'χωρκου'), ('χωρκου', 'εν'), ('εν', 'καλλυτερον'), ('καλλυτερον', 'που'), ('που', 'τουτα'), ('τουτα', 'που'), ('που', 'θωρουμε'), ('θωρουμε', 'ποψε')], [('να', 'τρως'), ('τρως', 'μονον'), ('μονον', 'του'), ('του', 'ψουμι'), ('ψουμι', 'απ'), ('απ', 'εφκιασαν'), ('εφκιασαν', 'οι'), ('οι', 'φτουχοι')], [('με', 'το'), ('το', 'ιδιο'), ('ιδιο', 'αλευρι'), ('αλευρι', 'τζιαι'), ('τζιαι', 'το'), ('το', 'ιδιο'), ('ιδιο', 'νερο'), ('νερο', 'το'), ('το', 'ιδιο'), ('ιδιο', 'ψουμι'), ('ψουμι', 'εννα'), ('εννα', 'φκαλεις')], [('εξυπνησα', 'πιο'), ('πιο', 'πελλος'), ('πελλος', 'που'), ('που', 'ποττε')], [('τουτο', 'εν'), ('εν', 'θα'), ('θα', 'αλλαξει'), ('αλλαξει', 'ποττε')], [('το', 'μουχτιν'), ('μουχτιν', 'εν'), ('εν', 'το'), ('το', 'καλυττερο'), ('καλυττερο', 'σιορ')], [('κοπελια', 'τζαμε'), ('τζαμε', 'στη'), ('στη', 'εταιρεια'), ('εταιρεια', 'οι'), ('οι', 'μιτσιοι'), ('μιτσιοι', 'που'), ('που', 'γραφουντε'), ('γραφουντε', 'στες'), ('στες', 'ακαδημιες'), ('ακαδημιες', 'πιαννουν'), ('πιαννουν', 'τζαι'), ('τζαι', 'σεζον'), ('σεζον', 'μουχτιν'), ('μουχτιν', 'εννεν')], [('αμπα', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'ειμαι'), ('ειμαι', 'εγιω')], [('εν', 'για'), ('για', 'τζεινα'), ('τζεινα', 'που'), ('που', 'λαλω')], [('τον', 'τζαιρο'), ('τζαιρο', 'που'), ('που', 'εγιω'), ('εγιω', 'επηεννα'), ('επηεννα', 'εσου'), ('εσου', 'ερκεσουν')], [('μα', 'ιντα'), ('ιντα', 'χωρκον'), ('χωρκον', 'εν'), ('εν', 'τουτο')], [('πιο', 'χωρκαθκιον'), ('χωρκαθκιον', 'πεθανισκεις')], [('ηντα', 'εσου'), ('εσου', 'εν'), ('εν', 'ερκεσαι'), ('ερκεσαι', 'ρε'), ('ρε', 'σιαχουρη'), ('σιαχουρη', 'π'), ('π', 'εν'), ('εν', 'τζιαι'), ('τζιαι', 'διπλα'), ('διπλα', 'σ'), ('σ', 'οποταν'), ('οποταν', 'μεν'), ('μεν', 'κατσιαριζεις'), ('κατσιαριζεις', 'ολαν')], [('τοσα', 'λαλουσιν'), ('λαλουσιν', 'και'), ('και', 'ουτε'), ('ουτε', 'ενα'), ('ενα', 'σκασε'), ('σκασε', 'εσουνι'), ('εσουνι', 'ρα')], [('επουλησεν', 'το'), ('το', 'αυτοκινητο'), ('αυτοκινητο', 'οξα'), ('οξα', 'ακομα')], [('οι', 'οτι'), ('οτι', 'εσιει'), ('εσιει', 'σκονη'), ('σκονη', 'απλα'), ('απλα', 'αν'), ('αν', 'βουττησω'), ('βουττησω', 'τη'), ('τη', 'κκελλε'), ('κκελλε', 'μου'), ('μου', 'μες'), ('μες', 'το'), ('το', 'χωμα'), ('χωμα', 'τζιαι'), ('τζιαι', 'χωσω'), ('χωσω', 'την'), ('την', 'εν'), ('εν', 'να'), ('να', 'αναπνεω'), ('αναπνεω', 'καλλυτερα')], [('εγεμωσεν', 'η'), ('η', 'κκελλε'), ('κκελλε', 'της'), ('της', 'με'), ('με', 'αεριο'), ('αεριο', 'επειδη'), ('επειδη', 'ηρτεν')], [('εν', 'καταλαβαινω'), ('καταλαβαινω', 'τι'), ('τι', 'μου'), ('μου', 'λαλεις'), ('λαλεις', 'τωρα'), ('τωρα', 'θες'), ('θες', 'να'), ('να', 'το'), ('το', 'καμνουμε'), ('καμνουμε', 'η'), ('η', 'οι')], [('τουτο', 'πραμα'), ('πραμα', 'εν'), ('εν', 'σολα'), ('σολα', 'παπουτσιου'), ('παπουτσιου', 'εμποτισμενη'), ('εμποτισμενη', 'με'), ('με', 'λια'), ('λια', 'κομμαθκια'), ('κομμαθκια', 'απο'), ('απο', 'κωλο'), ('κωλο', 'κοτοπουλου')], [('τι', 'βλεπουν'), ('βλεπουν', 'τα'), ('τα', 'ματουθκια'), ('ματουθκια', 'μας'), ('μας', 'να'), ('να', 'φκουμε'), ('φκουμε', 'τελικα')], [('τεσσερα', 'τζαι'), ('τζαι', 'τεσσερα'), ('τεσσερα', 'καμνουσιν'), ('καμνουσιν', 'οκτω'), ('οκτω', 'τεσσερα'), ('τεσσερα', 'παλικαρκια'), ('παλικαρκια', 'πασι'), ('πασι', 'στον'), ('στον', 'πολεμο'), ('πολεμο', 'αλλα'), ('αλλα', 'εχουσιν'), ('εχουσιν', 'τζαι'), ('τζαι', 'στολον')], [('οι', 'χοχοι'), ('χοχοι', 'ουλλοι'), ('ουλλοι', 'εγιναν'), ('εγιναν', 'φανατικοι'), ('φανατικοι', 'ρουμανοι'), ('ρουμανοι', 'ονειρα'), ('ονειρα', 'καμνουσιν'), ('καμνουσιν', 'τρελλα')], [('απο', 'εν'), ('εν', 'εξω'), ('εξω', 'του'), ('του', 'χορου'), ('χορου', 'πολλα'), ('πολλα', 'τραουθκια'), ('τραουθκια', 'ξερει')], [('τουτα', 'τα'), ('τα', 'τραουθκια'), ('τραουθκια', 'λαλει'), ('λαλει', 'της'), ('της', 'μανας'), ('μανας', 'της'), ('της', 'και'), ('και', 'αρεσαν'), ('αρεσαν', 'της')], [('τουτο', 'που'), ('που', 'φκαινουν'), ('φκαινουν', 'τραουδιστες'), ('τραουδιστες', 'που'), ('που', 'εν'), ('εν', 'τους'), ('τους', 'ηξερε'), ('ηξερε', 'η'), ('η', 'μανα'), ('μανα', 'τους'), ('τους', 'ξεπερνα'), ('ξεπερνα', 'με')], [('τρεις', 'ελιες'), ('ελιες', 'τζιε'), ('τζιε', 'μιαν'), ('μιαν', 'ντοματα'), ('ντοματα', 'αγαπω'), ('αγαπω', 'μιαν'), ('μιαν', 'μαυρομματα')], [('ετσι', 'κοπελλαν'), ('κοπελλαν', 'μπορει'), ('μπορει', 'τζαι'), ('τζαι', 'εγω'), ('εγω', 'να'), ('να', 'την'), ('την', 'ερωτευτω')], [('που', 'εν'), ('εν', 'η'), ('η', 'λουλουδου'), ('λουλουδου', 'σιορ'), ('σιορ', 'εσσιει'), ('εσσιει', 'μιση'), ('μιση', 'ωρα'), ('ωρα', 'που'), ('που', 'την'), ('την', 'γυρευκω'), ('γυρευκω', 'να'), ('να', 'συρω'), ('συρω', 'κανενα'), ('κανενα', 'γαρυφαλλο')], [('γυρευκω', 'τη'), ('τη', 'φατσουα'), ('φατσουα', 'με'), ('με', 'το'), ('το', 'ππουρτου'), ('ππουρτου', 'δασκαλε')], [('εν', 'να'), ('να', 'γυρευκω'), ('γυρευκω', 'να'), ('να', 'εβρω'), ('εβρω', 'την'), ('την', 'καμερα'), ('καμερα', 'για'), ('για', 'αορατο'), ('αορατο', 'φακο')], [('αμα', 'παουριζετε'), ('παουριζετε', 'στο'), ('στο', 'τηλεφωνο'), ('τηλεφωνο', 'οτι'), ('οτι', 'θελετε'), ('θελετε', 'ραντεβου'), ('ραντεβου', 'εδω'), ('εδω', 'και'), ('και', 'τωρα'), ('τωρα', 'εν'), ('εν', 'μου'), ('μου', 'διατε'), ('διατε', 'κινητρο'), ('κινητρο', 'να'), ('να', 'εβρω'), ('εβρω', 'ωρα')], [('θελω', 'να'), ('να', 'φυω'), ('φυω', 'να'), ('να', 'εβρω'), ('εβρω', 'μια'), ('μια', 'σπηλια'), ('σπηλια', 'να'), ('να', 'μπω'), ('μπω', 'τζιει'), ('τζιει', 'μεσα'), ('μεσα', 'να'), ('να', 'ζω'), ('ζω', 'με'), ('με', 'τες'), ('τες', 'αρκουδες'), ('αρκουδες', 'με'), ('με', 'τα'), ('τα', 'σιονια')], [('τα', 'ιδια'), ('ιδια', 'ελαλουν'), ('ελαλουν', 'τοτε'), ('τοτε', 'εσιει'), ('εσιει', 'τροπο'), ('τροπο', 'για'), ('για', 'να'), ('να', 'τα'), ('τα', 'εβρω'), ('εβρω', 'ευκολα')], [('ε', 'γεναικα'), ('γεναικα', 'αυριον'), ('αυριον', 'θα'), ('θα', 'εσιει'), ('εσιει', 'σιουρα'), ('σιουρα', 'κουπεπια')], [('που', 'επιττωσα'), ('επιττωσα', 'εφαα'), ('εφαα', 'ενα'), ('ενα', 'σινι'), ('σινι', 'μπουγατσα'), ('μπουγατσα', 'τζαι'), ('τζαι', 'πουπανω'), ('πουπανω', 'κουπεπια'), ('κουπεπια', 'εν'), ('εν', 'ηταν'), ('ηταν', 'τοσο'), ('τοσο', 'νεκατομενα')], [('ο', 'λοος'), ('λοος', 'του'), ('του', 'αδρωπου'), ('αδρωπου', 'ο,τι'), ('ο,τι', 'τζιαι'), ('τζιαι', 'αν'), ('αν', 'ενει')], [('να', 'σας'), ('σας', 'παρακαλεσω'), ('παρακαλεσω', 'χωρις'), ('χωρις', 'να'), ('να', 'θελω'), ('θελω', 'να'), ('να', 'σας'), ('σας', 'προσβαλω'), ('προσβαλω', 'να'), ('να', 'μεν'), ('μεν', 'μου'), ('μου', 'ξανα'), ('ξανα', 'στειλετε'), ('στειλετε', 'οτι'), ('οτι', 'εσιη'), ('εσιη', 'σχεση'), ('σχεση', 'με'), ('με', 'μηνυματα'), ('μηνυματα', 'που'), ('που', 'πρεπει'), ('πρεπει', 'να'), ('να', 'καμω'), ('καμω', 'κοπυ'), ('κοπυ', 'σε'), ('σε', 'φιλους'), ('φιλους', 'ειδαλλως'), ('ειδαλλως', 'εννα'), ('εννα', 'χασω'), ('χασω', 'την'), ('την', 'τυχη'), ('τυχη', 'που'), ('που', 'τρεσιη'), ('τρεσιη', 'που'), ('που', 'τα'), ('τα', 'πουναρκα'), ('πουναρκα', 'μου')], [('ποττε', 'εν'), ('εν', 'ημουν'), ('ημουν', 'τυχερος'), ('τυχερος', 'αντιθετως'), ('αντιθετως', 'ειμαι'), ('ειμαι', 'τοσο'), ('τοσο', 'καχτος'), ('καχτος', 'τις'), ('τις', 'πλειστες'), ('πλειστες', 'φορες'), ('φορες', 'που'), ('που', 'εσκεφτηκα'), ('εσκεφτηκα', 'να'), ('να', 'χτυπησω'), ('χτυπησω', 'τον'), ('τον', 'καχτο'), ('καχτο', 'σε'), ('σε', 'ταττοο')], [('ειμαι', 'σιουρος'), ('σιουρος', 'οτι'), ('οτι', 'αν'), ('αν', 'δεν'), ('δεν', 'ενοχλησω'), ('ενοχλησω', 'ατομα'), ('ατομα', 'με'), ('με', 'τα'), ('τα', 'μηνυματα'), ('μηνυματα', 'περι'), ('περι', 'θεων'), ('θεων', 'αγγελων'), ('αγγελων', 'και'), ('και', 'τυχης'), ('τυχης', 'ο'), ('ο', 'θεος'), ('θεος', 'εν'), ('εν', 'θα'), ('θα', 'με'), ('με', 'τιμωρησει'), ('τιμωρησει', 'ουτε'), ('ουτε', 'εννα'), ('εννα', 'χασω'), ('χασω', 'το'), ('το', 'θαυμα'), ('θαυμα', 'που'), ('που', 'θα'), ('θα', 'εγινετουν'), ('εγινετουν', 'σε'), ('σε', 'αντιθετη'), ('αντιθετη', 'περιπτωση')], [('αντι', 'να'), ('να', 'μου'), ('μου', 'ξανα'), ('ξανα', 'στειλετε'), ('στειλετε', 'οτιδηποτε'), ('οτιδηποτε', 'αφορα'), ('αφορα', 'μπορειτε'), ('μπορειτε', 'να'), ('να', 'μου'), ('μου', 'πειτε'), ('πειτε', 'ενα'), ('ενα', 'γεια'), ('γεια', 'σου'), ('σου', 'ενα'), ('ενα', 'τι'), ('τι', 'καμνεις'), ('καμνεις', 'χωρις'), ('χωρις', 'να'), ('να', 'υπαρξουν'), ('υπαρξουν', 'στην'), ('στην', 'συνομιλια'), ('συνομιλια', 'μας'), ('μας', 'μακρυσκελες'), ('μακρυσκελες', 'ανουσια'), ('ανουσια', 'μηνυματα'), ('μηνυματα', 'τζιαι'), ('τζιαι', 'εικονες'), ('εικονες', 'που'), ('που', 'αναβοσβηννουν'), ('αναβοσβηννουν', 'οπως'), ('οπως', 'τζινες'), ('τζινες', 'που'), ('που', 'βαλουν'), ('βαλουν', 'οι'), ('οι', 'θκιαες'), ('θκιαες', 'μου'), ('μου', 'τα'), ('τα', 'χριστουγεννα'), ('χριστουγεννα', 'τζιαι'), ('τζιαι', 'παθαινω'), ('παθαινω', 'ενα'), ('ενα', 'ειδος'), ('ειδος', 'επιληψιας')], [('εν', 'καλο'), ('καλο', 'εν'), ('εν', 'το'), ('το', 'εκατεβασα'), ('εκατεβασα', 'ακομα')], [('εν', 'εσιει'), ('εσιει', 'ακομα'), ('ακομα', 'φαινεται')], [('θα', 'ρωτησω'), ('ρωτησω', 'κανενα'), ('κανενα', 'πληροφορηκαριο'), ('πληροφορηκαριο', 'να'), ('να', 'δω'), ('δω', 'τι'), ('τι', 'εννα'), ('εννα', 'μου'), ('μου', 'πει')], [('η', 'δουλεια'), ('δουλεια', 'που'), ('που', 'εγινηκε'), ('εγινηκε', 'εννεν'), ('εννεν', 'μονο'), ('μονο', 'για'), ('για', 'μπυρες')], [('πρεπει', 'να'), ('να', 'φερεις'), ('φερεις', 'τζιαι'), ('τζιαι', 'σουβλακια')], [('χρωστω', 'τα'), ('τα', 'εγω')], [('εννα', 'καμουμε'), ('καμουμε', 'κατι'), ('κατι', 'πυραβλουθκια')], [('βαρκουμαι', 'εκατσα'), ('εκατσα', 'εκαμα'), ('εκαμα', 'ολοκληρο'), ('ολοκληρο', 'καστρο'), ('καστρο', 'που'), ('που', 'τη'), ('τη', 'βαρεμαρα')], [('ως', 'τζιαι'), ('τζιαι', 'εμαθαινα'), ('εμαθαινα', 'τζιαι'), ('τζιαι', 'καμω'), ('καμω', 'τες')], [('εν', 'καθηγητης'), ('καθηγητης', 'ο'), ('ο', 'πελλος')], [('εκαμα', 'μιαν'), ('μιαν', 'ερευνα'), ('ερευνα', 'που'), ('που', 'εδημιουργησα')], [('τζεινο', 'π'), ('π', 'εχουν'), ('εχουν', 'τα'), ('τα', 'τζαινουρκα'), ('τζαινουρκα', 'τζιαι'), ('τζιαι', 'να'), ('να', 'αλλαζουν'), ('αλλαζουν', 'τραουθκια'), ('τραουθκια', 'με'), ('με', 'κινητο'), ('κινητο', 'αρεσαν'), ('αρεσαν', 'μ')], [('α', 'ρε'), ('ρε', 'κοπελια'), ('κοπελια', 'αντρικκο'), ('αντρικκο', 'μ'), ('μ', 'μολις'), ('μολις', 'ερτεις'), ('ερτεις', 'να'), ('να', 'τους'), ('τους', 'καμουμε'), ('καμουμε', 'παρτυ'), ('παρτυ', 'εσσω'), ('εσσω', 'μ')], [('κατσε', 'θκιαβασε'), ('θκιαβασε', 'τζι'), ('τζι', 'ελα'), ('ελα', 'ελλαδα')], [('κατι', 'τρελλο'), ('τρελλο', 'ετοιμαζει'), ('ετοιμαζει', 'ο'), ('ο', 'αντρικκος'), ('αντρικκος', 'παλε')], [('σε', 'ουλλα'), ('ουλλα', 'τα'), ('τα', 'κοπελια'), ('κοπελια', 'που'), ('που', 'μου'), ('μου', 'ευχηθηκασιν'), ('ευχηθηκασιν', 'σημερα'), ('σημερα', 'λαλω'), ('λαλω', 'τους'), ('τους', 'ευχαριστω'), ('ευχαριστω', 'πολλα')], [('γιατι', 'με'), ('με', 'αφηκετε'), ('αφηκετε', 'μονη'), ('μονη', 'μου'), ('μου', 'μαζι'), ('μαζι', 'τους')], [('αν', 'ακουσετε'), ('ακουσετε', 'κανενα'), ('κανενα', 'μες'), ('μες', 'το'), ('το', 'χωρκο'), ('χωρκο', 'μεν'), ('μεν', 'φοηθειτε'), ('φοηθειτε', 'εγιω'), ('εγιω', 'ημουν')], [('περπατα', 'τζιαι'), ('τζιαι', 'λλιο'), ('λλιο', 'να'), ('να', 'καμεις'), ('καμεις', 'τα'), ('τα', 'αφκα')], [('ποσο', 'σιεσμενοι'), ('σιεσμενοι', 'ειμαστε'), ('ειμαστε', 'που'), ('που', 'τον'), ('τον', 'κοσμο'), ('κοσμο', 'ρε'), ('ρε', 'ενα'), ('ενα', 'προτζεκτ'), ('προτζεκτ', 'να'), ('να', 'καμουμε'), ('καμουμε', 'εν'), ('εν', 'γινεται')], [('ουλλοι', 'εχουν'), ('εχουν', 'μας'), ('μας', 'για'), ('για', 'κκιλιτζιρους'), ('κκιλιτζιρους', 'τζιαι'), ('τζιαι', 'εχουν'), ('εχουν', 'δικαιο'), ('δικαιο', 'τζιαι'), ('τζιαι', 'μεις'), ('μεις', 'μαχουμαστε'), ('μαχουμαστε', 'να'), ('να', 'καμουμε'), ('καμουμε', 'τους'), ('τους', 'πολιτισμενους')], [('ατε', 'ολαν'), ('ολαν', 'τζιαι'), ('τζιαι', 'κανει')], [('κορη', 'εν'), ('εν', 'για'), ('για', 'να'), ('να', 'μπαινει'), ('μπαινει', 'μπροστα'), ('μπροστα', 'που'), ('που', 'καθε'), ('καθε', 'βιντεο'), ('βιντεο', 'που'), ('που', 'τωρα'), ('τωρα', 'τζιαι'), ('τζιαι', 'να'), ('να', 'παει')], [('εν', 'τζι'), ('τζι', 'εικονα')], [('εκαμα', 'ενα'), ('ενα', 'καναλουι'), ('καναλουι', 'για'), ('για', 'τες'), ('τες', 'πελλαρες'), ('πελλαρες', 'που'), ('που', 'καμμω'), ('καμμω', 'εσσω'), ('εσσω', 'ουλλη'), ('ουλλη', 'μερα')], [('το', 'λοιπον'), ('λοιπον', 'ετο'), ('ετο', 'δαμε')], [('αλλο', 'λιο'), ('λιο', 'ετσι'), ('ετσι', 'εννα'), ('εννα', 'καμουμεν')], [('ειδαμεν', 'τα'), ('τα', 'ουλλα'), ('ουλλα', 'πιον')], [('οξα', 'κομα')], [('ευχαριστω', 'σε'), ('σε', 'ολους'), ('ολους', 'σας'), ('σας', 'κοπελια'), ('κοπελια', 'τζιαι'), ('τζιαι', 'στα'), ('στα', 'δικα'), ('δικα', 'σας')], [('ρε', 'πεθκια'), ('πεθκια', 'χτιπατε'), ('χτιπατε', 'μ'), ('μ', 'ενας'), ('ενας', 'το'), ('το', 'κινητο'), ('κινητο', 'μ'), ('μ', 'τζι'), ('τζι', 'εχασα'), ('εχασα', 'το')], [('παιθκια', 'εν'), ('εν', 'ξερω'), ('ξερω', 'για'), ('για', 'σας'), ('σας', 'εγω'), ('εγω', 'ουτε'), ('ουτε', 'φλαουνες'), ('φλαουνες', 'εκαμα'), ('εκαμα', 'ουτε'), ('ουτε', 'αφκα'), ('αφκα', 'εβαψα'), ('εβαψα', 'ουτε'), ('ουτε', 'επιταφιο'), ('επιταφιο', 'εκαταφερα'), ('εκαταφερα', 'να'), ('να', 'παω'), ('παω', 'ουτε'), ('ουτε', 'τιποτε')], [('οποταν', 'μετα'), ('μετα', 'το'), ('το', 'αρχικο'), ('αρχικο', 'μισαωρο'), ('μισαωρο', 'που'), ('που', 'εκαμνα'), ('εκαμνα', 'σαν'), ('σαν', 'το'), ('το', 'διχρονο'), ('διχρονο', 'εσηκωστηκα'), ('εσηκωστηκα', 'ετσαππισα'), ('ετσαππισα', 'εφυτεψα'), ('εφυτεψα', 'εξαπλωσα'), ('εξαπλωσα', 'στον'), ('στον', 'ηλιο'), ('ηλιο', 'τζαι'), ('τζαι', 'εθωρουν'), ('εθωρουν', 'τες'), ('τες', 'μελισσες'), ('μελισσες', 'τζαι'), ('τζαι', 'τες'), ('τες', 'πεταλουδες'), ('πεταλουδες', 'που'), ('που', 'πανω'), ('πανω', 'μου'), ('μου', 'ακουσα'), ('ακουσα', 'τα'), ('τα', 'χελιδονια'), ('χελιδονια', 'ετζημηθηκα'), ('ετζημηθηκα', 'εκατσα'), ('εκατσα', 'κατω'), ('κατω', 'που'), ('που', 'τ'), ('τ', 'αστρα'), ('αστρα', 'τζαι'), ('τζαι', 'ηπια'), ('ηπια', 'κρασουι'), ('κρασουι', 'τζαι'), ('τζαι', 'ακουσα'), ('ακουσα', 'τη'), ('τη', 'μουσικη'), ('μουσικη', 'μου'), ('μου', 'ουλλα'), ('ουλλα', 'εν'), ('εν', 'προσκαιρα')], [('καλη', 'καρθκιαν'), ('καρθκιαν', 'καλα'), ('καλα', 'ποτα'), ('ποτα', 'τζαι'), ('τζαι', 'με'), ('με', 'υγεια'), ('υγεια', 'τζαι'), ('τζαι', 'αγαπη'), ('αγαπη', 'στη'), ('στη', 'ζωη'), ('ζωη', 'μας')], [('ποιες', 'εκφρασεις'), ('εκφρασεις', 'φακκουν'), ('φακκουν', 'σου'), ('σου', 'στα'), ('στα', 'νευρα'), ('νευρα', 'ρε'), ('ρε', 'κουμπαρε')], [('ριγος', 'χωρκαθκιου'), ('χωρκαθκιου', 'με'), ('με', 'διαπερασε'), ('διαπερασε', 'ειμαι'), ('ειμαι', 'σιουρη'), ('σιουρη', 'τζ'), ('τζ', 'εσενα')], [('νομιζω', 'οπου'), ('οπου', 'τζαι'), ('τζαι', 'να'), ('να', 'παμε'), ('παμε', 'καπου'), ('καπου', 'ενα'), ('ενα', 'δειξουμε'), ('δειξουμε', 'το'), ('το', 'νου'), ('νου', 'μας')], [('τρεσιει', 'τιποτες'), ('τιποτες', 'λαμνε'), ('λαμνε', 'να'), ('να', 'μεν'), ('μεν', 'νευριασω')], [('αμαν', 'το'), ('το', 'δωρο'), ('δωρο', 'που'), ('που', 'πιανεις'), ('πιανεις', 'του'), ('του', 'παρεα'), ('παρεα', 'σου'), ('σου', 'εν'), ('εν', 'καλλυττερο'), ('καλλυττερο', 'που'), ('που', 'το'), ('το', 'δικο'), ('δικο', 'σου')], [('η', 'περιπαιζει'), ('περιπαιζει', 'μας'), ('μας', 'η'), ('η', 'εσιει'), ('εσιει', 'υπομονη'), ('υπομονη', 'γαδαρου')], [('εκαμα', 'το'), ('το', 'επιτελους'), ('επιτελους', 'ουλλο'), ('ουλλο', 'μαλακιες'), ('μαλακιες', 'λαλειτε'), ('λαλειτε', 'τελευταιως')], [('ο', 'θκειος'), ('θκειος', 'μου'), ('μου', 'ηξερεν'), ('ηξερεν', 'τον'), ('τον', 'θκειον'), ('θκειον', 'του')], [('οι', 'πως'), ('πως', 'εσιει'), ('εσιει', 'σκονη'), ('σκονη', 'σημερα'), ('σημερα', 'αλλα'), ('αλλα', 'μεν'), ('μεν', 'κατσετε'), ('κατσετε', 'εξω'), ('εξω', 'να'), ('να', 'πιειτε'), ('πιειτε', 'τον'), ('τον', 'καφε'), ('καφε', 'σας'), ('σας', 'εννα'), ('εννα', 'εσιει'), ('εσιει', 'λλιον'), ('λλιον', 'γευση'), ('γευση', 'λασπη')], [('ρε', 'παιθκια'), ('παιθκια', 'πναστε'), ('πναστε', 'λλιο'), ('λλιο', 'με'), ('με', 'το'), ('το', 'να'), ('να', 'γραψεις'), ('γραψεις', 'στα'), ('στα', 'σιερκα'), ('σιερκα', 'σου'), ('σου', 'θκυο'), ('θκυο', 'λεξεις'), ('λεξεις', 'ουτε'), ('ουτε', 'καν'), ('καν', 'τρυπα'), ('τρυπα', 'στο'), ('στο', 'νερο'), ('νερο', 'εν'), ('εν', 'καμνεις')], [('ηρτεν', 'η'), ('η', 'τσικνοπεμπτη'), ('τσικνοπεμπτη', 'να'), ('να', 'φαμε'), ('φαμε', 'λλιη'), ('λλιη', 'σαλατα')], [('ειμαι', 'κοπελλουι'), ('κοπελλουι', 'τζαι'), ('τζαι', 'καταλαβαινω'), ('καταλαβαινω', 'πολλα'), ('πολλα', 'καλα'), ('καλα', 'ιντα'), ('ιντα', 'που'), ('που', 'εννοει'), ('εννοει', 'αρα'), ('αρα', 'εγερασα')], [('εν', 'ιξερω'), ('ιξερω', 'παω'), ('παω', 'να'), ('να', 'ππεσω'), ('ππεσω', 'σε'), ('σε', 'κανενα'), ('κανενα', 'λακκο')], [('οι', 'ρε'), ('ρε', 'παιθκια'), ('παιθκια', 'τζαι'), ('τζαι', 'ειμαστεν'), ('ειμαστεν', 'κριμα'), ('κριμα', 'ολαν')], [('πρεπει', 'να'), ('να', 'εσιει'), ('εσιει', 'εναν'), ('εναν', 'να'), ('να', 'κανονισει'), ('κανονισει', 'τον'), ('τον', 'μιτσην')], [('εχαρισαμεν', 'σας'), ('σας', 'τουτην'), ('τουτην', 'την'), ('την', 'στρατα')], [('οι', 'πως'), ('πως', 'εν'), ('εν', 'πυρα'), ('πυρα', 'αλλα'), ('αλλα', 'επηρα'), ('επηρα', 'το'), ('το', 'γουρουνακι'), ('γουρουνακι', 'μου'), ('μου', 'περιπατο'), ('περιπατο', 'τζιαι'), ('τζιαι', 'εφερα'), ('εφερα', 'πισω'), ('πισω', 'σουβλα')], [('παπα', 'μου'), ('μου', 'να'), ('να', 'ζησεις'), ('ζησεις', 'κι'), ('κι', 'ο,τι'), ('ο,τι', 'ποθεις'), ('ποθεις', 'γι'), ('γι', 'τη'), ('τη', 'γιορτη'), ('γιορτη', 'σου'), ('σου', 'περρισσευκει'), ('περρισσευκει', 'σου'), ('σου', 'κανενα'), ('κανενα', 'εικοσαευρο')], [('γιατι', 'ολαν'), ('ολαν', 'το'), ('το', 'λαλεις'), ('λαλεις', 'τουτον'), ('τουτον', 'εν'), ('εν', 'πελλαμος')], [('φαε', 'παττιχα'), ('παττιχα', 'εν'), ('εν', 'γλυτζια'), ('γλυτζια', 'μελι')], [('τρωεις', 'παττιχα'), ('παττιχα', 'με'), ('με', 'χαλλουμι'), ('χαλλουμι', 'λειφκει'), ('λειφκει', 'σου'), ('σου', 'το'), ('το', 'χαλλουμι'), ('χαλλουμι', 'βαλλεις'), ('βαλλεις', 'αλλο'), ('αλλο', 'ενα'), ('ενα', 'κομματι')], [('εμεις', 'που'), ('που', 'το'), ('το', 'εδωκαμεν'), ('εδωκαμεν', 'μια'), ('μια', 'χαρα'), ('χαρα', 'το'), ('το', 'εκαταλαβαμεν')], [('ιντα', 'να'), ('να', 'ζησω'), ('ζησω', 'να'), ('να', 'με'), ('με', 'βασανιατε'), ('βασανιατε', 'τζιαλλο')], [('η', 'μπαταρια'), ('μπαταρια', 'εθυμισε'), ('εθυμισε', 'μ'), ('μ', 'εσενα'), ('εσενα', 'ποττε'), ('ποττε', 'δεν'), ('δεν', 'εχεις')], [('δουλευκουμε', 'για'), ('για', 'να'), ('να', 'παμε'), ('παμε', 'πουποτε'), ('πουποτε', 'τζιαι'), ('τζιαι', 'τελικα'), ('τελικα', 'εν'), ('εν', 'παμε'), ('παμε', 'πουποτε'), ('πουποτε', 'γιατι'), ('γιατι', 'δουλευκουμε')], [('εγω', 'τουτους'), ('τουτους', 'που'), ('που', 'εν'), ('εν', 'ετοιμοι'), ('ετοιμοι', 'την'), ('την', 'ωρα'), ('ωρα', 'που'), ('που', 'εκανονισαμε'), ('εκανονισαμε', 'εν'), ('εν', 'τους'), ('τους', 'εμπιστευκουμε')], [('εσηκωσα', 'το'), ('το', 'σιερι'), ('σιερι', 'μου'), ('μου', 'να'), ('να', 'φωναξω'), ('φωναξω', 'του'), ('του', 'σερβιτορου'), ('σερβιτορου', 'τζιαι'), ('τζιαι', 'που'), ('που', 'συνηθεια'), ('συνηθεια', 'εφκαλα'), ('εφκαλα', 'σελφι')], [('φερτε', 'μου'), ('μου', 'τζιεινον'), ('τζιεινον', 'που'), ('που', 'ειπε'), ('ειπε', 'ο'), ('ο', 'πελατης'), ('πελατης', 'εσιει'), ('εσιει', 'παντα'), ('παντα', 'δικαιο'), ('δικαιο', 'να'), ('να', 'τον'), ('τον', 'δερω'), ('δερω', 'αλλιως'), ('αλλιως', 'εν'), ('εν', 'να'), ('να', 'δερω'), ('δερω', 'τον'), ('τον', 'πελατη')], [('φκαινεις', 'που'), ('που', 'την'), ('την', 'θαλασσα'), ('θαλασσα', 'φακκα'), ('φακκα', 'το'), ('το', 'δαχτυλουι'), ('δαχτυλουι', 'σου'), ('σου', 'στον'), ('στον', 'βραχο'), ('βραχο', 'παιζεις'), ('παιζεις', 'το'), ('το', 'κουλ'), ('κουλ', 'παεις'), ('παεις', 'στο'), ('στο', 'κρεβατακι'), ('κρεβατακι', 'βαλλεις'), ('βαλλεις', 'γυαλια'), ('γυαλια', 'καππελο'), ('καππελο', 'κλαιεις')], [('παπα', 'αρεσκουν'), ('αρεσκουν', 'σου'), ('σου', 'τα'), ('τα', 'τζιαινουρκα'), ('τζιαινουρκα', 'μου'), ('μου', 'τακκουνια')], [('εσεις', 'που'), ('που', 'μιλατε'), ('μιλατε', 'πισω'), ('πισω', 'που'), ('που', 'τη'), ('τη', 'ρασιη'), ('ρασιη', 'μου'), ('μου', 'φυετε'), ('φυετε', 'λιο')], [('εσηκωσα', 'την'), ('την', 'κκελλε'), ('κκελλε', 'μου'), ('μου', 'που'), ('που', 'το'), ('το', 'κινητο')], [('θυμουμαι', 'τη'), ('τη', 'μανα'), ('μανα', 'μου'), ('μου', 'να'), ('να', 'με'), ('με', 'βουρα'), ('βουρα', 'σε'), ('σε', 'ουλλο'), ('ουλλο', 'το'), ('το', 'σπιτι'), ('σπιτι', 'για'), ('για', 'να'), ('να', 'πιω'), ('πιω', 'το'), ('το', 'γαλα'), ('γαλα', 'μου')], [('η', 'μανα'), ('μανα', 'μου'), ('μου', 'παντα'), ('παντα', 'λαλει'), ('λαλει', 'μου'), ('μου', 'οτι'), ('οτι', 'θελει'), ('θελει', 'να'), ('να', 'καμω'), ('καμω', 'κοπελλουθκια'), ('κοπελλουθκια', 'χειροτερα'), ('χειροτερα', 'που'), ('που', 'μενα')], [('για', 'εσας'), ('εσας', 'που'), ('που', 'σας'), ('σας', 'αρεσκει'), ('αρεσκει', 'ο'), ('ο', 'ηχος'), ('ηχος', 'της'), ('της', 'βροσιης')], [('ζητω', 'συγγνωμη'), ('συγγνωμη', 'που'), ('που', 'ουλλους'), ('ουλλους', 'σας')], [('φυε', 'που'), ('που', 'τη'), ('τη', 'κυπρο'), ('κυπρο', 'ωσπου'), ('ωσπου', 'εν'), ('εν', 'γλιορα')], [('η', 'κοινωνια'), ('κοινωνια', 'μας'), ('μας', 'εν'), ('εν', 'πολλα'), ('πολλα', 'συντηρητικη'), ('συντηρητικη', 'παρολο'), ('παρολο', 'που'), ('που', 'πιστευκω'), ('πιστευκω', 'οτι'), ('οτι', 'η'), ('η', 'πλειστη'), ('πλειστη', 'νεολαια'), ('νεολαια', 'εν'), ('εν', 'πιο'), ('πιο', 'προοδευτικη')], [('εχει', 'καποιους'), ('καποιους', 'που'), ('που', 'οντως'), ('οντως', 'ηρταν'), ('ηρταν', 'να'), ('να', 'δκιαβασουν'), ('δκιαβασουν', 'λιον'), ('λιον', 'ζιλικουρτι')], [('φκαλε', 'φαουσα'), ('φαουσα', 'γαμωτο'), ('γαμωτο', 'μισω'), ('μισω', 'σας'), ('σας', 'ουλλους'), ('ουλλους', 'θκιαολε'), ('θκιαολε', 'μαυρε')], [('εψες', 'εκαμνα'), ('εκαμνα', 'σεξ'), ('σεξ', 'με'), ('με', 'την'), ('την', 'κοπελλουα'), ('κοπελλουα', 'μου'), ('μου', 'τζιαι'), ('τζιαι', 'καταλαθως'), ('καταλαθως', 'ετραβηχτηκεν'), ('ετραβηχτηκεν', 'το'), ('το', 'συρμα'), ('συρμα', 'για'), ('για', 'το'), ('το', 'ρευμα'), ('ρευμα', 'τζιαι'), ('τζιαι', 'εκλεισεν')], [('δηλαδη', 'να'), ('να', 'μας'), ('μας', 'δερνει'), ('δερνει', 'ο'), ('ο', 'αντρας'), ('αντρας', 'δεν'), ('δεν', 'πειραζει'), ('πειραζει', 'εχει'), ('εχει', 'αντρισμο')], [('λαλειτε', 'εχαθηκαν'), ('εχαθηκαν', 'οι'), ('οι', 'ιπποτες'), ('ιπποτες', 'αλλα'), ('αλλα', 'αμαν'), ('αμαν', 'σας'), ('σας', 'προσεγγιζει'), ('προσεγγιζει', 'καποιος'), ('καποιος', 'ειστε'), ('ειστε', 'τοσο'), ('τοσο', 'αππωμενες'), ('αππωμενες', 'εν'), ('εν', 'εχει'), ('εχει', 'χειροτερο'), ('χειροτερο', 'πραμα')], [('νιωθω', 'πολλα'), ('πολλα', 'ασσιημα'), ('ασσιημα', 'αμαν'), ('αμαν', 'μαιρεφκω'), ('μαιρεφκω', 'για'), ('για', 'ενα'), ('ενα', 'λοχο'), ('λοχο', 'γιατι'), ('γιατι', 'εν'), ('εν', 'ηξερω'), ('ηξερω', 'να'), ('να', 'υπολογιζω'), ('υπολογιζω', 'ποσοτητες')], [('ειμαι', 'κυπρια'), ('κυπρια', 'τζαι'), ('τζαι', 'οπου'), ('οπου', 'παω'), ('παω', 'ξεκινουν'), ('ξεκινουν', 'τζαι'), ('τζαι', 'μιλουν'), ('μιλουν', 'μου'), ('μου', 'αγγλικα')], [('ηρτα', 'να'), ('να', 'δω'), ('δω', 'την'), ('την', 'γιαγια'), ('γιαγια', 'μου'), ('μου', 'εβιδωσεν'), ('εβιδωσεν', 'με'), ('με', 'παστον'), ('παστον', 'καναπε'), ('καναπε', 'να'), ('να', 'δουμε'), ('δουμε', 'πετρινο'), ('πετρινο', 'ποταμι'), ('ποταμι', 'τζαι'), ('τζαι', 'εν'), ('εν', 'με'), ('με', 'αφηνει'), ('αφηνει', 'να'), ('να', 'φιω')], [('θωρω', 'γαρους'), ('γαρους', 'τζιαι'), ('τζιαι', 'γαουρες'), ('γαουρες', 'να'), ('να', 'καμνουν'), ('καμνουν', 'καποιους'), ('καποιους', 'που'), ('που', 'νοιαζονται'), ('νοιαζονται', 'για'), ('για', 'τζεινους'), ('τζεινους', 'χωμα'), ('χωμα', 'γιατι'), ('γιατι', 'εν'), ('εν', 'μισιη'), ('μισιη', 'μου'), ('μου', 'ασσιημος'), ('ασσιημος', 'η'), ('η', 'ασσιημη'), ('ασσιημη', 'οξα'), ('οξα', 'πασσιης')], [('το', 'οτι'), ('οτι', 'βαλεις'), ('βαλεις', 'λαικ'), ('λαικ', 'σε'), ('σε', 'τεθκια'), ('τεθκια', 'κορουα'), ('κορουα', 'εν'), ('εν', 'υποσυνηδειτο'), ('υποσυνηδειτο', 'επδ'), ('επδ', 'ελκυει'), ('ελκυει', 'σε'), ('σε', 'εξωτερικα'), ('εξωτερικα', 'εν'), ('εν', 'και'), ('και', 'σημαινει'), ('σημαινει', 'οτι'), ('οτι', 'τερκαζετε'), ('τερκαζετε', 'η'), ('η', 'οτι'), ('οτι', 'θελεις'), ('θελεις', 'να'), ('να', 'καμετε'), ('καμετε', 'σχεση')], [('εν', 'να'), ('να', 'νευριαζω'), ('νευριαζω', 'αησμε'), ('αησμε', 'κορη'), ('κορη', 'μου')], [('υγραινομαι', 'αμαν'), ('αμαν', 'ακουω'), ('ακουω', 'τουτες'), ('τουτες', 'τις'), ('τις', 'λεξεις'), ('λεξεις', 'σαν'), ('σαν', 'τουτες'), ('τουτες', 'εν'), ('εν', 'εσχει')], [('παιθκια', 'χρειαζουμαι'), ('χρειαζουμαι', 'συμβουλη'), ('συμβουλη', 'τι'), ('τι', 'να'), ('να', 'καμω')], [('η', 'καλαμαρου'), ('καλαμαρου', 'βαρκεται'), ('βαρκεται', 'τοσο'), ('τοσο', 'πολλα')], [('εν', 'πολλα'), ('πολλα', 'μεγαλη'), ('μεγαλη', 'εν'), ('εν', 'την'), ('την', 'φωρει'), ('φωρει', 'το'), ('το', 'στομα'), ('στομα', 'μου')], [('εν', 'την'), ('την', 'κκελε'), ('κκελε', 'σου'), ('σου', 'που'), ('που', 'εν'), ('εν', 'να'), ('να', 'σσιησω'), ('σσιησω', 'αλλα'), ('αλλα', 'σσιηστο'), ('σσιηστο', 'εν'), ('εν', 'διω'), ('διω', 'μπακκιρα'), ('μπακκιρα', 'να'), ('να', 'πιαεις'), ('πιαεις', 'αλλο'), ('αλλο', 'να'), ('να', 'παεννεις'), ('παεννεις', 'θαλλασσα'), ('θαλλασσα', 'με'), ('με', 'τες'), ('τες', 'σοβρακες')], [('μανα', 'μου'), ('μου', 'κοπελια'), ('κοπελια', 'εφαμεν'), ('εφαμεν', 'τζε'), ('τζε', 'φετος'), ('φετος', 'την'), ('την', 'σουβλα'), ('σουβλα', 'μας')], [('ατε', 'μανα'), ('μανα', 'μου'), ('μου', 'να'), ('να', 'βρεξει'), ('βρεξει', 'να'), ('να', 'φαμε'), ('φαμε', 'κανενα'), ('κανενα', 'μανιταρι'), ('μανιταρι', 'που'), ('που', 'εν'), ('εν', 'τζαι'), ('τζαι', 'μουχτιν')], [('φιλε', 'μου'), ('μου', 'εν'), ('εν', 'λια'), ('λια', 'που'), ('που', 'τους'), ('τους', 'ειπες'), ('ειπες', 'ακομα')], [('αν', 'σε'), ('σε', 'πιασει'), ('πιασει', 'κανενας'), ('κανενας', 'και'), ('και', 'πει'), ('πει', 'σου'), ('σου', 'ομως'), ('ομως', 'εχω'), ('εχω', 'ρασιη'), ('ρασιη', 'πισω'), ('πισω', 'μου'), ('μου', 'ημουν'), ('ημουν', 'ουκ'), ('ουκ', 'λοκατζιης'), ('λοκατζιης', 'κλπ'), ('κλπ', 'πετου'), ('πετου', 'να'), ('να', 'ερτει'), ('ερτει', 'να'), ('να', 'καμει'), ('καμει', 'μιαν'), ('μιαν', 'βιδωτην'), ('βιδωτην', 'να'), ('να', 'δει'), ('δει', 'την'), ('την', 'γλυκα')], [('ενηξερουν', 'ποθεν'), ('ποθεν', 'κατουρα'), ('κατουρα', 'η'), ('η', 'ορνιχα'), ('ορνιχα', 'ρε'), ('ρε', 'τουτοοι'), ('τουτοοι', 'τσιαι'), ('τσιαι', 'εκαρτερουσετε'), ('εκαρτερουσετε', 'να'), ('να', 'καμουν'), ('καμουν', 'κατι'), ('κατι', 'καλυτερο')], [('εν', 'για'), ('για', 'τες'), ('τες', 'μπηχτες'), ('μπηχτες', 'ρε'), ('ρε', 'τουτοι')], [('εν', 'εκαταλαβαν'), ('εκαταλαβαν', 'ακομα'), ('ακομα', 'οτι'), ('οτι', 'εν'), ('εν', 'τα'), ('τα', 'περνει'), ('περνει', 'κανενας'), ('κανενας', 'μαζι'), ('μαζι', 'τους')], [('ουλους', 'τρωει'), ('τρωει', 'του'), ('του', 'το'), ('το', 'χωμα'), ('χωμα', 'σε'), ('σε', 'καποια'), ('καποια', 'φαση')], [('αηστους', 'τσιαμε'), ('τσιαμε', 'εγιω'), ('εγιω', 'σιερομαι'), ('σιερομαι', 'που'), ('που', 'εν'), ('εν', 'ετσι'), ('ετσι', 'μαππουροι'), ('μαππουροι', 'τσιλλιαραες')], [('εκατσε', 'ο'), ('ο', 'υπουργος'), ('υπουργος', 'με'), ('με', 'τεσσεροις'), ('τεσσεροις', 'βλακες'), ('βλακες', 'που'), ('που', 'καμνουν'), ('καμνουν', 'πως'), ('πως', 'καταλαβουν'), ('καταλαβουν', 'που'), ('που', 'κυνηγη'), ('κυνηγη', 'και'), ('και', 'ο'), ('ο', 'υπουργος'), ('υπουργος', 'εκαμνε'), ('εκαμνε', 'πως'), ('πως', 'ηξερε'), ('ηξερε', 'που'), ('που', 'κυνηγη'), ('κυνηγη', 'πως'), ('πως', 'καταλαβει'), ('καταλαβει', 'τουτοι'), ('τουτοι', 'ουλοι'), ('ουλοι', 'πληρωνοντε'), ('πληρωνοντε', 'με'), ('με', 'ενα'), ('ενα', 'σορο'), ('σορο', 'λεφτα'), ('λεφτα', 'που'), ('που', 'μπορουν'), ('μπορουν', 'να'), ('να', 'ζησουν'), ('ζησουν', 'πολλες'), ('πολλες', 'οικογενεις'), ('οικογενεις', 'τζε'), ('τζε', 'ηβραν'), ('ηβραν', 'την'), ('την', 'λυση'), ('λυση', 'για'), ('για', 'να'), ('να', 'σωσουν'), ('σωσουν', 'την'), ('την', 'κατασταση'), ('κατασταση', 'να'), ('να', 'κοψουμε'), ('κοψουμε', 'τεσσερις'), ('τεσσερις', 'εξορμησεις'), ('εξορμησεις', 'και'), ('και', 'ελυθηκε'), ('ελυθηκε', 'το'), ('το', 'προβλημα')], [('εφκηκαν', 'και'), ('και', 'μες'), ('μες', 'την'), ('την', 'τηλεοραση'), ('τηλεοραση', 'με'), ('με', 'ενα'), ('ενα', 'καπαρτησμα'), ('καπαρτησμα', 'οτι'), ('οτι', 'κοπελια'), ('κοπελια', 'εμεις'), ('εμεις', 'ειμαστε'), ('ειμαστε', 'εξυπνοι'), ('εξυπνοι', 'και'), ('και', 'ηβραμε'), ('ηβραμε', 'την'), ('την', 'λυση')], [('οχι', 'εσεις'), ('εσεις', 'που'), ('που', 'φωναζετε'), ('φωναζετε', 'τοσα'), ('τοσα', 'χρονια'), ('χρονια', 'εν'), ('εν', 'η'), ('η', 'λαθροθηρια'), ('λαθροθηρια', 'εν'), ('εν', 'ο'), ('ο', 'αλουπος')], [('την', 'αδεια'), ('αδεια', 'επληρωσετε'), ('επληρωσετε', 'την'), ('την', 'τα'), ('τα', 'λεφτα'), ('λεφτα', 'μας'), ('μας', 'επιασαμε'), ('επιασαμε', 'τα'), ('τα', 'και'), ('και', 'εσεις'), ('εσεις', 'φακκατε'), ('φακκατε', 'τσεραθκιες'), ('τσεραθκιες', 'αφου'), ('αφου', 'εμεις'), ('εμεις', 'παλε'), ('παλε', 'δαμε'), ('δαμε', 'εν'), ('εν', 'να'), ('να', 'ειμαστε'), ('ειμαστε', 'με'), ('με', 'γεματες'), ('γεματες', 'τις'), ('τις', 'τσεπες')], [('μα', 'ελογαριαζαν'), ('ελογαριαζαν', 'χωρις'), ('χωρις', 'τον'), ('τον', 'ξενοδοχο'), ('ξενοδοχο', 'ενα'), ('ενα', 'εχω'), ('εχω', 'να'), ('να', 'τους'), ('τους', 'πω'), ('πω', 'τα'), ('τα', 'λεφτα'), ('λεφτα', 'της'), ('της', 'αδειας'), ('αδειας', 'και'), ('και', 'για'), ('για', 'την'), ('την', 'κοροιδια'), ('κοροιδια', 'και'), ('και', 'μονο'), ('μονο', 'εν'), ('εν', 'να'), ('να', 'την'), ('την', 'πληρωσουν'), ('πληρωσουν', 'ακριβα'), ('ακριβα', 'γιατι'), ('γιατι', 'ο'), ('ο', 'καθενας'), ('καθενας', 'εν'), ('εν', 'να'), ('να', 'βλεπει'), ('βλεπει', 'το'), ('το', 'συμφερον'), ('συμφερον', 'του'), ('του', 'τωρα'), ('τωρα', 'και'), ('και', 'να'), ('να', 'παει')], [('ενοια', 'σας'), ('σας', 'και'), ('και', 'εν'), ('εν', 'εμειναμεν'), ('εμειναμεν', 'μονο'), ('μονο', 'στο'), ('στο', 'να'), ('να', 'του'), ('του', 'πουμεν'), ('πουμεν', 'το'), ('το', 'μονο'), ('μονο', 'που'), ('που', 'ενε'), ('ενε', 'μπορουσα'), ('μπορουσα', 'να'), ('να', 'καμω'), ('καμω', 'και'), ('και', 'εν'), ('εν', 'τζιαμε'), ('τζιαμε', 'που'), ('που', 'εχτιτζιασα'), ('εχτιτζιασα', 'παραπανω'), ('παραπανω', 'ηταν'), ('ηταν', 'να'), ('να', 'πιασω'), ('πιασω', 'την'), ('την', 'θηρα')], [('τον', 'λογο'), ('λογο', 'καταλαβαινετε'), ('καταλαβαινετε', 'τον'), ('τον', 'νομιζω'), ('νομιζω', 'και'), ('και', 'σιουρα'), ('σιουρα', 'εν'), ('εν', 'θα'), ('θα', 'επιαννα'), ('επιαννα', 'τον'), ('τον', 'κουμπαρο'), ('κουμπαρο', 'μου'), ('μου', 'στο'), ('στο', 'λαιμο'), ('λαιμο', 'μου'), ('μου', 'για'), ('για', 'ενναν'), ('ενναν', 'βλακα'), ('βλακα', 'αφου'), ('αφου', 'ηταν'), ('ηταν', 'απο'), ('απο', 'το'), ('το', 'επαγγελματικο'), ('επαγγελματικο', 'του'), ('του', 'περιβαλλον'), ('περιβαλλον', 'και'), ('και', 'σημερα'), ('σημερα', 'ξερετε'), ('ξερετε', 'τα'), ('τα', 'ουλλοι'), ('ουλλοι', 'οτι'), ('οτι', 'οι'), ('οι', 'δουλειες'), ('δουλειες', 'ειναι'), ('ειναι', 'αφαντες')], [('και', 'καποιοι'), ('καποιοι', 'που'), ('που', 'επια'), ('επια', 'ννα'), ('ννα', 'μου'), ('μου', 'πουν'), ('πουν', 'οτι'), ('οτι', 'εν'), ('εν', 'μιτσιοι'), ('μιτσιοι', 'εν'), ('εν', 'κοφκει'), ('κοφκει', 'ο'), ('ο', 'νους'), ('νους', 'τους'), ('τους', 'επιαν'), ('επιαν', 'τζιε'), ('τζιε', 'τζινοι'), ('τζινοι', 'εσσω'), ('εσσω', 'σιεσμενοι'), ('σιεσμενοι', 'γιατι'), ('γιατι', 'εν'), ('εν', 'τουτη'), ('τουτη', 'η'), ('η', 'νοοτροπια'), ('νοοτροπια', 'που'), ('που', 'μας'), ('μας', 'εφαεν')], [('μιλω', 'σας'), ('σας', 'ποιος'), ('ποιος', 'με'), ('με', 'ειδεν'), ('ειδεν', 'και'), ('και', 'δεν'), ('δεν', 'με'), ('με', 'φοβηθηκε')], [('καλα', 'του'), ('του', 'καμες')], [('ε', 'να'), ('να', 'ισιωσει'), ('ισιωσει', 'ο'), ('ο', 'νουρος'), ('νουρος', 'του'), ('του', 'ομως')], [('κοπελια', 'φετος'), ('φετος', 'εν'), ('εν', 'να'), ('να', 'εχουμε'), ('εχουμε', 'μεγαλο'), ('μεγαλο', 'προβλημα'), ('προβλημα', 'με'), ('με', 'τις'), ('τις', 'κουφαες')], [('να', 'παρακαλατε'), ('παρακαλατε', 'να'), ('να', 'βρεξει'), ('βρεξει', 'καλα'), ('καλα', 'περκει'), ('περκει', 'παρει'), ('παρει', 'το'), ('το', 'νερο'), ('νερο', 'κανενα'), ('κανενα', 'αυκο'), ('αυκο', 'γιατι'), ('γιατι', 'κατα'), ('κατα', 'που'), ('που', 'θωρω'), ('θωρω', 'εν'), ('εν', 'να'), ('να', 'μας'), ('μας', 'φαν'), ('φαν', 'φετος')], [('εχτος', 'που'), ('που', 'τουτο'), ('τουτο', 'φετος'), ('φετος', 'εν'), ('εν', 'κ'), ('κ', 'πιο'), ('πιο', 'επικινδυνες'), ('επικινδυνες', 'ειχαμε'), ('ειχαμε', 'βαρυχειμωνια'), ('βαρυχειμωνια', 'εν'), ('εν', 'νιστιτζιες'), ('νιστιτζιες', 'εχει'), ('εχει', 'τοσο'), ('τοσο', 'καιρο')], [('αλλες', 'χρονιες'), ('χρονιες', 'εβρισκες'), ('εβρισκες', 'ολοχρονα'), ('ολοχρονα', 'εκυνηγουσαν'), ('εκυνηγουσαν', 'εν'), ('εν', 'ηταν'), ('ηταν', 'τοσο'), ('τοσο', 'το'), ('το', 'δηλητηριο')], [('τουτη', 'την'), ('την', 'περιοδο'), ('περιοδο', 'ομως'), ('ομως', 'εν'), ('εν', 'θανατος')], [('επαρχια', 'λεμεσου'), ('λεμεσου', 'κ'), ('κ', 'εχτες'), ('εχτες', 'και'), ('και', 'σημερα'), ('σημερα', 'ηβραμε')], [('ηντα', 'που'), ('που', 'γινεται'), ('γινεται', 'ρε'), ('ρε', 'κοπελια')], [('ειπα', 'να'), ('να', 'καμω'), ('καμω', 'τουτο'), ('τουτο', 'το'), ('το', 'τοπουι'), ('τοπουι', 'δαμε'), ('δαμε', 'να'), ('να', 'λαλουμε'), ('λαλουμε', 'τα'), ('τα', 'δικα'), ('δικα', 'μας'), ('μας', 'τζιαι'), ('τζιαι', 'ας'), ('ας', 'μεν'), ('μεν', 'καταλαβουν'), ('καταλαβουν', 'οι'), ('οι', 'καλαμαραες')], [('να', 'δουμε'), ('δουμε', 'ποσοι'), ('ποσοι', 'εν'), ('εν', 'να'), ('να', 'συναχτουμε')], [('γραφετε', 'οτι'), ('οτι', 'θελετε'), ('θελετε', 'αλλα'), ('αλλα', 'οι'), ('οι', 'ξιτιμασιες'), ('ξιτιμασιες', 'για'), ('για', 'να'), ('να', 'μεν'), ('μεν', 'μας'), ('μας', 'το'), ('το', 'κλειωσουν')], [('ειπα', 'να'), ('να', 'ξιθαψω'), ('ξιθαψω', 'λιον'), ('λιον', 'τουτο'), ('τουτο', 'το'), ('το', 'θεμαν'), ('θεμαν', 'ρε'), ('ρε', 'κοπεγια')], [('ατε', 'να'), ('να', 'δουμεν'), ('δουμεν', 'ποσοι'), ('ποσοι', 'κυπρεοι'), ('κυπρεοι', 'ειμαστεν'), ('ειμαστεν', 'δαμεσα')], [('ατε', 'ρε'), ('ρε', 'τζε'), ('τζε', 'μεν'), ('μεν', 'ιξιανετε')], [('εν', 'πολλοι'), ('πολλοι', 'οι'), ('οι', 'κυπραιοι'), ('κυπραιοι', 'τελικα'), ('τελικα', 'απ'), ('απ', 'οτι'), ('οτι', 'θωρω'), ('θωρω', 'δαμε')], [('μα', 'πως'), ('πως', 'να'), ('να', 'πεισουν'), ('πεισουν', 'για'), ('για', 'ξενο'), ('ξενο', 'δακτυλο'), ('δακτυλο', 'οταν'), ('οταν', 'ο'), ('ο', 'κυπριακος'), ('κυπριακος', 'κωλοδακτυλος'), ('κωλοδακτυλος', 'εν'), ('εν', 'τοσο'), ('τοσο', 'προφανης')], [('εφκηκα', 'γυρον'), ('γυρον', 'τα'), ('τα', 'μπλογκς'), ('μπλογκς', 'που'), ('που', 'θκιαβαζω'), ('θκιαβαζω', 'να'), ('να', 'πω'), ('πω', 'τα'), ('τα', 'καλαντα')], [('ατε', 'καλη'), ('καλη', 'χρονια'), ('χρονια', 'τζιαι'), ('τζιαι', 'μεν'), ('μεν', 'συγχυζεσαι'), ('συγχυζεσαι', 'πολλα')], [('ο', 'φασισμος'), ('φασισμος', 'μονον'), ('μονον', 'αμαν'), ('αμαν', 'γινει'), ('γινει', 'εξουσια'), ('εξουσια', 'εν'), ('εν', 'επικινδυνος')], [('μανα', 'μου'), ('μου', 'εν'), ('εν', 'ξερετε'), ('ξερετε', 'για'), ('για', 'το'), ('το', 'εγκλημα'), ('εγκλημα', 'ζει'), ('ζει', 'στην'), ('στην', 'κυπρο')], [('αμαν', 'συλλαβουν'), ('συλλαβουν', 'εμπορο'), ('εμπορο', 'ναρκωτικων'), ('ναρκωτικων', 'τζε'), ('τζε', 'εν'), ('εν', 'αλλοδαπος'), ('αλλοδαπος', 'δηλαδη'), ('δηλαδη', 'ακελικος'), ('ακελικος', 'εκ'), ('εκ', 'τουρκιας'), ('τουρκιας', 'η'), ('η', 'κινεζος'), ('κινεζος', 'ρωσσος'), ('ρωσσος', 'η'), ('η', 'αραπης'), ('αραπης', 'αφηνουν'), ('αφηνουν', 'τον')], [('πιαννει', 'μονο'), ('μονο', 'τζεινους'), ('τζεινους', 'που'), ('που', 'ταυτισαν'), ('ταυτισαν', 'ουλλη'), ('ουλλη', 'την'), ('την', 'πολιτικη'), ('πολιτικη', 'τους'), ('τους', 'υπαρξη')], [('τελικα', 'εμας'), ('εμας', 'που'), ('που', 'εν'), ('εν', 'ημασταν'), ('ημασταν', 'τζιαμε'), ('τζιαμε', 'εν'), ('εν', 'να'), ('να', 'μας'), ('μας', 'πει'), ('πει', 'κανενας'), ('κανενας', 'ποιοι'), ('ποιοι', 'εν'), ('εν', 'υπερ'), ('υπερ', 'τζαι'), ('τζαι', 'ποιοι'), ('ποιοι', 'κατα')], [('εκαμαμεν', 'τον'), ('τον', 'πονο'), ('πονο', 'μας'), ('μας', 'ανεκδοτο'), ('ανεκδοτο', 'τα'), ('τα', 'αγρινα'), ('αγρινα', 'εν'), ('εν', 'πας'), ('πας', 'τα'), ('τα', 'βουνα'), ('βουνα', 'ειμαστεν'), ('ειμαστεν', 'τετραποδα')], [('ε', 'για'), ('για', 'τους'), ('τους', 'χαρακτηρισμους'), ('χαρακτηρισμους', 'που'), ('που', 'λαλεις'), ('λαλεις', 'οτι'), ('οτι', 'εχρησιμοποιαν'), ('εχρησιμοποιαν', 'ο'), ('ο', 'συγκεκριμενος'), ('συγκεκριμενος', 'τοτε'), ('τοτε', 'εν'), ('εν', 'αθυμουμαι'), ('αθυμουμαι', 'ακριβως'), ('ακριβως', 'αν'), ('αν', 'τζαι'), ('τζαι', 'εννα'), ('εννα', 'τα'), ('τα', 'κοιταξω'), ('κοιταξω', 'να'), ('να', 'θυμηθω')], [('εν', 'μπορω'), ('μπορω', 'φυσικα'), ('φυσικα', 'να'), ('να', 'καμω'), ('καμω', 'τον'), ('τον', 'δικηγορο'), ('δικηγορο', 'ισως'), ('ισως', 'τζαι'), ('τζαι', 'οι')], [('μα', 'εν'), ('εν', 'για'), ('για', 'τζιεινο'), ('τζιεινο', 'το'), ('το', 'δειπνον'), ('δειπνον', 'που'), ('που', 'εστησεν'), ('εστησεν', 'που'), ('που', 'λαλεις')], [('οι', 'οι'), ('οι', 'η'), ('η', 'κουβεντα'), ('κουβεντα', 'με'), ('με', 'το'), ('το', 'δειπνο'), ('δειπνο', 'εν'), ('εν', 'μετα')], [('η', 'φωτογραφια'), ('φωτογραφια', 'δαμε'), ('δαμε', 'εν'), ('εν', 'που'), ('που', 'το'), ('το', 'εθνικο'), ('εθνικο', 'συμβουλιο')], [('τζαι', 'εγω'), ('εγω', 'που'), ('που', 'ενομισα'), ('ενομισα', 'πως'), ('πως', 'εν'), ('εν', 'οι'), ('οι', 'ψηφοφοροι'), ('ψηφοφοροι', 'του'), ('του', 'δηκο'), ('δηκο', 'που'), ('που', 'την'), ('την', 'εφκαλαν')], [('περκει', 'να'), ('να', 'ξερεις'), ('ξερεις', 'τζαι'), ('τζαι', 'τι'), ('τι', 'εψηφισα'), ('εψηφισα', 'τζαι'), ('τζαι', 'γιατι'), ('γιατι', 'τζαι'), ('τζαι', 'ας'), ('ας', 'μεν'), ('μεν', 'ειμαι'), ('ειμαι', 'ακελικος'), ('ακελικος', 'αν'), ('αν', 'εθκιεβαζες'), ('εθκιεβαζες', 'την'), ('την', 'αναρτηση'), ('αναρτηση', 'πιο'), ('πιο', 'πανω'), ('πανω', 'μπορει'), ('μπορει', 'να'), ('να', 'το'), ('το', 'επροσεχες')], [('τζιαι', 'τζεινο'), ('τζεινο', 'του'), ('του', 'λαικου'), ('λαικου', 'μετωπου'), ('μετωπου', 'που'), ('που', 'καθε'), ('καθε', 'θκυο'), ('θκυο', 'μερες'), ('μερες', 'φκαλλει'), ('φκαλλει', 'διαγγελμα'), ('διαγγελμα', 'του'), ('του', 'στυλ'), ('στυλ', 'πατριωτες'), ('πατριωτες', 'στ'), ('στ', 'αρματα'), ('αρματα', 'εσιει'), ('εσιει', 'χαζι')], [('εν', 'φανερο'), ('φανερο', 'θελει'), ('θελει', 'σε'), ('σε', 'μανα'), ('μανα', 'μου'), ('μου', 'θελει'), ('θελει', 'σε')], [('φιλε', 'μου'), ('μου', 'ειπα'), ('ειπα', 'σου'), ('σου', 'το'), ('το', 'τζιαι'), ('τζιαι', 'τζιαμε')], [('λες', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'εξερες'), ('εξερες', 'με'), ('με', 'τι'), ('τι', 'ρεμαλια'), ('ρεμαλια', 'εισιες'), ('εισιες', 'να'), ('να', 'καμεις')], [('εν', 'ο'), ('ο', 'θεος'), ('θεος', 'που'), ('που', 'εφωτισεν'), ('εφωτισεν', 'τον'), ('τον', 'χριστοφια'), ('χριστοφια', 'τζιαι'), ('τζιαι', 'αλλαξεν'), ('αλλαξεν', 'πορεια')], [('οξα', 'αμπα'), ('αμπα', 'τζι'), ('τζι', 'εν'), ('εν', 'τουτοι')], [('για', 'αυτο'), ('αυτο', 'αναμενω'), ('αναμενω', 'να'), ('να', 'κρινω'), ('κρινω', 'που'), ('που', 'το'), ('το', 'αν'), ('αν', 'τζαι'), ('τζαι', 'τι'), ('τι', 'θα'), ('θα', 'συζητηθει'), ('συζητηθει', 'στο'), ('στο', 'πλαισιο'), ('πλαισιο', 'της'), ('της', 'ασφαλειας'), ('ασφαλειας', 'οταν'), ('οταν', 'ερτει'), ('ερτει', 'η'), ('η', 'ωρα')], [('τζιαι', 'σε'), ('σε', 'διαβεβαιω'), ('διαβεβαιω', 'οτι'), ('οτι', 'οταν'), ('οταν', 'εγινην'), ('εγινην', 'η'), ('η', 'επιθεση'), ('επιθεση', 'πας'), ('πας', 'τους'), ('τους', 'πυργους'), ('πυργους', 'ειπαν'), ('ειπαν', 'μου'), ('μου', 'το'), ('το', 'εναν'), ('εναν', 'τεταρτον'), ('τεταρτον', 'μετα'), ('μετα', 'τζιαι'), ('τζιαι', 'οταν'), ('οταν', 'επεθανεν'), ('επεθανεν', 'ο'), ('ο', 'χατζιηδακης'), ('χατζιηδακης', 'εμαθα'), ('εμαθα', 'το'), ('το', 'την'), ('την', 'ιδιαν'), ('ιδιαν', 'ημεραν')], [('μαλακιες', 'τουτα'), ('τουτα', 'ουλλα'), ('ουλλα', 'ασχολειθειτε'), ('ασχολειθειτε', 'λιο'), ('λιο', 'με'), ('με', 'κανενα'), ('κανενα', 'τηλεσκουπιδι'), ('τηλεσκουπιδι', 'να'), ('να', 'περασει'), ('περασει', 'η'), ('η', 'ωρα'), ('ωρα', 'σας'), ('σας', 'τζαι'), ('τζαι', 'κανει')], [('αφου', 'ξερουμεν'), ('ξερουμεν', 'πως'), ('πως', 'εν'), ('εν', 'μας'), ('μας', 'παν'), ('παν', 'στο'), ('στο', 'εξωτερικο'), ('εξωτερικο', 'γιατι'), ('γιατι', 'σκαλιζουμε'), ('σκαλιζουμε', 'συνεχως'), ('συνεχως', 'το'), ('το', 'θεμα'), ('θεμα', 'ετο'), ('ετο', 'να'), ('να', 'ασχολουμαστε'), ('ασχολουμαστε', 'ο'), ('ο', 'ενας'), ('ενας', 'με'), ('με', 'τον'), ('τον', 'αλλον'), ('αλλον', 'μες'), ('μες', 'την'), ('την', 'κυπρο'), ('κυπρο', 'μεταξυ'), ('μεταξυ', 'μας'), ('μας', 'τζαι'), ('τζαι', 'κανει')], [('μιλουμε', 'για'), ('για', 'καμποση'), ('καμποση', 'προπαγανδα'), ('προπαγανδα', 'πκιο'), ('πκιο', 'γελοιο'), ('γελοιο', 'τζιαι'), ('τζιαι', 'που'), ('που', 'τη'), ('τη', 'σημερινη'), ('σημερινη', 'χωρις'), ('χωρις', 'να'), ('να', 'εν'), ('εν', 'τοσο'), ('τοσο', 'ακραιο'), ('ακραιο', 'επειδη'), ('επειδη', 'δαμε'), ('δαμε', 'εν'), ('εν', 'η'), ('η', 'προπαγανδα'), ('προπαγανδα', 'ενος'), ('ενος', 'κρατους'), ('κρατους', 'προς'), ('προς', 'τα'), ('τα', 'εξω'), ('εξω', 'ασε'), ('ασε', 'που'), ('που', 'τωρα'), ('τωρα', 'εν'), ('εν', 'θα'), ('θα', 'εσιεις'), ('εσιεις', 'υποθεση'), ('υποθεση', 'αμα'), ('αμα', 'καμεις'), ('καμεις', 'αιτηση'), ('αιτηση', 'στον'), ('στον', 'αστρα'), ('αστρα', 'τζιαι'), ('τζιαι', 'δουν'), ('δουν', 'βιογραφικο')], [('ειδα', 'τζαι'), ('τζαι', 'εγω'), ('εγω', 'τεθκοια'), ('τεθκοια', 'οι'), ('οι', 'εν'), ('εν', 'εζησα'), ('εζησα', 'ποτε'), ('ποτε', 'ετσι'), ('ετσι', 'πραμα')], [('να', 'εξερες'), ('εξερες', 'ποσες'), ('ποσες', 'φορες'), ('φορες', 'εκλαψα'), ('εκλαψα', 'θωρωντας'), ('θωρωντας', 'τουτον'), ('τουτον', 'το'), ('το', 'βουνον'), ('βουνον', 'καθε'), ('καθε', 'πρωιν'), ('πρωιν', 'που'), ('που', 'παω'), ('παω', 'δουλειαν'), ('δουλειαν', 'τον'), ('τον', 'τελευταιον'), ('τελευταιον', 'τζαιρον')], [('λαλεις', 'ο'), ('ο', 'συνδιασμος'), ('συνδιασμος', 'του'), ('του', 'ταλεντου'), ('ταλεντου', 'με'), ('με', 'την'), ('την', 'δουλειαν'), ('δουλειαν', 'τζαι'), ('τζαι', 'τη'), ('τη', 'μουσικη'), ('μουσικη', 'γνωσην'), ('γνωσην', 'να'), ('να', 'δια'), ('δια', 'στην'), ('στην', 'δημιουργιαν'), ('δημιουργιαν', 'μιαν'), ('μιαν', 'διοχρονικην'), ('διοχρονικην', 'αξιαν'), ('αξιαν', 'περαν'), ('περαν', 'που'), ('που', 'την'), ('την', 'εμπορικην'), ('εμπορικην', 'τζαι'), ('τζαι', 'ναν'), ('ναν', 'για'), ('για', 'τουτο')], [('εν', 'το'), ('το', 'πιστευκω'), ('πιστευκω', 'πως'), ('πως', 'εξεχασες'), ('εξεχασες', 'πισω'), ('πισω', 'το'), ('το', 'πιο'), ('πιο', 'σημαντικο')], [('νομιζω', 'ειμαι'), ('ειμαι', 'τζιαι'), ('τζιαι', 'γω'), ('γω', 'σε'), ('σε', 'μιαν'), ('μιαν', 'πορταν'), ('πορταν', 'τζιαι'), ('τζιαι', 'στεκουμαι'), ('στεκουμαι', 'χωρις'), ('χωρις', 'να'), ('να', 'την'), ('την', 'ανοιξω'), ('ανοιξω', 'τζιαι'), ('τζιαι', 'χωρις'), ('χωρις', 'να'), ('να', 'παω'), ('παω', 'πισω')], [('τζιαι', 'μιας'), ('μιας', 'τζιαι'), ('τζιαι', 'εφερεν'), ('εφερεν', 'τουντες'), ('τουντες', 'χαζοβιολες'), ('χαζοβιολες', 'η'), ('η', 'κουβεντα'), ('κουβεντα', 'τι'), ('τι', 'παιζει'), ('παιζει', 'με'), ('με', 'σχεδον'), ('σχεδον', 'ουλλες'), ('ουλλες', 'που'), ('που', 'βαλλουν'), ('βαλλουν', 'φωτογραφιες'), ('φωτογραφιες', 'σελφι'), ('σελφι', 'σε'), ('σε', 'σεξι'), ('σεξι', 'ποζες'), ('ποζες', 'κορτωμενους'), ('κορτωμενους', 'κωλους'), ('κωλους', 'βυζια'), ('βυζια', 'εξω'), ('εξω', 'κλπ'), ('κλπ', 'τζιαι'), ('τζιαι', 'αλλα'), ('αλλα', 'που'), ('που', 'τουτα'), ('τουτα', 'γιατι'), ('γιατι', 'εν'), ('εν', 'πεθανισκετε'), ('πεθανισκετε', 'να'), ('να', 'σας'), ('σας', 'καμουμεν'), ('καμουμεν', 'νεκροψιαν'), ('νεκροψιαν', 'να'), ('να', 'δουμεν'), ('δουμεν', 'τες'), ('τες', 'εσωτερικες'), ('εσωτερικες', 'σας'), ('σας', 'ομορφιες')], [('εσκεφτουμουν', 'ποτε'), ('ποτε', 'εννα'), ('εννα', 'φκαλεις'), ('φκαλεις', 'φαουσαν'), ('φαουσαν', 'να'), ('να', 'πνασει'), ('πνασει', 'η'), ('η', 'κκελλε'), ('κκελλε', 'μου')], [('το', 'θετικον'), ('θετικον', 'εν'), ('εν', 'οτι'), ('οτι', 'σε'), ('σε', 'καθε'), ('καθε', 'περιπτωσην'), ('περιπτωσην', 'εφκαλες'), ('εφκαλες', 'τα'), ('τα', 'τζιαι'), ('τζιαι', 'επνασες'), ('επνασες', 'εστω'), ('εστω', 'τζιαι'), ('τζιαι', 'προσωρινα')], [('εγω', 'ηρτα'), ('ηρτα', 'στο'), ('στο', 'τσακ'), ('τσακ', 'να'), ('να', 'το'), ('το', 'χρησιμοποιησω'), ('χρησιμοποιησω', 'αλλα'), ('αλλα', 'εκρατηθηκα'), ('εκρατηθηκα', 'τζιαι'), ('τζιαι', 'θωρουν'), ('θωρουν', 'με'), ('με', 'ουλλοι'), ('ουλλοι', 'παραξενα'), ('παραξενα', 'λες'), ('λες', 'τζιαι'), ('τζιαι', 'εχω'), ('εχω', 'κατι'), ('κατι', 'στο'), ('στο', 'κουτελλο')], [('πε', 'μου'), ('μου', 'αληθκεια'), ('αληθκεια', 'ρε'), ('ρε', 'ειπες'), ('ειπες', 'ετσι'), ('ετσι', 'ατακα')], [('εν', 'τζιαμε'), ('τζιαμε', 'που'), ('που', 'αλλασσει'), ('αλλασσει', 'η'), ('η', 'ζωη'), ('ζωη', 'σου'), ('σου', 'τζιαι'), ('τζιαι', 'αραιωνουν'), ('αραιωνουν', 'δραματικα'), ('δραματικα', 'οι'), ('οι', 'εξοδοι')], [('κατα', 'τα'), ('τα', 'αλλα'), ('αλλα', 'ο'), ('ο', 'καθενας'), ('καθενας', 'καμνει'), ('καμνει', 'ο,τι'), ('ο,τι', 'θελει'), ('θελει', 'τζιαι'), ('τζιαι', 'ο,τι'), ('ο,τι', 'κοφκει'), ('κοφκει', 'ο'), ('ο', 'νους'), ('νους', 'του')], [('φορω', 'τες'), ('τες', 'χαζες'), ('χαζες', 'ροζ'), ('ροζ', 'ριαρες'), ('ριαρες', 'τζαι'), ('τζαι', 'μολις'), ('μολις', 'φυω'), ('φυω', 'φκαλλω'), ('φκαλλω', 'τες')], [('να', 'σε'), ('σε', 'δω'), ('δω', 'με'), ('με', 'βιλλουθκια'), ('βιλλουθκια', 'πανω'), ('πανω', 'στην'), ('στην', 'κελλε'), ('κελλε', 'τζιαι'), ('τζιαι', 'ειδα'), ('ειδα', 'τα'), ('τα', 'ουλλα')], [('αμπα', 'τζιαι'), ('τζιαι', 'φτασεις'), ('φτασεις', 'τζιαι'), ('τζιαι', 'καμεις'), ('καμεις', 'το'), ('το', 'μια'), ('μια', 'φορα'), ('φορα', 'θα'), ('θα', 'πρεπει'), ('πρεπει', 'να'), ('να', 'το'), ('το', 'καμνεις'), ('καμνεις', 'συνεχεια')], [('γιατι', 'τουτο'), ('τουτο', 'να'), ('να', 'καλυφκει'), ('καλυφκει', 'τες'), ('τες', 'ψυχικες'), ('ψυχικες', 'μας'), ('μας', 'αναγκες'), ('αναγκες', 'τζιαμε'), ('τζιαμε', 'που'), ('που', 'σαφεστατα'), ('σαφεστατα', 'δεν'), ('δεν', 'υπαρχει'), ('υπαρχει', 'σοβαρο'), ('σοβαρο', 'προβλημα'), ('προβλημα', 'ζιαμε'), ('ζιαμε', 'εν'), ('εν', 'πιο'), ('πιο', 'φροντιδα'), ('φροντιδα', 'με'), ('με', 'καλλωπισμο'), ('καλλωπισμο', 'μαζι'), ('μαζι', 'που'), ('που', 'ενιγουεις'), ('ενιγουεις', 'καμνουν'), ('καμνουν', 'το'), ('το', 'τζιαι'), ('τζιαι', 'αντρες'), ('αντρες', 'τζιαι'), ('τζιαι', 'γεναιτζιες'), ('γεναιτζιες', 'που'), ('που', 'τον'), ('τον', 'τζιαιρο'), ('τζιαιρο', 'που'), ('που', 'εσταματησαμεν'), ('εσταματησαμεν', 'να'), ('να', 'ζουμε'), ('ζουμε', 'στες'), ('στες', 'σπηλιες'), ('σπηλιες', 'τζιαι'), ('τζιαι', 'εκοινωνικοποιηθηκαμεν')], [('εν', 'ρητορικες'), ('ρητορικες', 'οι'), ('οι', 'ερωτησεις'), ('ερωτησεις', 'μου'), ('μου', 'εν'), ('εν', 'σαν'), ('σαν', 'να'), ('να', 'τζιαι'), ('τζιαι', 'μιλω'), ('μιλω', 'τζιαι'), ('τζιαι', 'του'), ('του', 'εαυτου'), ('εαυτου', 'μου')], [('μια', 'κορουα'), ('κορουα', 'που'), ('που', 'ξερω'), ('ξερω', 'εβαλε'), ('εβαλε', 'χειλη'), ('χειλη', 'εν'), ('εν', 'εισιε'), ('εισιε', 'τιποτε'), ('τιποτε', 'πριν'), ('πριν', 'εν'), ('εν', 'αληθκεια')], [('εν', 'θεμα'), ('θεμα', 'ισορροπιας'), ('ισορροπιας', 'πρεπει'), ('πρεπει', 'να'), ('να', 'μεγαλωνουμε'), ('μεγαλωνουμε', 'τα'), ('τα', 'κοπελλουθκια'), ('κοπελλουθκια', 'με'), ('με', 'τσαγανο'), ('τσαγανο', 'με'), ('με', 'δυναμη')], [('φυσικα', 'εσιει'), ('εσιει', 'σημασια'), ('σημασια', 'τζιαι'), ('τζιαι', 'καμνει'), ('καμνει', 'διαφορα'), ('διαφορα', 'το'), ('το', 'πως'), ('πως', 'συμπεριφερεσαι')], [('εσιει', 'ομως'), ('ομως', 'τζιαι'), ('τζιαι', 'καποιους'), ('καποιους', 'αθρωπους'), ('αθρωπους', 'που'), ('που', 'ειτε'), ('ειτε', 'δυσκολευουνται'), ('δυσκολευουνται', 'ειτε'), ('ειτε', 'εν'), ('εν', 'ηξερουν'), ('ηξερουν', 'πως'), ('πως', 'να'), ('να', 'εκφραστουν')], [('εν', 'σημαινει'), ('σημαινει', 'οτι'), ('οτι', 'εν'), ('εν', 'εννουν'), ('εννουν', 'με'), ('με', 'καλο'), ('καλο', 'τροπο'), ('τροπο', 'οσα'), ('οσα', 'λαλουν')], [('εβαλες', 'με'), ('με', 'σε'), ('σε', 'σκεψεις'), ('σκεψεις', 'τωρα'), ('τωρα', 'εν'), ('εν', 'τα'), ('τα', 'θυμουμαι'), ('θυμουμαι', 'τζιαι'), ('τζιαι', 'εγω'), ('εγω', 'μνημη'), ('μνημη', 'ψαρκου'), ('ψαρκου', 'εχω'), ('εχω', 'φιλουθκια')], [('μα', 'που'), ('που', 'να'), ('να', 'αρκεψω'), ('αρκεψω', 'τζιαι'), ('τζιαι', 'που'), ('που', 'να'), ('να', 'τελιωσω')], [('σκεφτου', 'ποσο'), ('ποσο', 'ευκολο'), ('ευκολο', 'εν'), ('εν', 'για'), ('για', 'σενα'), ('σενα', 'να'), ('να', 'παραδεκτεις'), ('παραδεκτεις', 'η'), ('η', 'να'), ('να', 'πεις'), ('πεις', 'οτι'), ('οτι', 'αυνανιζεσαι')], [('τζαι', 'μιλουν'), ('μιλουν', 'σου'), ('σου', 'τζαι'), ('τζαι', 'μιλας'), ('μιλας', 'τους'), ('τους', 'πισω')], [('μιλω', 'σε'), ('σε', 'ουλλον'), ('ουλλον', 'τον'), ('τον', 'κοσμο'), ('κοσμο', 'εν'), ('εν', 'εχω'), ('εχω', 'ετσι'), ('ετσι', 'κολληματα'), ('κολληματα', 'επειδη'), ('επειδη', 'ενα'), ('ενα', 'θεμα'), ('θεμα', 'μπορει'), ('μπορει', 'να'), ('να', 'μεν'), ('μεν', 'μας'), ('μας', 'αφορα'), ('αφορα', 'αμεσα'), ('αμεσα', 'εν'), ('εν', 'σημαινει'), ('σημαινει', 'δεν'), ('δεν', 'γινεται'), ('γινεται', 'γυρον'), ('γυρον', 'μας')], [('το', 'προβλημα'), ('προβλημα', 'εννεν'), ('εννεν', 'η'), ('η', 'συνηθεια'), ('συνηθεια', 'καθαυτη'), ('καθαυτη', 'αλλα'), ('αλλα', 'ο'), ('ο', 'λογος'), ('λογος', 'που'), ('που', 'την'), ('την', 'χρησιμοποιας')], [('αν', 'εισαι'), ('εισαι', 'μονος'), ('μονος', 'σου'), ('σου', 'καλη'), ('καλη', 'ωρα'), ('ωρα', 'τζιαι'), ('τζιαι', 'βλεπεις'), ('βλεπεις', 'τσοντες'), ('τσοντες', 'γιατι'), ('γιατι', 'φοασαι'), ('φοασαι', 'να'), ('να', 'μιλησεις'), ('μιλησεις', 'στο'), ('στο', 'φυλο'), ('φυλο', 'που'), ('που', 'σε'), ('σε', 'ενδιαφερει'), ('ενδιαφερει', 'τοτε'), ('τοτε', 'ναι'), ('ναι', 'εσιεις'), ('εσιεις', 'προβλημα')], [('αν', 'εισαι'), ('εισαι', 'παντρεμενος'), ('παντρεμενος', 'τζιαι'), ('τζιαι', 'βλεπεις'), ('βλεπεις', 'τσοντες'), ('τσοντες', 'γιατι'), ('γιατι', 'εν'), ('εν', 'μπορεις'), ('μπορεις', 'να'), ('να', 'επικοινωνησεις'), ('επικοινωνησεις', 'τις'), ('τις', 'επιθυμιες'), ('επιθυμιες', 'σου'), ('σου', 'με'), ('με', 'τοντην'), ('τοντην', 'συντροφο'), ('συντροφο', 'σου'), ('σου', 'τοτε'), ('τοτε', 'ναι'), ('ναι', 'εσιεις'), ('εσιεις', 'προβλημα')], [('αυτο', 'που'), ('που', 'ηθελα'), ('ηθελα', 'να'), ('να', 'θιξω'), ('θιξω', 'χωρις'), ('χωρις', 'να'), ('να', 'συμφωνησω'), ('συμφωνησω', 'η'), ('η', 'να'), ('να', 'διαφωνησω'), ('διαφωνησω', 'με'), ('με', 'τους'), ('τους', 'αντρες'), ('αντρες', 'που'), ('που', 'θωρουν'), ('θωρουν', 'πορνο'), ('πορνο', 'ηταν'), ('ηταν', 'το'), ('το', 'ποσο'), ('ποσο', 'πολυ'), ('πολυ', 'πονο'), ('πονο', 'προκαλουν'), ('προκαλουν', 'στις'), ('στις', 'συντροφους'), ('συντροφους', 'τους'), ('τους', 'τζιαι'), ('τζιαι', 'επροσπαθησα'), ('επροσπαθησα', 'να'), ('να', 'το'), ('το', 'δω'), ('δω', 'λλιο'), ('λλιο', 'πιο'), ('πιο', 'σφαιρικα'), ('σφαιρικα', 'το'), ('το', 'θεμα')], [('εν', 'ηξερω'), ('ηξερω', 'αν'), ('αν', 'εν'), ('εν', 'θεμα'), ('θεμα', 'προβληματος'), ('προβληματος', 'η'), ('η', 'θεμα'), ('θεμα', 'κατανοησης'), ('κατανοησης', 'η'), ('η', 'αν'), ('αν', 'εν'), ('εν', 'λογια'), ('λογια', 'της'), ('της', 'παρηορκας'), ('παρηορκας', 'οποταν'), ('οποταν', 'μπορει'), ('μπορει', 'να'), ('να', 'μεν'), ('μεν', 'διαφωνουμεν'), ('διαφωνουμεν', 'καθολου'), ('καθολου', 'στην'), ('στην', 'ουσια')], [('τωρα', 'για'), ('για', 'οσους'), ('οσους', 'εν'), ('εν', 'μονοι'), ('μονοι', 'τους'), ('τους', 'θεωρω'), ('θεωρω', 'το'), ('το', 'πιο'), ('πιο', 'φυσιολογικο'), ('φυσιολογικο', 'εστω'), ('εστω', 'γιατι'), ('γιατι', 'εν'), ('εν', 'πολλοι'), ('πολλοι', 'οι'), ('οι', 'λογοι'), ('λογοι', 'που'), ('που', 'μπορουν'), ('μπορουν', 'να'), ('να', 'οδηγησουν'), ('οδηγησουν', 'εναν'), ('εναν', 'αντρα'), ('αντρα', 'μονο'), ('μονο', 'του'), ('του', 'να'), ('να', 'θωρει'), ('θωρει', 'πορνο')], [('εν', 'ακομα'), ('ακομα', 'πιο'), ('πιο', 'δυσκολο'), ('δυσκολο', 'οι'), ('οι', 'γυναικες'), ('γυναικες', 'να'), ('να', 'το'), ('το', 'κατανοησουν'), ('κατανοησουν', 'και'), ('και', 'να'), ('να', 'το'), ('το', 'δεκτουν'), ('δεκτουν', 'μεν'), ('μεν', 'σου'), ('σου', 'πω'), ('πω', 'τζιαι'), ('τζιαι', 'πολλα'), ('πολλα', 'ασχημο'), ('ασχημο', 'γι'), ('γι', 'αυτες')], [('εφκαλα', 'τα'), ('τα', 'μμαθκια'), ('μμαθκια', 'μου'), ('μου', 'η'), ('η', 'αληθκεια')], [('να', 'μεν'), ('μεν', 'θωρειτε'), ('θωρειτε', 'τιποτε'), ('τιποτε', 'ομως'), ('ομως', 'τζιαι'), ('τζιαι', 'κανει')], [('θκιαβαζω', 'σε'), ('σε', 'αλλα'), ('αλλα', 'το'), ('το', 'εχεις'), ('εχεις', 'ολο'), ('ολο', 'λαθος')], [('πε', 'μου'), ('μου', 'εναν'), ('εναν', 'που'), ('που', 'εν'), ('εν', 'αυνανιζεται'), ('αυνανιζεται', 'σιορ')], [('εν', 'τζιαι'), ('τζιαι', 'σημαινει'), ('σημαινει', 'οτι'), ('οτι', 'φανταζεσαι'), ('φανταζεσαι', 'τζιεινα'), ('τζιεινα', 'που'), ('που', 'βλεπεις')], [('οι', 'αντρες'), ('αντρες', 'εν'), ('εν', 'πιο'), ('πιο', 'οπτικοι'), ('οπτικοι', 'τυποι')], [('εξαρταται', 'που'), ('που', 'την'), ('την', 'ηλικια'), ('ηλικια', 'και'), ('και', 'τες'), ('τες', 'περιστασεις'), ('περιστασεις', 'τωρα'), ('τωρα', 'αν'), ('αν', 'εισαι'), ('εισαι', 'με'), ('με', 'συντροφο'), ('συντροφο', 'τζιαι'), ('τζιαι', 'εξακολουθεις'), ('εξακολουθεις', 'να'), ('να', 'το'), ('το', 'καμνεις'), ('καμνεις', 'εν'), ('εν', 'ηξερω'), ('ηξερω', 'εν'), ('εν', 'ειμαι'), ('ειμαι', 'ειδικη')], [('η', 'αν'), ('αν', 'εν'), ('εν', 'απλα'), ('απλα', 'για'), ('για', 'εκτονωση'), ('εκτονωση', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'οκ'), ('οκ', 'ο'), ('ο', 'αλλος'), ('αλλος', 'που'), ('που', 'εν'), ('εν', 'μαζι'), ('μαζι', 'σου')], [('εν', 'φυσιολογικο'), ('φυσιολογικο', 'ομως'), ('ομως', 'να'), ('να', 'σαι'), ('σαι', 'σε'), ('σε', 'σχεση'), ('σχεση', 'τζιαι'), ('τζιαι', 'να'), ('να', 'φανταζεσαι'), ('φανταζεσαι', 'διαφορα')], [('καλο', 'θα'), ('θα', 'ηταν'), ('ηταν', 'να'), ('να', 'μην'), ('μην', 'περιστρεφουνται'), ('περιστρεφουνται', 'ουλλα'), ('ουλλα', 'στη'), ('στη', 'ζωη'), ('ζωη', 'σου'), ('σου', 'που'), ('που', 'το'), ('το', 'σεξ'), ('σεξ', 'τζιαι'), ('τζιαι', 'πολλα'), ('πολλα', 'καλλυττερο'), ('καλλυττερο', 'να'), ('να', 'επιστρεψεις'), ('επιστρεψεις', 'στην'), ('στην', 'πραγματικοτητα'), ('πραγματικοτητα', 'γιατι'), ('γιατι', 'εν'), ('εν', 'τζιαιν'), ('τζιαιν', 'ουλλα'), ('ουλλα', 'οσα'), ('οσα', 'θωρεις'), ('θωρεις', 'τζιαι'), ('τζιαι', 'φανταζεσαι'), ('φανταζεσαι', 'αληθινα')], [('εν', 'τοσο'), ('τοσο', 'εξωφρενικο'), ('εξωφρενικο', 'που'), ('που', 'εν'), ('εν', 'πολλα'), ('πολλα', 'αστειο'), ('αστειο', 'ατε'), ('ατε', 'ρε'), ('ρε', 'εν'), ('εν', 'γινεται')], [('αμπα', 'και'), ('και', 'εκαμναν'), ('εκαμναν', 'απου'), ('απου', 'πει'), ('πει', 'την'), ('την', 'πιο'), ('πιο', 'μεγαλη'), ('μεγαλη', 'τσιοφτα'), ('τσιοφτα', 'για'), ('για', 'να'), ('να', 'θωρουν'), ('θωρουν', 'αντιδρασεις'), ('αντιδρασεις', 'τζιαι'), ('τζιαι', 'να'), ('να', 'καμνουν'), ('καμνουν', 'πλακα')], [('ειμαι', 'κυπρο'), ('κυπρο', 'για'), ('για', 'λιες'), ('λιες', 'μερες'), ('μερες', 'διακοπες'), ('διακοπες', 'τζιαι'), ('τζιαι', 'τα'), ('τα', 'λιοπετριτικα'), ('λιοπετριτικα', 'δινουν'), ('δινουν', 'τζιαι'), ('τζιαι', 'περνουν')], [('μασιαλλα', 'μπραβο'), ('μπραβο', 'σου'), ('σου', 'ειμαι'), ('ειμαι', 'τζι'), ('τζι', 'εγω'), ('εγω', 'κουμερα'), ('κουμερα', 'αλλα'), ('αλλα', 'δεν'), ('δεν', 'εχω'), ('εχω', 'τις'), ('τις', 'ικανοτητες'), ('ικανοτητες', 'τζαι'), ('τζαι', 'τα'), ('τα', 'προσοντα'), ('προσοντα', 'σου'), ('σου', 'ουτε'), ('ουτε', 'καλη'), ('καλη', 'διοργανωτρια'), ('διοργανωτρια', 'ειμαι'), ('ειμαι', 'ουτε'), ('ουτε', 'ευφανταστες'), ('ευφανταστες', 'ιδεες'), ('ιδεες', 'εχω'), ('εχω', 'ετο'), ('ετο', 'απλα'), ('απλα', 'εν'), ('εν', 'πολλα'), ('πολλα', 'καλα'), ('καλα', 'πλασματα'), ('πλασματα', 'οι'), ('οι', 'κουμερες'), ('κουμερες', 'μου'), ('μου', 'τζι'), ('τζι', 'αγαπουν'), ('αγαπουν', 'με')], [('παντως', 'εσυζητουσαμεν'), ('εσυζητουσαμεν', 'τζιαι'), ('τζιαι', 'μεις'), ('μεις', 'με'), ('με', 'την'), ('την', 'παρεα'), ('παρεα', 'για'), ('για', 'την'), ('την', 'κατασταση'), ('κατασταση', 'τζαμε'), ('τζαμε', 'στην'), ('στην', 'πανδωρα'), ('πανδωρα', 'εν'), ('εν', 'εσιει'), ('εσιει', 'ετσι'), ('ετσι', 'πραμα'), ('πραμα', 'χαιρομαι'), ('χαιρομαι', 'που'), ('που', 'εστω'), ('εστω', 'και'), ('και', 'καποιος'), ('καποιος', 'εφκυκεν'), ('εφκυκεν', 'να'), ('να', 'πει'), ('πει', 'κατι')], [('καλα', 'τζιαι'), ('τζιαι', 'συγκρατηθηκες'), ('συγκρατηθηκες', 'εγω'), ('εγω', 'εν'), ('εν', 'μπορω')], [('αφηκα', 'το'), ('το', 'λαπτοπ'), ('λαπτοπ', 'ανοικτο'), ('ανοικτο', 'στο'), ('στο', 'μπλογκ'), ('μπλογκ', 'σου'), ('σου', 'πας'), ('πας', 'το'), ('το', 'τραπεζι'), ('τραπεζι', 'να'), ('να', 'καμω'), ('καμω', 'καφε')], [('εγω', 'παντως'), ('παντως', 'με'), ('με', 'τες'), ('τες', 'φιλες'), ('φιλες', 'μου'), ('μου', 'που'), ('που', 'εν'), ('εν', 'ωραιες'), ('ωραιες', 'τζιαι'), ('τζιαι', 'αρεσκουν'), ('αρεσκουν', 'μου'), ('μου', 'εξωτερικα'), ('εξωτερικα', 'θα'), ('θα', 'επηαινα'), ('επηαινα', 'μαζι'), ('μαζι', 'τους'), ('τους', 'ανετα'), ('ανετα', 'αν'), ('αν', 'ημουν'), ('ημουν', 'σε'), ('σε', 'φαση'), ('φαση', 'τζιαι'), ('τζιαι', 'ειχα'), ('ειχα', 'ευκαιρια')], [('εν', 'θα'), ('θα', 'επηαινα'), ('επηαινα', 'με'), ('με', 'τζεινες'), ('τζεινες', 'που'), ('που', 'εν'), ('εν', 'μου'), ('μου', 'αρεσκουν'), ('αρεσκουν', 'τζιαι'), ('τζιαι', 'θωρω'), ('θωρω', 'τες'), ('τες', 'καθαρα'), ('καθαρα', 'φιλικα')], [('εγω', 'εν'), ('εν', 'εκαταλαβα'), ('εκαταλαβα', 'που'), ('που', 'επηε'), ('επηε', 'λαθος')], [('και', 'μετα'), ('μετα', 'λαλουν'), ('λαλουν', 'για'), ('για', 'τες'), ('τες', 'γυναικες'), ('γυναικες', 'οτι'), ('οτι', 'δινουν'), ('δινουν', 'μικτα'), ('μικτα', 'σηματα')], [('ν', 'υπαρχει'), ('υπαρχει', 'φιλια'), ('φιλια', 'μεταξυ'), ('μεταξυ', 'αντρων'), ('αντρων', 'τζιαι'), ('τζιαι', 'γυναικων'), ('γυναικων', 'ουλλοι'), ('ουλλοι', 'στο'), ('στο', 'τελος'), ('τελος', 'σκεφτουνται'), ('σκεφτουνται', 'πως'), ('πως', 'να'), ('να', 'σε'), ('σε', 'γωνιασουν')], [('εκαμες', 'τζαι'), ('τζαι', 'εσυ'), ('εσυ', 'λαθος'), ('λαθος', 'ομως'), ('ομως', 'διοτι'), ('διοτι', 'λαλεις'), ('λαλεις', 'πως'), ('πως', 'εν'), ('εν', 'του'), ('του', 'ελαλες'), ('ελαλες', 'τιποτε'), ('τιποτε', 'προσωπικο'), ('προσωπικο', 'αλλα'), ('αλλα', 'ειπες'), ('ειπες', 'του'), ('του', 'για'), ('για', 'την'), ('την', 'παλια'), ('παλια', 'σχεση')], [('πως', 'να'), ('να', 'μεν'), ('μεν', 'του'), ('του', 'γυρισει'), ('γυρισει', 'τζαι'), ('τζαι', 'να'), ('να', 'τριππαρει')], [('εν', 'επηες'), ('επηες', 'παρακατω')], [('και', 'τουτος'), ('τουτος', 'που'), ('που', 'ενει'), ('ενει', 'τωρα'), ('τωρα', 'τζαι'), ('τζαι', 'αφηκε'), ('αφηκε', 'σε'), ('σε', 'ενω'), ('ενω', 'αλλοι'), ('αλλοι', 'βουρουν'), ('βουρουν', 'κομα')], [('συμφωνω', 'τζ'), ('τζ', 'εγω'), ('εγω', 'με'), ('με', 'οσα'), ('οσα', 'εγραψες')], [('εννα', 'με'), ('με', 'πεισεις'), ('πεισεις', 'τζιαι'), ('τζιαι', 'μετα'), ('μετα', 'αν'), ('αν', 'ξαναδιαπιστωσω'), ('ξαναδιαπιστωσω', 'αυτην'), ('αυτην', 'μου'), ('μου', 'τη'), ('τη', 'διαπιστωση'), ('διαπιστωση', 'θα'), ('θα', 'σε'), ('σε', 'εβρω'), ('εβρω', 'να'), ('να', 'μου'), ('μου', 'τραουδας'), ('τραουδας', 'λαλω'), ('λαλω', 'σου'), ('σου', 'το')], [('χωρκαθκιον', 'δεν'), ('δεν', 'ειναι'), ('ειναι', 'τα'), ('τα', 'καρναβαλια'), ('καρναβαλια', 'ειναι'), ('ειναι', 'να'), ('να', 'εν'), ('εν', 'τοσο'), ('τοσο', 'ψηλα'), ('ψηλα', 'η'), ('η', 'μουτη'), ('μουτη', 'τους'), ('τους', 'που'), ('που', 'να'), ('να', 'μεν'), ('μεν', 'μπορεις'), ('μπορεις', 'να'), ('να', 'δουν'), ('δουν', 'λιη'), ('λιη', 'ελαφροτητα'), ('ελαφροτητα', 'γυρω'), ('γυρω', 'τους'), ('τους', 'και'), ('και', 'μεσα'), ('μεσα', 'τους')], [('καμετε', 'χαζι'), ('χαζι', 'σιορ'), ('σιορ', 'εννα'), ('εννα', 'σας'), ('σας', 'καμει'), ('καμει', 'καλο')], [('εν', 'καταλαβω'), ('καταλαβω', 'τι'), ('τι', 'σημαινει'), ('σημαινει', 'χωρκαθκιον'), ('χωρκαθκιον', 'πιον'), ('πιον', 'εβαρεθηκα'), ('εβαρεθηκα', 'να'), ('να', 'ακουω'), ('ακουω', 'το'), ('το', 'εναν'), ('εναν', 'εν'), ('εν', 'χωρκατικον'), ('χωρκατικον', 'το'), ('το', 'αλλο'), ('αλλο', 'εν'), ('εν', 'χωρκατικον')], [('αν', 'εν'), ('εν', 'ετσι'), ('ετσι', 'η'), ('η', 'κυπρος'), ('κυπρος', 'εν'), ('εν', 'ενα'), ('ενα', 'χωρκο'), ('χωρκο', 'μεγαλο'), ('μεγαλο', 'αρα'), ('αρα', 'ειμαστε'), ('ειμαστε', 'ουλλοι'), ('ουλλοι', 'χωρκατοι')], [('και', 'αφηστους'), ('αφηστους', 'τζεινους'), ('τζεινους', 'που'), ('που', 'την'), ('την', 'βρισκουν'), ('βρισκουν', 'να'), ('να', 'παν'), ('παν', 'και'), ('και', 'να'), ('να', 'περασουν'), ('περασουν', 'καλα'), ('καλα', 'ταχα'), ('ταχα', 'φακκα'), ('φακκα', 'μου'), ('μου', 'και'), ('και', 'εμενα'), ('εμενα', 'το'), ('το', 'καρναβαλι'), ('καρναβαλι', 'αλλα'), ('αλλα', 'εν'), ('εν', 'το'), ('το', 'καμνω'), ('καμνω', 'θεμα')], [('εν', 'ηξερω'), ('ηξερω', 'σιουρα'), ('σιουρα', 'μαλλον'), ('μαλλον', 'εν'), ('εν', 'παραλογισμοι'), ('παραλογισμοι', 'του'), ('του', 'πονεμενου'), ('πονεμενου', 'εν'), ('εν', 'σου'), ('σου', 'ξανατυχε'), ('ξανατυχε', 'να'), ('να', 'σκεφτεσαι'), ('σκεφτεσαι', 'λλιο'), ('λλιο', 'καχυποπτα'), ('καχυποπτα', 'αμα'), ('αμα', 'πονεις'), ('πονεις', 'με'), ('με', 'τοσα'), ('τοσα', 'λεφτα'), ('λεφτα', 'που'), ('που', 'μαζευκουνται'), ('μαζευκουνται', 'θα'), ('θα', 'ηθελα'), ('ηθελα', 'να'), ('να', 'γινουνταν'), ('γινουνταν', 'καλλυττερα'), ('καλλυττερα', 'πραματα')], [('να', 'σαι'), ('σαι', 'καλα'), ('καλα', 'παντα'), ('παντα', 'τζιαι'), ('τζιαι', 'μπραβο'), ('μπραβο', 'στη'), ('στη', 'μανα'), ('μανα', 'σου'), ('σου', 'που'), ('που', 'ενικησεν')], [('εν', 'πολλα'), ('πολλα', 'δυσκολος'), ('δυσκολος', 'τουτος'), ('τουτος', 'ο'), ('ο', 'αγωνας')], [('ξερεις', 'τα'), ('τα', 'τζιαι'), ('τζιαι', 'συ'), ('συ', 'που'), ('που', 'πρωτον'), ('πρωτον', 'σιεριν')], [('ξερεις', 'ο'), ('ο', 'καθενας'), ('καθενας', 'εσιει'), ('εσιει', 'τα'), ('τα', 'δικα'), ('δικα', 'του'), ('του', 'αλλα'), ('αλλα', 'τουτη'), ('τουτη', 'η'), ('η', 'αρρωσκεια'), ('αρρωσκεια', 'εφαεν'), ('εφαεν', 'πολλους'), ('πολλους', 'τζιαι'), ('τζιαι', 'επηρεν'), ('επηρεν', 'τους'), ('τους', 'μιτα'), ('μιτα', 'της')], [('ετο', 'μανα'), ('μανα', 'μου'), ('μου', 'εν'), ('εν', 'ενας'), ('ενας', 'κοσμος'), ('κοσμος', 'για'), ('για', 'αντρες'), ('αντρες', 'ευνουχιστε'), ('ευνουχιστε', 'τους'), ('τους', 'ουλλους'), ('ουλλους', 'τωρα'), ('τωρα', 'που'), ('που', 'το'), ('το', 'σκεφτομαι'), ('σκεφτομαι', 'μεν'), ('μεν', 'τους'), ('τους', 'ευνουχισετε'), ('ευνουχισετε', 'γιατι'), ('γιατι', 'καμνουν'), ('καμνουν', 'και'), ('και', 'ενα'), ('ενα', 'καλο')], [('εθκιαβασα', 'τζαι'), ('τζαι', 'εγω'), ('εγω', 'την'), ('την', 'ποιηση'), ('ποιηση', 'της'), ('της', 'μου'), ('μου', 'το'), ('το', 'εδειξαν'), ('εδειξαν', 'προχτες'), ('προχτες', 'νομιζω'), ('νομιζω', 'εν'), ('εν', 'εχω'), ('εχω', 'να'), ('να', 'σχολιασω'), ('σχολιασω', 'τιποτε')], [('εν', 'εσιει'), ('εσιει', 'μανα'), ('μανα', 'μου'), ('μου', 'σαν'), ('σαν', 'την'), ('την', 'λεμεσο'), ('λεμεσο', 'εχεις'), ('εχεις', 'και'), ('και', 'εσυ'), ('εσυ', 'το'), ('το', 'δικαιο'), ('δικαιο', 'σου'), ('σου', 'λιο'), ('λιο', 'τατσιλλικκι'), ('τατσιλλικκι', 'εχουμε'), ('εχουμε', 'το')], [('εν', 'τουτο'), ('τουτο', 'που'), ('που', 'λαλουμε'), ('λαλουμε', 'καμε'), ('καμε', 'μου'), ('μου', 'τοσο'), ('τοσο', 'τζαι'), ('τζαι', 'θα'), ('θα', 'σου'), ('σου', 'καμω'), ('καμω', 'αλλο'), ('αλλο', 'τοσο')], [('ασε', 'του'), ('του', 'πατας'), ('πατας', 'για'), ('για', 'να'), ('να', 'προλαβεις'), ('προλαβεις', 'να'), ('να', 'μπεις'), ('μπεις', 'τζαι'), ('τζαι', 'ερκεται'), ('ερκεται', 'τζαι'), ('τζαι', 'ο'), ('ο', 'αλλος'), ('αλλος', 'ο'), ('ο', 'βλακας'), ('βλακας', 'τζαι'), ('τζαι', 'πατα'), ('πατα', 'του'), ('του', 'παραπανω'), ('παραπανω', 'για'), ('για', 'να'), ('να', 'σε'), ('σε', 'σκοτωσει'), ('σκοτωσει', 'σιουρα'), ('σιουρα', 'πραματα')], [('παντως', 'σιουρα'), ('σιουρα', 'τουτα'), ('τουτα', 'ουλλα'), ('ουλλα', 'τα'), ('τα', 'χορτα'), ('χορτα', 'που'), ('που', 'λαλεις'), ('λαλεις', 'πιο'), ('πιο', 'πανω'), ('πανω', 'πολλα'), ('πολλα', 'λιοι'), ('λιοι', 'εως'), ('εως', 'κανενας'), ('κανενας', 'αλλος'), ('αλλος', 'λαος'), ('λαος', 'δεν'), ('δεν', 'τα'), ('τα', 'τρωει')], [('εν', 'ανθρωποι'), ('ανθρωποι', 'τζιαι'), ('τζιαι', 'τουτοι'), ('τουτοι', 'ποιος'), ('ποιος', 'ξερει'), ('ξερει', 'ποιες'), ('ποιες', 'δυσκολιες'), ('δυσκολιες', 'τζιαι'), ('τζιαι', 'τι'), ('τι', 'τραυματα'), ('τραυματα', 'κουβαλουν'), ('κουβαλουν', 'τουτες'), ('τουτες', 'οι'), ('οι', 'ψυσιες'), ('ψυσιες', 'τζιαι'), ('τζιαι', 'γιατι'), ('γιατι', 'υιοθετησαν'), ('υιοθετησαν', 'τουτο'), ('τουτο', 'τον'), ('τον', 'τροπο'), ('τροπο', 'ζωης')], [('μπορω', 'να'), ('να', 'συζητω'), ('συζητω', 'με'), ('με', 'τις'), ('τις', 'ωρες'), ('ωρες', 'για'), ('για', 'τουντο'), ('τουντο', 'θεμα'), ('θεμα', 'αλλα'), ('αλλα', 'να'), ('να', 'ξεπεφτουν'), ('ξεπεφτουν', 'τοσο'), ('τοσο', 'πολυ'), ('πολυ', 'καποιοι'), ('καποιοι', 'τζιαι'), ('τζιαι', 'να'), ('να', 'κρινουν'), ('κρινουν', 'να'), ('να', 'πειραζουν'), ('πειραζουν', 'να'), ('να', 'ενοχλουν'), ('ενοχλουν', 'τζιαι'), ('τζιαι', 'να'), ('να', 'ξετυμαζουν'), ('ξετυμαζουν', 'εν'), ('εν', 'το'), ('το', 'καρτερουσα')], [('ευχαριστω', 'σας'), ('σας', 'ουλλους'), ('ουλλους', 'παρα'), ('παρα', 'πολλα'), ('πολλα', 'για'), ('για', 'το'), ('το', 'ενδιαφερον')], [('σιερουμαι', 'που'), ('που', 'σας'), ('σας', 'αρεσε'), ('αρεσε', 'τζιαι'), ('τζιαι', 'σκεφτεστε'), ('σκεφτεστε', 'το'), ('το', 'ιδιο')], [('προσφατα', 'εβλεπα'), ('εβλεπα', 'τζαι'), ('τζαι', 'εγω'), ('εγω', 'φωτογραφιες'), ('φωτογραφιες', 'που'), ('που', 'το'), ('το', 'δημοτικο'), ('δημοτικο', 'μετα'), ('μετα', 'που'), ('που', 'παρα'), ('παρα', 'πολλα'), ('πολλα', 'χρονια')], [('μα', 'ιντα'), ('ιντα', 'ωραιο'), ('ωραιο', 'ποστ'), ('ποστ', 'τι'), ('τι', 'μας'), ('μας', 'εθυμισες'), ('εθυμισες', 'τωρα')], [('υπομονη', 'με'), ('με', 'τες'), ('τες', 'δουλειες'), ('δουλειες', 'σου'), ('σου', 'τζαι'), ('τζαι', 'γραφε'), ('γραφε', 'τους'), ('τους', 'κουτσομπολιες'), ('κουτσομπολιες', 'να'), ('να', 'μεν'), ('μεν', 'σε'), ('σε', 'κοφτει'), ('κοφτει', 'ακομα'), ('ακομα', 'τζαι'), ('τζαι', 'που'), ('που', 'τζεινους'), ('τζεινους', 'που'), ('που', 'τους'), ('τους', 'διουν'), ('διουν', 'σημασια'), ('σημασια', 'τζαι'), ('τζαι', 'πιστευκουν'), ('πιστευκουν', 'τους')], [('εσιει', 'κανεναν'), ('κανεναν', 'εξαμηνο'), ('εξαμηνο', 'που'), ('που', 'σε'), ('σε', 'ανακαλυψα'), ('ανακαλυψα', 'εν'), ('εν', 'η'), ('η', 'χαρα'), ('χαρα', 'μου'), ('μου', 'να'), ('να', 'σε'), ('σε', 'θκιαβαζω')], [('ακομα', 'βεβαια'), ('βεβαια', 'προσπαθω'), ('προσπαθω', 'να'), ('να', 'καταλαβω'), ('καταλαβω', 'τι'), ('τι', 'δουλεια'), ('δουλεια', 'καμνεις'), ('καμνεις', 'γιατι'), ('γιατι', 'που'), ('που', 'τη'), ('τη', 'μια'), ('μια', 'μιλας'), ('μιλας', 'για'), ('για', 'καλλιτεχνικα'), ('καλλιτεχνικα', 'πραματα'), ('πραματα', 'τζιαι'), ('τζιαι', 'που'), ('που', 'την'), ('την', 'αλλη'), ('αλλη', 'θκιαβαζω'), ('θκιαβαζω', 'κατι'), ('κατι', 'παλια'), ('παλια', 'ποστ'), ('ποστ', 'με'), ('με', 'ψυχολογιες'), ('ψυχολογιες', 'τζιαι'), ('τζιαι', 'κοπελλουθκια')], [('ενιγουει', 'ο,τι'), ('ο,τι', 'τζιαι'), ('τζιαι', 'αν'), ('αν', 'καμνεις'), ('καμνεις', 'στη'), ('στη', 'ζωη'), ('ζωη', 'σου'), ('σου', 'σημασια'), ('σημασια', 'εσιει'), ('εσιει', 'να'), ('να', 'το'), ('το', 'αγαπας')], [('εν', 'δαμε'), ('δαμε', 'που'), ('που', 'επρεπε'), ('επρεπε', 'να'), ('να', 'γραψεις'), ('γραψεις', 'το'), ('το', 'κειμενο'), ('κειμενο', 'με'), ('με', 'το'), ('το', 'σιεσιμο'), ('σιεσιμο', 'τζιαι'), ('τζιαι', 'το'), ('το', 'καζανακι'), ('καζανακι', 'εσιει'), ('εσιει', 'μιαν'), ('μιαν', 'ηλιθια'), ('ηλιθια', 'που'), ('που', 'ξερω'), ('ξερω', 'που'), ('που', 'επειδη'), ('επειδη', 'εσπουδασεν'), ('εσπουδασεν', 'στο'), ('στο', 'εξωτερικο'), ('εξωτερικο', 'λαλει'), ('λαλει', 'οτι'), ('οτι', 'οι'), ('οι', 'αλλοι'), ('αλλοι', 'ουλλοι'), ('ουλλοι', 'εν'), ('εν', 'πρεπει'), ('πρεπει', 'να'), ('να', 'θεωρουνται'), ('θεωρουνται', 'καλλιτεχνες'), ('καλλιτεχνες', 'τζιαι'), ('τζιαι', 'κακοφημει'), ('κακοφημει', 'τους'), ('τους', 'εχθρους'), ('εχθρους', 'της'), ('της', 'τζιαι'), ('τζιαι', 'υποτιμα'), ('υποτιμα', 'τους')], [('εν', 'κατι'), ('κατι', 'καλλιτεχνες'), ('καλλιτεχνες', 'που'), ('που', 'επηαν'), ('επηαν', 'στο'), ('στο', 'εξωτερικον'), ('εξωτερικον', 'τζιαι'), ('τζιαι', 'εψηλωσεν'), ('εψηλωσεν', 'η'), ('η', 'μουττη'), ('μουττη', 'τους')], [('μεν', 'σου'), ('σου', 'πω'), ('πω', 'εν'), ('εν', 'η'), ('η', 'μονη'), ('μονη', 'η'), ('η', 'συγκεκριμενη'), ('συγκεκριμενη', 'που'), ('που', 'ξερω'), ('ξερω', 'τζιαι'), ('τζιαι', 'πρεπει'), ('πρεπει', 'να'), ('να', 'ντρεπεται'), ('ντρεπεται', 'που'), ('που', 'θεωρει'), ('θεωρει', 'τον'), ('τον', 'εαυτο'), ('εαυτο', 'της'), ('της', 'καλλιτεχνη'), ('καλλιτεχνη', 'ξερει'), ('ξερει', 'τα'), ('τα', 'ουλλα'), ('ουλλα', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'θελει'), ('θελει', 'να'), ('να', 'ερτει'), ('ερτει', 'κυπρο'), ('κυπρο', 'για'), ('για', 'να'), ('να', 'καμει'), ('καμει', 'καριερα')], [('τουτοι', 'κρατουν'), ('κρατουν', 'οι'), ('οι', 'ιδιοι'), ('ιδιοι', 'τον'), ('τον', 'τοπο'), ('τοπο', 'τους'), ('τους', 'πισω'), ('πισω', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'πραγματικα'), ('πραγματικα', 'ψυχοπαθεις'), ('ψυχοπαθεις', 'τζιαι'), ('τζιαι', 'πρεπει'), ('πρεπει', 'να'), ('να', 'κλειστουν'), ('κλειστουν', 'σε'), ('σε', 'ιδρυμα')], [('εν', 'θελω'), ('θελω', 'να'), ('να', 'καταγγειλω'), ('καταγγειλω', 'κανεναν'), ('κανεναν', 'αν'), ('αν', 'και'), ('και', 'τους'), ('τους', 'αξιζεν'), ('αξιζεν', 'ετσι'), ('ετσι', 'πραμα'), ('πραμα', 'εν'), ('εν', 'εξαναζησα')], [('μην', 'νομιζεις'), ('νομιζεις', 'πως'), ('πως', 'θα'), ('θα', 'κερδιζα'), ('κερδιζα', 'οτιδηποτε'), ('οτιδηποτε', 'οταν'), ('οταν', 'οι'), ('οι', 'σκαταες'), ('σκαταες', 'εχουν'), ('εχουν', 'μεσο'), ('μεσο', 'αγωνιζεσαι'), ('αγωνιζεσαι', 'αδικα'), ('αδικα', 'τζιαι'), ('τζιαι', 'εν'), ('εν', 'εχω'), ('εχω', 'χρονο'), ('χρονο', 'για'), ('για', 'πεταμα')], [('μονο', 'τζιαι'), ('τζιαι', 'μονο'), ('μονο', 'που'), ('που', 'εκαμαν'), ('εκαμαν', 'ετσι'), ('ετσι', 'ουλλοι'), ('ουλλοι', 'που'), ('που', 'γυρω'), ('γυρω', 'οι'), ('οι', 'μονο'), ('μονο', 'οι'), ('οι', 'νοσοκομοι'), ('νοσοκομοι', 'τζιαι'), ('τζιαι', 'οι'), ('οι', 'μπατσοι'), ('μπατσοι', 'αποδυκνυει'), ('αποδυκνυει', 'πως'), ('πως', 'εν'), ('εν', 'ουλλα'), ('ουλλα', 'που'), ('που', 'την'), ('την', 'κοινωνια'), ('κοινωνια', 'που'), ('που', 'ξεκινουν')], [('εσκεφτηκες', 'οι'), ('οι', 'πολλα'), ('πολλα', 'οι'), ('οι', 'βαρετοι'), ('βαρετοι', 'οι'), ('οι', 'ταξιτζιες'), ('ταξιτζιες', 'που'), ('που', 'αφηνουν'), ('αφηνουν', 'ενα'), ('ενα', 'νιχουι'), ('νιχουι', 'πας'), ('πας', 'το'), ('το', 'δακτυλο'), ('δακτυλο', 'το'), ('το', 'μιτσι'), ('μιτσι', 'ξερεις'), ('ξερεις', 'τζεινο'), ('τζεινο', 'το'), ('το', 'νυχουι'), ('νυχουι', 'που'), ('που', 'καθαριζει'), ('καθαριζει', 'τα'), ('τα', 'αφτια'), ('αφτια', 'τζαι'), ('τζαι', 'τη'), ('τη', 'μουτη')], [('οι', 'νεραιδες'), ('νεραιδες', 'ενναιν'), ('ενναιν', 'μεσα'), ('μεσα', 'εσυ'), ('εσυ', 'δικαιουσαι'), ('δικαιουσαι', 'εν'), ('εν', 'απλα'), ('απλα', 'μια'), ('μια', 'αποψη')], [('για', 'να'), ('να', 'σου'), ('σου', 'περασει'), ('περασει', 'λλιο'), ('λλιο', 'η'), ('η', 'καουρα'), ('καουρα', 'παντως'), ('παντως', 'προτεινω'), ('προτεινω', 'να'), ('να', 'θκιαβασεις'), ('θκιαβασεις', 'τους'), ('τους', 'θκυο'), ('θκυο', 'τομους'), ('τομους', 'με'), ('με', 'ιστοριες')], [('ηβρα', 'τουντην'), ('τουντην', 'φιλεναδα'), ('φιλεναδα', 'που'), ('που', 'λαλεις')], [('εν', 'τζιαι'), ('τζιαι', 'καρτερω'), ('καρτερω', 'τα'), ('τα', 'ουλλα'), ('ουλλα', 'που'), ('που', 'τον'), ('τον', 'θεο'), ('θεο', 'κινω'), ('κινω', 'τζιαι'), ('τζιαι', 'τα'), ('τα', 'ποθκια'), ('ποθκια', 'μου'), ('μου', 'τζιαι'), ('τζιαι', 'τα'), ('τα', 'σιερκα'), ('σιερκα', 'μου')], [('αλλα', 'εσιει'), ('εσιει', 'πραματα'), ('πραματα', 'που'), ('που', 'εν'), ('εν', 'εξαρτουνται'), ('εξαρτουνται', 'μονο'), ('μονο', 'που'), ('που', 'τες'), ('τες', 'κινησεις'), ('κινησεις', 'μου')], [('οσο', 'για'), ('για', 'το'), ('το', 'στομασι'), ('στομασι', 'εν'), ('εν', 'καταλαβει'), ('καταλαβει', 'τιποτε'), ('τιποτε', 'τωρα'), ('τωρα', 'που'), ('που', 'την'), ('την', 'πολλη'), ('πολλη', 'σουβλα'), ('σουβλα', 'τζιαι'), ('τζιαι', 'την'), ('την', 'πολλη'), ('πολλη', 'παττιχα')], [('βρισκω', 'σε'), ('σε', 'πολλα'), ('πολλα', 'συνειδητοποιημενη'), ('συνειδητοποιημενη', 'για'), ('για', 'το'), ('το', 'θεμα'), ('θεμα', 'τζιαι'), ('τζιαι', 'αφου'), ('αφου', 'ξερεις'), ('ξερεις', 'γιατι'), ('γιατι', 'ξεκινησες'), ('ξεκινησες', 'ξερεις'), ('ξερεις', 'γιατι'), ('γιατι', 'εφυες'), ('εφυες', 'τζιαι'), ('τζιαι', 'εισαι'), ('εισαι', 'αποφασισμενη'), ('αποφασισμενη', 'πιστευκω'), ('πιστευκω', 'θα'), ('θα', 'το'), ('το', 'προσπερασεις')], [('εισαι', 'μια'), ('μια', 'λεβεντια'), ('λεβεντια', 'τζιαι'), ('τζιαι', 'ενα'), ('ενα', 'πλασμα'), ('πλασμα', 'με'), ('με', 'ψυσιην')], [('εισαι', 'τσιακκος'), ('τσιακκος', 'γιατι'), ('γιατι', 'εσιεις'), ('εσιεις', 'αρχες'), ('αρχες', 'δυναμη'), ('δυναμη', 'να'), ('να', 'αντισταθεις')], [('ανοιξες', 'μας'), ('μας', 'για'), ('για', 'λιο'), ('λιο', 'τη'), ('τη', 'ψυσιη'), ('ψυσιη', 'σου'), ('σου', 'τζαι'), ('τζαι', 'ειδαμε'), ('ειδαμε', 'τζαι'), ('τζαι', 'μεις'), ('μεις', 'την'), ('την', 'αληθκεια')], [('πονουν', 'με'), ('με', 'τα'), ('τα', 'μμαθκια'), ('μμαθκια', 'μου'), ('μου', 'αμαν'), ('αμαν', 'δκιαβαζω')], [('εκατσα', 'μιαν'), ('μιαν', 'εφτομαδα'), ('εφτομαδα', 'τζιαι'), ('τζιαι', 'εμαθα'), ('εμαθα', 'να'), ('να', 'γραφω'), ('γραφω', 'στα'), ('στα', 'ελληνικα'), ('ελληνικα', 'επειδη'), ('επειδη', 'μου'), ('μου', 'το'), ('το', 'εζητησετε'), ('εζητησετε', 'τζιαι'), ('τζιαι', 'επειδη'), ('επειδη', 'πιστευκω'), ('πιστευκω', 'εν'), ('εν', 'καλλυττερα')], [('νιωθω', 'περιφανη'), ('περιφανη', 'γιατι'), ('γιατι', 'εν'), ('εν', 'εξαναγραψα'), ('εξαναγραψα', 'ποττε'), ('ποττε', 'μου'), ('μου', 'ελληνικα')], [('μα', 'τι'), ('τι', 'ομορφα'), ('ομορφα', 'που'), ('που', 'ενουν'), ('ενουν', 'μανα'), ('μανα', 'μου'), ('μου', 'τα')], [('πε', 'μου'), ('μου', 'τουτα'), ('τουτα', 'ουλλα'), ('ουλλα', 'που'), ('που', 'σε'), ('σε', 'νευριαζουν'), ('νευριαζουν', 'εν'), ('εν', 'στην'), ('στην', 'κυπρον'), ('κυπρον', 'που'), ('που', 'τα'), ('τα', 'θωρεις'), ('θωρεις', 'η'), ('η', 'εν'), ('εν', 'φαινομενον'), ('φαινομενον', 'στον'), ('στον', 'τοπον'), ('τοπον', 'που'), ('που', 'εισαι'), ('εισαι', 'τωρα')], [('αν', 'δεν'), ('δεν', 'εθκιαβαζεν'), ('εθκιαβαζεν', 'η'), ('η', 'οικουμενη'), ('οικουμενη', 'οτι'), ('οτι', 'σου'), ('σου', 'λαλω'), ('λαλω', 'δαμαι'), ('δαμαι', 'ηταν'), ('ηταν', 'να'), ('να', 'σου'), ('σου', 'πω'), ('πω', 'διαφορα')], [('εφαεν', 'τους'), ('τους', 'ο'), ('ο', 'ηπιος'), ('ηπιος', 'ιμπεριαλισμος'), ('ιμπεριαλισμος', 'που'), ('που', 'ερκεται'), ('ερκεται', 'εν'), ('εν', 'ειδη'), ('ειδη', 'πολιτισμου'), ('πολιτισμου', 'λεμεν'), ('λεμεν', 'τωρα')], [('εκτος', 'που'), ('που', 'τον'), ('τον', 'ταταν'), ('ταταν', 'του'), ('του', 'δεν'), ('δεν', 'τον'), ('τον', 'υποστηριζουν'), ('υποστηριζουν', 'καν'), ('καν', 'οι'), ('οι', 'φιλελευθεροι')], [('τον', 'τζαιρον'), ('τζαιρον', 'που'), ('που', 'ηταν'), ('ηταν', 'πρωταθλητης'), ('πρωταθλητης', 'του'), ('του', 'μπατμιττον'), ('μπατμιττον', 'οσον'), ('οσον', 'τζαιρον'), ('τζαιρον', 'ηταν'), ('ηταν', 'πρωταθλητης'), ('πρωταθλητης', 'ηταν'), ('ηταν', 'του'), ('του', 'ηθους')], [('ρε', 'φιλε'), ('φιλε', 'λες'), ('λες', 'τζι'), ('τζι', 'εν'), ('εν', 'λλιοι'), ('λλιοι', 'που'), ('που', 'εθησαυρισαν'), ('εθησαυρισαν', 'πανω'), ('πανω', 'στον'), ('στον', 'πονο'), ('πονο', 'του'), ('του', 'προσφυγα'), ('προσφυγα', 'τζιαι'), ('τζιαι', 'του'), ('του', 'συγγενη'), ('συγγενη', 'του'), ('του', 'αγνοουμενου')], [('ετυχε', 'μου'), ('μου', 'εχτες'), ('εχτες', 'ειπα'), ('ειπα', 'φευκω'), ('φευκω', 'γεια'), ('γεια', 'σας'), ('σας', 'και'), ('και', 'τους'), ('τους', 'αφησα')], [('εν', 'θα'), ('θα', 'το'), ('το', 'στηριξω'), ('στηριξω', 'αλλα'), ('αλλα', 'εν'), ('εν', 'τζαι'), ('τζαι', 'επεσεν'), ('επεσεν', 'που'), ('που', 'τον'), ('τον', 'ουρανο'), ('ουρανο', 'ουτε'), ('ουτε', 'εχει'), ('εχει', 'την'), ('την', 'ιδια'), ('ιδια', 'βαση'), ('βαση', 'με'), ('με', 'το'), ('το', 'ειμαι'), ('ειμαι', 'το'), ('το', 'ελαμ'), ('ελαμ', 'τζαι'), ('τζαι', 'δερνω'), ('δερνω', 'μαυρους')], [('πιστεφκουν', 'οτι'), ('οτι', 'πιανουν'), ('πιανουν', 'τες'), ('τες', 'δουλειες'), ('δουλειες', 'πιστεφκουν'), ('πιστεφκουν', 'οτι'), ('οτι', 'ειναι'), ('ειναι', 'εθνοτικα'), ('εθνοτικα', 'ανωτεροι')], [('εν', 'να'), ('να', 'σταματησει'), ('σταματησει', 'η'), ('η', 'αριστερα'), ('αριστερα', 'πκιον'), ('πκιον', 'οξα'), ('οξα', 'στο'), ('στο', 'ιραν'), ('ιραν', 'εν'), ('εν', 'κακοι'), ('κακοι', 'οι'), ('οι', 'ισλαμιστες'), ('ισλαμιστες', 'αλλα'), ('αλλα', 'εν'), ('εν', 'καλη'), ('καλη', 'η'), ('η', 'σαουδικη'), ('σαουδικη', 'αραβια')], [('να', 'μου'), ('μου', 'επιτρεψεις'), ('επιτρεψεις', 'λλιον'), ('λλιον', 'σκληρη'), ('σκληρη', 'κριτικη'), ('κριτικη', 'διοτι'), ('διοτι', 'οποιος'), ('οποιος', 'εδεχτηκεν'), ('εδεχτηκεν', 'βοηθεια'), ('βοηθεια', 'για'), ('για', 'να'), ('να', 'στρωσει'), ('στρωσει', 'μιαν'), ('μιαν', 'εξεγερση'), ('εξεγερση', 'που'), ('που', 'το'), ('το', 'κεφαλαιο'), ('κεφαλαιο', 'στο'), ('στο', 'τελος'), ('τελος', 'εχασεν')], [('εν', 'τζαι'), ('τζαι', 'εκεδρισεν'), ('εκεδρισεν', 'μανια'), ('μανια', 'ρε'), ('ρε', 'που'), ('που', 'την'), ('την', 'εσσιετε')], [('τι', 'εκερδισα'), ('εκερδισα', 'που'), ('που', 'την'), ('την', 'επιλογη'), ('επιλογη', 'μου'), ('μου', 'να'), ('να', 'ανοικω'), ('ανοικω', 'τζιαμε')], [('δεν', 'βγαινει'), ('βγαινει', 'οτι'), ('οτι', 'μας'), ('μας', 'εκαμαν'), ('εκαμαν', 'χατηρκα')], [('τι', 'εκερδισαμεν'), ('εκερδισαμεν', 'που'), ('που', 'τους'), ('τους', 'αδεσμευτους')], [('στελιο', 'εναν'), ('εναν', 'σεντονιν'), ('σεντονιν', 'σαν'), ('σαν', 'τον'), ('τον', 'φρεσκον'), ('φρεσκον', 'τον'), ('τον', 'αεραν')], [('η', 'αληθκεια'), ('αληθκεια', 'δεν'), ('δεν', 'εκαταλαβα'), ('εκαταλαβα', 'πως'), ('πως', 'οι'), ('οι', 'πομπες'), ('πομπες', 'του'), ('του', 'γριβα'), ('γριβα', 'σιουρα'), ('σιουρα', 'αποδυκνειουν'), ('αποδυκνειουν', 'σχεδιο'), ('σχεδιο', 'διχοτομισης')], [('τη', 'ζυριχη'), ('ζυριχη', 'δεν'), ('δεν', 'τη'), ('τη', 'καμνω'), ('καμνω', 'φοινιτζιαν')], [('τζαι', 'να'), ('να', 'σιεις'), ('σιεις', 'παντα'), ('παντα', 'υποψιν'), ('υποψιν', 'σου'), ('σου', 'οτι'), ('οτι', 'το'), ('το', 'να'), ('να', 'εσσιεις'), ('εσσιεις', 'μιαν'), ('μιαν', 'αλλαγη'), ('αλλαγη', 'θεωρητικα'), ('θεωρητικα', 'στο'), ('στο', 'νου'), ('νου', 'εν'), ('εν', 'τζαι'), ('τζαι', 'σημαινει'), ('σημαινει', 'οτι'), ('οτι', 'εννα'), ('εννα', 'σου'), ('σου', 'φκει'), ('φκει', 'τζαι'), ('τζαι', 'στη'), ('στη', 'πραξη'), ('πραξη', 'απαραιτητα')], [('οπως', 'λαλεις'), ('λαλεις', 'να'), ('να', 'μεν'), ('μεν', 'σουζουμεν'), ('σουζουμεν', 'τα'), ('τα', 'ποθκια'), ('ποθκια', 'μας'), ('μας', 'εν'), ('εν', 'ιχρειαζεται')], [('οπως', 'τζαι'), ('τζαι', 'να'), ('να', 'σσιει'), ('σσιει', 'οι'), ('οι', 'πως'), ('πως', 'απλα'), ('απλα', 'ησουν'), ('ησουν', 'εμπορας'), ('εμπορας', 'οι'), ('οι', 'σιορ')], [('καλαν', 'η'), ('η', 'ουσια'), ('ουσια', 'ειναι'), ('ειναι', 'ποιος'), ('ποιος', 'εν'), ('εν', 'να'), ('να', 'την'), ('την', 'κατσει'), ('κατσει', 'του'), ('του', 'αλλου'), ('αλλου', 'τζιαι'), ('τζιαι', 'να'), ('να', 'φανει'), ('φανει', 'πιο'), ('πιο', 'εξυπνος')], [('οξα', 'οτι'), ('οτι', 'βασικα'), ('βασικα', 'τουτη'), ('τουτη', 'η'), ('η', 'κυβερνηση'), ('κυβερνηση', 'χειριζεται'), ('χειριζεται', 'το'), ('το', 'θεμαν'), ('θεμαν', 'των'), ('των', 'πορων'), ('πορων', 'με'), ('με', 'στοχον'), ('στοχον', 'το'), ('το', 'κοινον'), ('κοινον', 'σσυφφερον')], [('ειναι', 'αληθκεια'), ('αληθκεια', 'οτι'), ('οτι', 'πριν'), ('πριν', 'να'), ('να', 'καλλιτζιεψουν'), ('καλλιτζιεψουν', 'σουζουν'), ('σουζουν', 'ουλλοι'), ('ουλλοι', 'τα'), ('τα', 'ποθκια'), ('ποθκια', 'τους')], [('πολλα', 'φοουμαι'), ('φοουμαι', 'πως'), ('πως', 'το'), ('το', 'ακελ'), ('ακελ', 'κατερριψε'), ('κατερριψε', 'και'), ('και', 'αυτο')], [('απορια', 'της'), ('της', 'χωρκατισσας'), ('χωρκατισσας', 'εκαμναν'), ('εκαμναν', 'κουπες'), ('κουπες', 'μες'), ('μες', 'το'), ('το', 'προεδρικο'), ('προεδρικο', 'εν'), ('εν', 'ηξεραμεν'), ('ηξεραμεν', 'λαλουσιν'), ('λαλουσιν', 'τωρα')], [('ενι', 'ξερω'), ('ξερω', 'εν'), ('εν', 'ειμαι'), ('ειμαι', 'σιουρη'), ('σιουρη', 'ακουσα'), ('ακουσα', 'πως'), ('πως', 'ηδη'), ('ηδη', 'εφκηκαν')], [('βεβαια', 'οπως'), ('οπως', 'εν'), ('εν', 'τα'), ('τα', 'πραματα'), ('πραματα', 'στην'), ('στην', 'ελλαδα'), ('ελλαδα', 'εσιει'), ('εσιει', 'πολλους'), ('πολλους', 'που'), ('που', 'θα'), ('θα', 'πασιν'), ('πασιν', 'με'), ('με', 'τουτα'), ('τουτα', 'τα'), ('τα', 'λεφτα'), ('λεφτα', 'να'), ('να', 'δουλεψουν')], [('κρατω', 'μιτσην'), ('μιτσην', 'καλαθιν'), ('καλαθιν', 'αν'), ('αν', 'οντως'), ('οντως', 'το'), ('το', 'συστημα'), ('συστημα', 'φκαινει'), ('φκαινει', 'μονο'), ('μονο', 'με'), ('με', 'τες'), ('τες', 'αποκοπες')], [('η', 'αληθκεια'), ('αληθκεια', 'ειναι'), ('ειναι', 'οτι'), ('οτι', 'πολλα'), ('πολλα', 'που'), ('που', 'τα'), ('τα', 'θεματα'), ('θεματα', 'που'), ('που', 'εθιξα'), ('εθιξα', 'σιγουρα'), ('σιγουρα', 'ακομα'), ('ακομα', 'ψαχνω'), ('ψαχνω', 'τα')], [('πολλα', 'πραματα'), ('πραματα', 'που'), ('που', 'σημερα'), ('σημερα', 'θεωρουμεν'), ('θεωρουμεν', 'δεδομενα'), ('δεδομενα', 'πιθανον'), ('πιθανον', 'να'), ('να', 'μην'), ('μην', 'ειναι'), ('ειναι', 'πολλα'), ('πολλα', 'συντομα')], [('δες', 'μονον'), ('μονον', 'τες'), ('τες', 'φοβερες'), ('φοβερες', 'αλλαγες'), ('αλλαγες', 'στη'), ('στη', 'ζωη'), ('ζωη', 'του'), ('του', 'αθρωπου'), ('αθρωπου', 'σε'), ('σε', 'μιαν'), ('μιαν', 'γενιαν'), ('γενιαν', 'τι'), ('τι', 'εζησαν'), ('εζησαν', 'οι'), ('οι', 'παππουδες'), ('παππουδες', 'μας'), ('μας', 'στην'), ('στην', 'τεχνολογιαν')], [('φοουμαι', 'με'), ('με', 'νακκον'), ('νακκον', 'μερικες'), ('μερικες', 'φορες'), ('φορες', 'η'), ('η', 'αληθεια'), ('αληθεια', 'γιατι'), ('γιατι', 'νομιζω'), ('νομιζω', 'αν'), ('αν', 'για'), ('για', 'καποιο'), ('καποιο', 'λογο'), ('λογο', 'εν'), ('εν', 'τα'), ('τα', 'καταφερνα'), ('καταφερνα', 'τζαι'), ('τζαι', 'εκαμναν'), ('εκαμναν', 'μου'), ('μου', 'τουτα'), ('τουτα', 'που'), ('που', 'περιγραφεις'), ('περιγραφεις', 'ηταν'), ('ηταν', 'να'), ('να', 'σπαζω'), ('σπαζω', 'μπουκαλες'), ('μπουκαλες', 'πας'), ('πας', 'στες'), ('στες', 'τζεφαλες'), ('τζεφαλες', 'τους'), ('τους', 'να'), ('να', 'σπαζω'), ('σπαζω', 'τα'), ('τα', 'πραματα'), ('πραματα', 'τους'), ('τους', 'που'), ('που', 'εν'), ('εν', 'καπου'), ('καπου', 'που'), ('που', 'εν'), ('εν', 'πρεπει'), ('πρεπει', 'να'), ('να', 'τα'), ('τα', 'πετασσω'), ('πετασσω', 'να'), ('να', 'παιρνω'), ('παιρνω', 'ξιμαρισμενα'), ('ξιμαρισμενα', 'πραματα'), ('πραματα', 'μες'), ('μες', 'στα'), ('στα', 'δωματια'), ('δωματια', 'τους')], [('εννα', 'καμω'), ('καμω', 'την'), ('την', 'πελλη'), ('πελλη', 'τζαι'), ('τζαι', 'σικκιμε')], [('μεν', 'το'), ('το', 'γενικευκεις'), ('γενικευκεις', 'τα'), ('τα', 'κινεζουθκια'), ('κινεζουθκια', 'μου'), ('μου', 'εμενα'), ('εμενα', 'εν'), ('εν', 'πολλα'), ('πολλα', 'καλα')], [('ιντα', 'μπου'), ('μπου', 'θωρεις'), ('θωρεις', 'ρε'), ('ρε', 'καλαμαρα'), ('καλαμαρα', 'το'), ('το', 'τραγικο'), ('τραγικο', 'ειναι'), ('ειναι', 'πως'), ('πως', 'το'), ('το', 'ξερουμεν'), ('ξερουμεν', 'τζιολας')], [('μπορουμεν', 'να'), ('να', 'το'), ('το', 'συζητουμεν'), ('συζητουμεν', 'για'), ('για', 'ωρες'), ('ωρες', 'αλλα'), ('αλλα', 'σιγουρος'), ('σιγουρος', 'δεν'), ('δεν', 'μπορει'), ('μπορει', 'να'), ('να', 'ειναι'), ('ειναι', 'κανενας')], [('εγω', 'δεν'), ('δεν', 'ηβρα'), ('ηβρα', 'κανεναν'), ('κανεναν', 'να'), ('να', 'μου'), ('μου', 'την'), ('την', 'εξηγησει'), ('εξηγησει', 'αμαν'), ('αμαν', 'φτασουμεν'), ('φτασουμεν', 'σε'), ('σε', 'τουτον'), ('τουτον', 'το'), ('το', 'θεμαν'), ('θεμαν', 'μασουν'), ('μασουν', 'τα'), ('τα', 'λογια'), ('λογια', 'τους'), ('τους', 'ππεφτει'), ('ππεφτει', 'τζιαι'), ('τζιαι', 'λλιον'), ('λλιον', 'αυτοκριτικη'), ('αυτοκριτικη', 'αλλα'), ('αλλα', 'δεν'), ('δεν', 'καταλαβω'), ('καταλαβω', 'τι'), ('τι', 'μου'), ('μου', 'λαλουσιν')], [('εν', 'δυσκολον'), ('δυσκολον', 'τζιαι'), ('τζιαι', 'επικινδυνον'), ('επικινδυνον', 'παντως'), ('παντως', 'να'), ('να', 'φκαλλουμεν'), ('φκαλλουμεν', 'τα'), ('τα', 'γεγονοτα'), ('γεγονοτα', 'οταν'), ('οταν', 'προχωρουμεν'), ('προχωρουμεν', 'με'), ('με', 'συμπερασματα'), ('συμπερασματα', 'οσα'), ('οσα', 'ντοκουμεντα'), ('ντοκουμεντα', 'τζιαι'), ('τζιαι', 'να'), ('να', 'συναξουμεν')], [('εφηρεν', 'μας'), ('μας', 'παλε'), ('παλε', 'ο'), ('ο', 'γιωρκος'), ('γιωρκος', 'κοφτει'), ('κοφτει', 'σε'), ('σε', 'οξα'), ('οξα', 'οι')], [('αν', 'το'), ('το', 'εφκαλεν'), ('εφκαλεν', 'στειλε'), ('στειλε', 'μου'), ('μου', 'αστειεφκω')], [('ο', 'παππους'), ('παππους', 'μου'), ('μου', 'τζ'), ('τζ', 'ο'), ('ο', 'τζυρης'), ('τζυρης', 'μου'), ('μου', 'ηταν'), ('ηταν', 'αριστεροι')], [('κινα', 'μου'), ('μου', 'την'), ('την', 'περιεργια'), ('περιεργια', 'να'), ('να', 'θκιαβασω'), ('θκιαβασω', 'το'), ('το', 'βιβλιο'), ('βιβλιο', 'γιατι'), ('γιατι', 'εχει'), ('εχει', 'πολλα'), ('πολλα', 'πραματα'), ('πραματα', 'που'), ('που', 'δεν'), ('δεν', 'εμπορεσα'), ('εμπορεσα', 'ποττε'), ('ποττε', 'να'), ('να', 'καταλαβω')], [('εννεν', 'μονον'), ('μονον', 'ετσι'), ('ετσι', 'που'), ('που', 'εννα'), ('εννα', 'θεσεις'), ('θεσεις', 'το'), ('το', 'ερωτημαν')], [('εν', 'αθυμουμαι'), ('αθυμουμαι', 'ποιαν'), ('ποιαν', 'αγγυση'), ('αγγυση', 'που'), ('που', 'εδωκαν'), ('εδωκαν', 'ως'), ('ως', 'τα'), ('τα', 'σημερα'), ('σημερα', 'εσταθηκε'), ('εσταθηκε', 'κανενας'), ('κανενας', 'στο'), ('στο', 'λοο'), ('λοο', 'του')], [('ποσο', 'σσιειροτερα'), ('σσιειροτερα', 'εννα'), ('εννα', 'ναι'), ('ναι', 'δηλαδη')], [('τουτο', 'το'), ('το', 'πραμα'), ('πραμα', 'θεωρουμεν'), ('θεωρουμεν', 'οτι'), ('οτι', 'δουλεφκει'), ('δουλεφκει', 'τι'), ('τι', 'θα'), ('θα', 'επρεπε'), ('επρεπε', 'να'), ('να', 'γινει'), ('γινει', 'δηλαδη'), ('δηλαδη', 'για'), ('για', 'να'), ('να', 'μεν'), ('μεν', 'δουλεφκει')], [('εκαμες', 'να'), ('να', 'μετρουμε'), ('μετρουμε', 'τες'), ('τες', 'γραμμες'), ('γραμμες', 'για'), ('για', 'του'), ('του', 'κειμενου'), ('κειμενου', 'για'), ('για', 'να'), ('να', 'δουμε'), ('δουμε', 'τι'), ('τι', 'συμφωνας'), ('συμφωνας', 'πιο'), ('πιο', 'λιο'), ('λιο', 'τζαι'), ('τζαι', 'τι'), ('τι', 'συμφωνας'), ('συμφωνας', 'παραπανω')], [('οσο', 'πιο'), ('πιο', 'γληορα'), ('γληορα', 'φυει'), ('φυει', 'τουτη'), ('τουτη', 'η'), ('η', 'κυβερνηση'), ('κυβερνηση', 'τοσο'), ('τοσο', 'το'), ('το', 'καλυτερο')], [('ποιος', 'μας'), ('μας', 'ειπεν'), ('ειπεν', 'εμας'), ('εμας', 'πως'), ('πως', 'εννε'), ('εννε', 'καλη'), ('καλη', 'η'), ('η', 'λιτοτητα'), ('λιτοτητα', 'η'), ('η', 'επιαεν'), ('επιαεν', 'η'), ('η', 'εννοια'), ('εννοια', 'τους'), ('τους', 'κεφαλαιουχους'), ('κεφαλαιουχους', 'αν'), ('αν', 'εμεις'), ('εμεις', 'πεινουμεν'), ('πεινουμεν', 'αφου'), ('αφου', 'αλλοσπως'), ('αλλοσπως', 'κερδη'), ('κερδη', 'δεν'), ('δεν', 'φκαινουν')], [('υγειαν', 'τζιαι'), ('τζιαι', 'ευτυχιαν'), ('ευτυχιαν', 'σε'), ('σε', 'ουλλην'), ('ουλλην', 'την'), ('την', 'οικογενειαν'), ('οικογενειαν', 'τζιαι'), ('τζιαι', 'καλον'), ('καλον', 'κουραγιον'), ('κουραγιον', 'κατα'), ('κατα', 'τη'), ('τη', 'βρεφικην'), ('βρεφικην', 'τζιαι'), ('τζιαι', 'τη'), ('τη', 'νηπιακην'), ('νηπιακην', 'ηλικιαν')], [('παντα', 'να'), ('να', 'θυμασαι'), ('θυμασαι', 'μιτσσια'), ('μιτσσια', 'κοπελλουθκια'), ('κοπελλουθκια', 'μιτσσιες'), ('μιτσσιες', 'εγνοιες')], [('αφου', 'οι'), ('οι', 'μιτσιες'), ('μιτσιες', 'εγνοιες'), ('εγνοιες', 'διαρκουν'), ('διαρκουν', 'λιγο'), ('λιγο', 'θα'), ('θα', 'καμουμε'), ('καμουμε', 'οτι'), ('οτι', 'μπορουμε'), ('μπορουμε', 'να'), ('να', 'τες'), ('τες', 'απολαυσουμε')], [('οτι', 'καμνει'), ('καμνει', 'η'), ('η', 'κυβερνηση'), ('κυβερνηση', 'κουστιζει'), ('κουστιζει', 'μονον'), ('μονον', 'στους'), ('στους', 'δανειστες'), ('δανειστες', 'που'), ('που', 'κλωτσουν')], [('τζιαμαι', 'θα'), ('θα', 'δουμεν'), ('δουμεν', 'τες'), ('τες', 'αντοχες'), ('αντοχες', 'της'), ('της', 'πλατειας'), ('πλατειας', 'τες'), ('τες', 'αντοχες'), ('αντοχες', 'της'), ('της', 'κοινωνιας')], [('ελπιζω', 'να'), ('να', 'μεν'), ('μεν', 'παμεν'), ('παμεν', 'σε'), ('σε', 'απλα'), ('απλα', 'πραματα'), ('πραματα', 'μιαν'), ('μιαν', 'διχτατοριαν'), ('διχτατοριαν', 'για'), ('για', 'ουλλους')], [('ασ', 'ησκεφτουμεν'), ('ησκεφτουμεν', 'τζιαι'), ('τζιαι', 'για'), ('για', 'λλοου'), ('λλοου', 'μας'), ('μας', 'με'), ('με', 'τες'), ('τες', 'ομοσπονδιες')], [('τωρα', 'που'), ('που', 'εφααμεν'), ('εφααμεν', 'το'), ('το', 'δολομαν'), ('δολομαν', 'ηρτεν'), ('ηρτεν', 'η'), ('η', 'καψη'), ('καψη', 'του'), ('του', 'αντζιστριου')], [('το', 'παραθυρον'), ('παραθυρον', 'καποιας'), ('καποιας', 'δημοκρατιας'), ('δημοκρατιας', 'ισως'), ('ισως', 'ετελειωσεν'), ('ετελειωσεν', 'μακαρι'), ('μακαρι', 'να'), ('να', 'φτασουν'), ('φτασουν', 'τζιαι'), ('τζιαι', 'λλιον'), ('λλιον', 'τα'), ('τα', 'παιθκια'), ('παιθκια', 'μας'), ('μας', 'την'), ('την', 'δυνατοτηταν')], [('αυτον', 'που'), ('που', 'δεν'), ('δεν', 'εκαμεν'), ('εκαμεν', 'η'), ('η', 'μαλλον'), ('μαλλον', 'που'), ('που', 'το'), ('το', 'κεφαλαιον'), ('κεφαλαιον', 'του'), ('του', 'δεν'), ('δεν', 'τον'), ('τον', 'αφηκεν'), ('αφηκεν', 'να'), ('να', 'καμει'), ('καμει', 'εκαμεν'), ('εκαμεν', 'το'), ('το', 'ο'), ('ο', 'τσιπρας'), ('τσιπρας', 'που'), ('που', 'δεν'), ('δεν', 'ειχεν'), ('ειχεν', 'ακομα'), ('ακομα', 'κεφαλαιον'), ('κεφαλαιον', 'δικον'), ('δικον', 'του'), ('του', 'ετα')], [('δεν', 'ξερω'), ('ξερω', 'για'), ('για', 'ποιαν'), ('ποιαν', 'συζητησην'), ('συζητησην', 'λαλεις')], [('οποτε', 'με'), ('με', 'ερωτουσαν'), ('ερωτουσαν', 'επρεπεν'), ('επρεπεν', 'να'), ('να', 'καμω'), ('καμω', 'διατριβην'), ('διατριβην', 'για'), ('για', 'να'), ('να', 'τους'), ('τους', 'πω'), ('πω', 'τι'), ('τι', 'ηταν'), ('ηταν', 'τουτον'), ('τουτον', 'τζιαι'), ('τζιαι', 'παλε'), ('παλε', 'εν'), ('εν', 'εκαταλαβαιναν')], [('εφκαλλα', 'καλα'), ('καλα', 'λεφτα'), ('λεφτα', 'βεβαια'), ('βεβαια', 'αλλα'), ('αλλα', 'ειχα'), ('ειχα', 'τουτον'), ('τουτον', 'το'), ('το', 'ανικανοποιητον'), ('ανικανοποιητον', 'του'), ('του', 'οτι'), ('οτι', 'εν'), ('εν', 'εφκαλλα'), ('εφκαλλα', 'κατι'), ('κατι', 'χειροπιαστον')], [('τζιαι', 'ξερουν'), ('ξερουν', 'τζιαι'), ('τζιαι', 'οι'), ('οι', 'μιτσιοι'), ('μιτσιοι', 'ναμπου'), ('ναμπου', 'να'), ('να', 'πουν'), ('πουν', 'στο'), ('στο', 'σχολειον'), ('σχολειον', 'οταν'), ('οταν', 'τους'), ('τους', 'ρωτουν'), ('ρωτουν', 'τι'), ('τι', 'καμνει'), ('καμνει', 'ο'), ('ο', 'παπακης')], [('αρεσκει', 'μου'), ('μου', 'που'), ('που', 'λαλεις'), ('λαλεις', 'αρκετα'), ('αρκετα', 'πολυπλοκα'), ('πολυπλοκα', 'πραματα'), ('πραματα', 'αλλα'), ('αλλα', 'με'), ('με', 'απλη'), ('απλη', 'γλωσσαν')], [('εγω', 'εν'), ('εν', 'θωρω'), ('θωρω', 'ως'), ('ως', 'μονην'), ('μονην', 'ελπιδαν'), ('ελπιδαν', 'την'), ('την', 'αριστεραν')], [('στην', 'κυπρον'), ('κυπρον', 'εν'), ('εν', 'θκιο'), ('θκιο', 'τα'), ('τα', 'κρισιμα'), ('κρισιμα', 'ερωτηματα')], [('κατ', 'εμεναν'), ('εμεναν', 'η'), ('η', 'απαντηση'), ('απαντηση', 'στο'), ('στο', 'πρωτον'), ('πρωτον', 'ερωτημαν'), ('ερωτημαν', 'εν'), ('εν', 'προφανως'), ('προφανως', 'ναι')], [('εσιεις', 'καλην'), ('καλην', 'ψυσιην'), ('ψυσιην', 'γαμω'), ('γαμω', 'το'), ('το', 'στανιο'), ('στανιο', 'σου')], [('εναν', 'κειμενον'), ('κειμενον', 'σου'), ('σου', 'να'), ('να', 'θκαβαζει'), ('θκαβαζει', 'ο'), ('ο', 'αλλος'), ('αλλος', 'νωθει'), ('νωθει', 'το'), ('το', 'να'), ('να', 'τρεσιει'), ('τρεσιει', 'που'), ('που', 'τα'), ('τα', 'περιθωρια')], [('οι', 'διακρισεις'), ('διακρισεις', 'εφκηκαν'), ('εφκηκαν', 'μου'), ('μου', 'σε'), ('σε', 'καλον'), ('καλον', 'τελικα')], [('αμαν', 'τα'), ('τα', 'σκεφτουμαι'), ('σκεφτουμαι', 'τουτα'), ('τουτα', 'ουλλα'), ('ουλλα', 'πιαννει'), ('πιαννει', 'με'), ('με', 'η'), ('η', 'λυπη')], [('κοφκεται', 'μου'), ('μου', 'η'), ('η', 'ορεξη'), ('ορεξη', 'ακομα'), ('ακομα', 'τζιαι'), ('τζιαι', 'για'), ('για', 'τα'), ('τα', 'χαλλουμια'), ('χαλλουμια', 'τζιαι'), ('τζιαι', 'για'), ('για', 'τες'), ('τες', 'παττισιες')], [('εγω', 'μεταναστρια'), ('μεταναστρια', 'ειμαι'), ('ειμαι', 'οποτε'), ('οποτε', 'πολλα'), ('πολλα', 'πραματα'), ('πραματα', 'εν'), ('εν', 'ηξερω')], [('ειδικα', 'το'), ('το', 'χαλλουμιν'), ('χαλλουμιν', 'εν'), ('εν', 'αδυνατον'), ('αδυνατον', 'να'), ('να', 'το'), ('το', 'ευρεις'), ('ευρεις', 'οπως'), ('οπως', 'πρεπει')], [('ηβρα', 'τζι'), ('τζι', 'εγω'), ('εγω', 'εναν'), ('εναν', 'πεθαμενον'), ('πεθαμενον', 'μια'), ('μια', 'φορα'), ('φορα', 'στη'), ('στη', 'γαλλια')], [('εν', 'τον'), ('τον', 'εξερα'), ('εξερα', 'πρωτη'), ('πρωτη', 'φορα'), ('φορα', 'τον'), ('τον', 'εθωρουν'), ('εθωρουν', 'ηταν'), ('ηταν', 'μολις'), ('μολις', 'εβρεθηκα'), ('εβρεθηκα', 'στην'), ('στην', 'πολη')], [('εν', 'εχω'), ('εχω', 'κατι'), ('κατι', 'με'), ('με', 'τα'), ('τα', 'κατοικιδια'), ('κατοικιδια', 'ισα'), ('ισα', 'ισα'), ('ισα', 'αλλα'), ('αλλα', 'οι'), ('οι', 'να'), ('να', 'μεν'), ('μεν', 'γοραζεις'), ('γοραζεις', 'ενα'), ('ενα', 'πραμα'), ('πραμα', 'για'), ('για', 'τον'), ('τον', 'αλλον'), ('αλλον', 'αθρωπο'), ('αθρωπο', 'που'), ('που', 'πεινα'), ('πεινα', 'δεν'), ('δεν', 'νομιζω'), ('νομιζω', 'να'), ('να', 'εκαμα'), ('εκαμα', 'κατι'), ('κατι', 'εκτος'), ('εκτος', 'που'), ('που', 'να'), ('να', 'δουλεψω'), ('δουλεψω', 'πας'), ('πας', 'τες'), ('τες', 'ενοχες'), ('ενοχες', 'μου')], [('πολλα', 'καλη'), ('καλη', 'αναλυση'), ('αναλυση', 'ξερω'), ('ξερω', 'πολλα'), ('πολλα', 'καλα'), ('καλα', 'το'), ('το', 'τραουδιν'), ('τραουδιν', 'αλλα'), ('αλλα', 'πολλα'), ('πολλα', 'που'), ('που', 'τουτα'), ('τουτα', 'που'), ('που', 'εγραψες'), ('εγραψες', 'εν'), ('εν', 'τα'), ('τα', 'εσκεφτηκα')], [('συγγνωμην', 'που'), ('που', 'σας'), ('σας', 'επελλανα')], [('το', 'τραουδιν'), ('τραουδιν', 'εν'), ('εν', 'ωραιον'), ('ωραιον', 'τζιαι'), ('τζιαι', 'ο'), ('ο', 'αλκινοος'), ('αλκινοος', 'εν'), ('εν', 'εσιει'), ('εσιει', 'λαθος')], [('επηεν', 'εν'), ('εν', 'χαμαι'), ('χαμαι', 'ηδη'), ('ηδη', 'πε'), ('πε', 'μας'), ('μας', 'λια'), ('λια', 'για'), ('για', 'την'), ('την', 'τενεδο')], [('ηυρα', 'το'), ('το', 'μιαν'), ('μιαν', 'φορα'), ('φορα', 'πας'), ('πας', 'το'), ('το', 'μπαλκονι'), ('μπαλκονι', 'αποκλειεται'), ('αποκλειεται', 'να'), ('να', 'σου'), ('σου', 'αρεσκουν'), ('αρεσκουν', 'οι'), ('οι', 'καττοι')], [('τωρα', 'που'), ('που', 'τα'), ('τα', 'εδκιαβασα'), ('εδκιαβασα', 'τουτα'), ('τουτα', 'που'), ('που', 'εγραψες'), ('εγραψες', 'οπως'), ('οπως', 'τα'), ('τα', 'εγραψες'), ('εγραψες', 'αθθυμηθηκαν'), ('αθθυμηθηκαν', 'την'), ('την', 'πρωτη'), ('πρωτη', 'βολαν'), ('βολαν', 'που'), ('που', 'επηα'), ('επηα', 'τζιαμε')], [('τωρα', 'θωρω'), ('θωρω', 'οτι'), ('οτι', 'αμαν'), ('αμαν', 'τα'), ('τα', 'θεωρησεις'), ('θεωρησεις', 'δεδομενα'), ('δεδομενα', 'τα'), ('τα', 'θωρεις'), ('θωρεις', 'με'), ('με', 'την'), ('την', 'ιδιαν'), ('ιδιαν', 'μμαθκιαν')], [('εμας', 'δακατω'), ('δακατω', 'αμα'), ('αμα', 'σε'), ('σε', 'καλιει'), ('καλιει', 'ο'), ('ο', 'αλλος'), ('αλλος', 'πρεπει'), ('πρεπει', 'να'), ('να', 'παρεις'), ('παρεις', 'κατι'), ('κατι', 'αλλα'), ('αλλα', 'ως'), ('ως', 'τζιειαμαι')], [('τζιαι', 'για'), ('για', 'να'), ('να', 'παεις'), ('παεις', 'εσσω'), ('εσσω', 'του'), ('του', 'αλλου'), ('αλλου', 'πρεπει'), ('πρεπει', 'να'), ('να', 'πιασεις'), ('πιασεις', 'ραντεβου')], [('τζιερασμαν', 'εξω'), ('εξω', 'δεν'), ('δεν', 'τζιερνουν'), ('τζιερνουν', 'ιδιως'), ('ιδιως', 'αμαν'), ('αμαν', 'εν'), ('εν', 'περαν'), ('περαν', 'του'), ('του', 'καφε')], [('ατε', 'ατε'), ('ατε', 'αν'), ('αν', 'εχεις'), ('εχεις', 'πολλα'), ('πολλα', 'στενες'), ('στενες', 'σχεσεις'), ('σχεσεις', 'μοιραζεις'), ('μοιραζεις', 'τον'), ('τον', 'λοαρκασμον'), ('λοαρκασμον', 'για'), ('για', 'να'), ('να', 'μεν'), ('μεν', 'υπολογιζεις'), ('υπολογιζεις', 'το'), ('το', 'πιατον'), ('πιατον', 'του'), ('του', 'καθενου'), ('καθενου', 'ξηχωριστα')], [('μια', 'γνωστη'), ('γνωστη', 'μου'), ('μου', 'ηρτεν'), ('ηρτεν', 'να'), ('να', 'φαμεν'), ('φαμεν', 'τζιαι'), ('τζιαι', 'εφερεν'), ('εφερεν', 'κρασιν'), ('κρασιν', 'επειδη'), ('επειδη', 'εν'), ('εν', 'το'), ('το', 'ηπκιαμεν'), ('ηπκιαμεν', 'επηρεν'), ('επηρεν', 'το'), ('το', 'παλε')], [('η', 'μανα'), ('μανα', 'μου'), ('μου', 'παντως'), ('παντως', 'αμμαν'), ('αμμαν', 'επαιζεν'), ('επαιζεν', 'το'), ('το', 'κουδουνιν'), ('κουδουνιν', 'την'), ('την', 'ωραν'), ('ωραν', 'που'), ('που', 'ετρωαμεν'), ('ετρωαμεν', 'ελαλεν'), ('ελαλεν', 'μα'), ('μα', 'ποιος'), ('ποιος', 'εννα'), ('εννα', 'νει'), ('νει', 'ετσι'), ('ετσι', 'ωραν')], [('φερε', 'εναν'), ('εναν', 'πκιατον')], [('πολλα', 'ενδιαφερον'), ('ενδιαφερον', 'τουτον'), ('τουτον', 'που'), ('που', 'λαλεις'), ('λαλεις', 'εθκιεβασα'), ('εθκιεβασα', 'τζιαι'), ('τζιαι', 'γω'), ('γω', 'μιαν'), ('μιαν', 'εθνολογικην'), ('εθνολογικην', 'μελετην'), ('μελετην', 'για'), ('για', 'τον'), ('τον', 'δωρισμον')], [('αννα', 'μου'), ('μου', 'ευκαριστω'), ('ευκαριστω', 'σε'), ('σε', 'για'), ('για', 'την'), ('την', 'πιο'), ('πιο', 'θετικη'), ('θετικη', 'εικονα'), ('εικονα', 'που'), ('που', 'μου'), ('μου', 'εδωκες')], [('φωτογραφιζεται', 'με'), ('με', 'την'), ('την', 'χαρτωμενην'), ('χαρτωμενην', 'του'), ('του', 'αμπα'), ('αμπα', 'τζιαι'), ('τζιαι', 'νομισει'), ('νομισει', 'κανενας'), ('κανενας', 'οτι'), ('οτι', 'εν'), ('εν', 'ππουστης')], [('εν', 'πολλα'), ('πολλα', 'σωστον'), ('σωστον', 'τζιεινον'), ('τζιεινον', 'που'), ('που', 'λαλεις'), ('λαλεις', 'για'), ('για', 'τζιεινους'), ('τζιεινους', 'που'), ('που', 'γραφουν'), ('γραφουν', 'τα'), ('τα', 'αρθρα')], [('η', 'ερωτηση'), ('ερωτηση', 'με'), ('με', 'την'), ('την', 'οποιαν'), ('οποιαν', 'με'), ('με', 'αφηκες'), ('αφηκες', 'εν'), ('εν', 'σαν'), ('σαν', 'τζιεινα'), ('τζιεινα', 'τα'), ('τα', 'παιχνιθκια'), ('παιχνιθκια', 'που'), ('που', 'παιζουμεν'), ('παιζουμεν', 'αμαν'), ('αμαν', 'πιουμεν'), ('πιουμεν', 'λλιον'), ('λλιον', 'τζιαι'), ('τζιαι', 'διουμεν'), ('διουμεν', 'ο'), ('ο', 'ενας'), ('ενας', 'του'), ('του', 'αλλου'), ('αλλου', 'αδυνατες'), ('αδυνατες', 'επιλογες'), ('επιλογες', 'για'), ('για', 'το'), ('το', 'χαζιν')], [('δυστυχως', 'ετσι'), ('ετσι', 'ενι'), ('ενι', 'πολλα'), ('πολλα', 'πραματα'), ('πραματα', 'φαινοντε'), ('φαινοντε', 'τελια'), ('τελια', 'εξωτερικα'), ('εξωτερικα', 'αλλα'), ('αλλα', 'που'), ('που', 'μεσα'), ('μεσα', 'εν'), ('εν', 'μαυρα')], [('ειδες', 'ιντα'), ('ιντα', 'κακον'), ('κακον', 'μας'), ('μας', 'καμνουν'), ('καμνουν', 'τα'), ('τα', 'κοπελλουθκια'), ('κοπελλουθκια', 'απιστευτα'), ('απιστευτα', 'τα'), ('τα', 'κοπελλουθκια'), ('κοπελλουθκια', 'ποσα'), ('ποσα', 'εχουμεν'), ('εχουμεν', 'να'), ('να', 'μαθουμε'), ('μαθουμε', 'που'), ('που', 'λοου'), ('λοου', 'τους')], [('για', 'το'), ('το', 'πουρκουριν'), ('πουρκουριν', 'δοτζιημασε'), ('δοτζιημασε', 'την'), ('την', 'παραδοσιακην'), ('παραδοσιακην', 'συνταγην'), ('συνταγην', 'μεσα'), ('μεσα', 'στην'), ('στην', 'οποιαν'), ('οποιαν', 'θα'), ('θα', 'βαλεις'), ('βαλεις', 'οτι'), ('οτι', 'χορτικον'), ('χορτικον', 'κοψει'), ('κοψει', 'ο'), ('ο', 'νους'), ('νους', 'σου')], [('εν', 'πραμαν'), ('πραμαν', 'σιορ'), ('σιορ', 'που'), ('που', 'τρια'), ('τρια', 'πραματα'), ('πραματα', 'που'), ('που', 'γραφει'), ('γραφει', 'ο'), ('ο', 'αλλος'), ('αλλος', 'να'), ('να', 'τον'), ('τον', 'συμπαθας'), ('συμπαθας', 'η'), ('η', 'να'), ('να', 'τον'), ('τον', 'αντιπαθας'), ('αντιπαθας', 'με'), ('με', 'σεναν'), ('σεναν', 'εννοειται'), ('εννοειται', 'οτι'), ('οτι', 'τσιαττα'), ('τσιαττα', 'ο'), ('ο', 'νους'), ('νους', 'μου'), ('μου', 'παμπο')], [('πριν', 'λλιον'), ('λλιον', 'τζιαιρον'), ('τζιαιρον', 'εππεσεν'), ('εππεσεν', 'εναν'), ('εναν', 'κοτσιηνολαιμουιν'), ('κοτσιηνολαιμουιν', 'που'), ('που', 'την'), ('την', 'φωλιαν'), ('φωλιαν', 'του'), ('του', 'τζιαι'), ('τζιαι', 'λυπηθηκα'), ('λυπηθηκα', 'το'), ('το', 'πολλα')], [('για', 'μεναν'), ('μεναν', 'φιλε'), ('φιλε', 'το'), ('το', 'παν'), ('παν', 'εν'), ('εν', 'να'), ('να', 'μεν'), ('μεν', 'ξιαννουμεν'), ('ξιαννουμεν', 'ποθθεν'), ('ποθθεν', 'ηρταμεν'), ('ηρταμεν', 'να'), ('να', 'μεν'), ('μεν', 'μεγαλοπιαννουμαστεν'), ('μεγαλοπιαννουμαστεν', 'τζιαι'), ('τζιαι', 'να'), ('να', 'νομιζουμεν'), ('νομιζουμεν', 'οτι'), ('οτι', 'τωρα'), ('τωρα', 'εγινηκαμεν'), ('εγινηκαμεν', 'αστικη'), ('αστικη', 'ταξη'), ('ταξη', 'τζιαι'), ('τζιαι', 'να'), ('να', 'περιφρονουμεν'), ('περιφρονουμεν', 'ανθρωπους'), ('ανθρωπους', 'με'), ('με', 'βασην'), ('βασην', 'την'), ('την', 'κοινωνικην'), ('κοινωνικην', 'ταξην'), ('ταξην', 'τζιαι'), ('τζιαι', 'το'), ('το', 'επαγγελμαν')], [('ειντα', 'καθαρα'), ('καθαρα', 'που'), ('που', 'θωρεις'), ('θωρεις', 'ειντα'), ('ειντα', 'γυαλλια'), ('γυαλλια', 'ηβρες'), ('ηβρες', 'να'), ('να', 'πιασω'), ('πιασω', 'τζιαι'), ('τζιαι', 'γω')], [('ακομα', 'να'), ('να', 'φκουμεν'), ('φκουμεν', 'πας'), ('πας', 'τον'), ('τον', 'αππαρον'), ('αππαρον', 'και'), ('και', 'σουζουμεν'), ('σουζουμεν', 'τα'), ('τα', 'ποθκια'), ('ποθκια', 'μας')], [('παντως', 'το'), ('το', 'να'), ('να', 'θαυμαζεις'), ('θαυμαζεις', 'εναν'), ('εναν', 'τεθκοιον'), ('τεθκοιον', 'ανθρωπο'), ('ανθρωπο', 'που'), ('που', 'εισιεν'), ('εισιεν', 'οραμα'), ('οραμα', 'σιουρα'), ('σιουρα', 'καμνει'), ('καμνει', 'σε'), ('σε', 'να'), ('να', 'τον'), ('τον', 'νοιωθεις'), ('νοιωθεις', 'σαν'), ('σαν', 'δικο'), ('δικο', 'σου')], [('η', 'αριστερη'), ('αριστερη', 'ιδεολογια'), ('ιδεολογια', 'καμνει'), ('καμνει', 'κακον'), ('κακον', 'μονον'), ('μονον', 'αμαν'), ('αμαν', 'παει'), ('παει', 'στραβα'), ('στραβα', 'η'), ('η', 'δεξια'), ('δεξια', 'εσιει'), ('εσιει', 'το'), ('το', 'μες'), ('μες', 'το'), ('το', 'προγραμμα'), ('προγραμμα', 'της'), ('της', 'να'), ('να', 'καμει'), ('καμει', 'κακον')], [('ειμαι', 'λλιον'), ('λλιον', 'απαισιοδοξος'), ('απαισιοδοξος', 'γενικα'), ('γενικα', 'γιατι'), ('γιατι', 'νομιζω'), ('νομιζω', 'τουντα'), ('τουντα', 'πραματα'), ('πραματα', 'πρεπει'), ('πρεπει', 'ναν'), ('ναν', 'λλιον'), ('λλιον', 'οργανικα'), ('οργανικα', 'που'), ('που', 'την'), ('την', 'βαση'), ('βαση', 'τζιαι'), ('τζιαι', 'πανω')], [('ελπιζουμεν', 'σε'), ('σε', 'εναν'), ('εναν', 'καλλυττερο'), ('καλλυττερο', 'μελλον')], [('ωσπου', 'παρατηρω'), ('παρατηρω', 'τους'), ('τους', 'κυπραιους'), ('κυπραιους', 'καταλαββαινω'), ('καταλαββαινω', 'οτι'), ('οτι', 'η'), ('η', 'φυσικη'), ('φυσικη', 'τους'), ('τους', 'κατασταση'), ('κατασταση', 'εν'), ('εν', 'να'), ('να', 'εν'), ('εν', 'παφτωσιοι'), ('παφτωσιοι', 'να'), ('να', 'δουλευκουν'), ('δουλευκουν', 'στα'), ('στα', 'χωραφκια')], [('μασιαλλα', 'σου'), ('σου', 'κορη'), ('κορη', 'ηταν'), ('ηταν', 'πολλα'), ('πολλα', 'καλη'), ('καλη', 'η'), ('η', 'ιδεα')], [('δηλαδη', 'να'), ('να', 'κοψει'), ('κοψει', 'πισω'), ('πισω', 'η'), ('η', 'κυπρος'), ('κυπρος', 'που'), ('που', 'την'), ('την', 'εξελιξη'), ('εξελιξη', 'για'), ('για', 'να'), ('να', 'εσιετε'), ('εσιετε', 'εσεις'), ('εσεις', 'να'), ('να', 'μας'), ('μας', 'κατακλεφκετε')], [('τουτο', 'που'), ('που', 'επροσεξα'), ('επροσεξα', 'εν'), ('εν', 'οτι'), ('οτι', 'οσοι'), ('οσοι', 'ειμαστεν'), ('ειμαστεν', 'λιο'), ('λιο', 'μπλεγμενοι'), ('μπλεγμενοι', 'στα'), ('στα', 'ακαδημαικα')], [('εθκιαβασα', 'το'), ('το', 'αφιερωμαν'), ('αφιερωμαν', 'σου'), ('σου', 'στον'), ('στον', 'παπα'), ('παπα', 'σας'), ('σας', 'κορη')], [('εν', 'χρειαζουμαι'), ('χρειαζουμαι', 'προσπαθεια'), ('προσπαθεια', 'για'), ('για', 'να'), ('να', 'ευρω'), ('ευρω', 'κατι'), ('κατι', 'να'), ('να', 'πω')], [('σιουρα', 'φκαλλει'), ('φκαλλει', 'πολλη'), ('πολλη', 'θετικη'), ('θετικη', 'ενεργεια'), ('ενεργεια', 'τουτο'), ('τουτο', 'το'), ('το', 'ποστ'), ('ποστ', 'ατε'), ('ατε', 'καλα'), ('καλα', 'εκαμες')], [('το', 'προβλημαν'), ('προβλημαν', 'ειναι'), ('ειναι', 'οτι'), ('οτι', 'πρεπει'), ('πρεπει', 'να'), ('να', 'αρκεψει'), ('αρκεψει', 'με'), ('με', 'μιαν'), ('μιαν', 'καλην'), ('καλην', 'χειρονομιαν')], [('οπως', 'τα'), ('τα', 'λαλεις'), ('λαλεις', 'εχουν'), ('εχουν', 'τα'), ('τα', 'πραματα'), ('πραματα', 'φιλε'), ('φιλε', 'μου')], [('εν', 'μακρυς'), ('μακρυς', 'ο'), ('ο', 'σιειμωνας'), ('σιειμωνας', 'εβαρεθηκα'), ('εβαρεθηκα', 'πιον'), ('πιον', 'θελω'), ('θελω', 'παραλιαν')], [('με', 'τη'), ('τη', 'φοασε'), ('φοασε', 'τρωει'), ('τρωει', 'που'), ('που', 'ουλλα'), ('ουλλα', 'εκαταλαβεν'), ('εκαταλαβεν', 'οτι'), ('οτι', 'διας'), ('διας', 'βαρος'), ('βαρος', 'στο'), ('στο', 'θεμαν'), ('θεμαν', 'του'), ('του', 'φαγιου')], [('εν', 'πολλη'), ('πολλη', 'η'), ('η', 'σημασια'), ('σημασια', 'που'), ('που', 'τους'), ('τους', 'διουμεν'), ('διουμεν', 'σε'), ('σε', 'συγκρισην')], [('ενομησες', 'οτι'), ('οτι', 'καμνω'), ('καμνω', 'ακελικη'), ('ακελικη', 'προπαγανδα'), ('προπαγανδα', 'ρε'), ('ρε', 'χοχοι'), ('χοχοι', 'μα'), ('μα', 'εν'), ('εν', 'να'), ('να', 'μας'), ('μας', 'πελλανετε')], [('εννα', 'με'), ('με', 'πελλανουν'), ('πελλανουν', 'αρκεψα'), ('αρκεψα', 'γυμναστηριον'), ('γυμναστηριον', 'πριν'), ('πριν', 'θκυο'), ('θκυο', 'εφτομαδες'), ('εφτομαδες', 'για'), ('για', 'να'), ('να', 'χασω'), ('χασω', 'κιλα'), ('κιλα', 'ταχα'), ('ταχα', 'καμνω'), ('καμνω', 'τες'), ('τες', 'γυμναστικες'), ('γυμναστικες', 'που'), ('που', 'μου'), ('μου', 'βαλλει'), ('βαλλει', 'τζιαι'), ('τζιαι', 'ξαναζυγιζουμαι'), ('ξαναζυγιζουμαι', 'εχτες')], [('εβαλα', 'αναμιση'), ('αναμιση', 'κιλο'), ('κιλο', 'γαμω'), ('γαμω', 'το'), ('το', 'σσιηστο'), ('σσιηστο', 'μου'), ('μου', 'ηνταλως'), ('ηνταλως', 'θκιαολον'), ('θκιαολον', 'καμνω'), ('καμνω', 'ασκησεις'), ('ασκησεις', 'να'), ('να', 'χασω'), ('χασω', 'ταχα'), ('ταχα', 'τζιαι'), ('τζιαι', 'γινουμαι'), ('γινουμαι', 'πιο'), ('πιο', 'βαρετος')], [('τουτα', 'τα'), ('τα', 'ονειρα'), ('ονειρα', 'εκαμαμε'), ('εκαμαμε', 'τα'), ('τα', 'μεγαλο'), ('μεγαλο', 'θεμα'), ('θεμα', 'μες'), ('μες', 'τουτες'), ('τουτες', 'τες'), ('τες', 'μερες')], [('τζεινο', 'που'), ('που', 'φοαστε'), ('φοαστε', 'θα'), ('θα', 'το'), ('το', 'παθετε')], [('σαν', 'την'), ('την', 'κυπρον'), ('κυπρον', 'εν'), ('εν', 'εσιει'), ('εσιει', 'ναμπου'), ('ναμπου', 'καμνεις')], [('φιλε', 'μιλω'), ('μιλω', 'σου'), ('σου', 'εχασα'), ('εχασα', 'τα'), ('τα', 'φωτα'), ('φωτα', 'μου')], [('τουντο', 'ποστ'), ('ποστ', 'ηρτεν'), ('ηρτεν', 'πας'), ('πας', 'την'), ('την', 'ωρα')], [('α', 'μανα'), ('μανα', 'μου'), ('μου', 'ειντα'), ('ειντα', 'αππαρα'), ('αππαρα', 'εν'), ('εν', 'τουτη'), ('τουτη', 'αρεσκει'), ('αρεσκει', 'μου'), ('μου', 'πολλα')], [('γιατι', 'ρε'), ('ρε', 'παιδακι'), ('παιδακι', 'μου'), ('μου', 'ρε'), ('ρε', 'κορουδα'), ('κορουδα', 'μου'), ('μου', 'να'), ('να', 'μου'), ('μου', 'καλαμαριζεις'), ('καλαμαριζεις', 'αμαν'), ('αμαν', 'μιλας')], [('νευριαζουμεν', 'με'), ('με', 'τζαι'), ('τζαι', 'εμενα'), ('εμενα', 'πολλα')], [('εγω', 'εν'), ('εν', 'καλαμαριζω'), ('καλαμαριζω', 'αλλα'), ('αλλα', 'βρισκω'), ('βρισκω', 'πολλα'), ('πολλα', 'αστειες'), ('αστειες', 'καποιες'), ('καποιες', 'καλαμαριστικες'), ('καλαμαριστικες', 'εκφρασεις')]] ###Markdown 3.2.3 Chars `Text` objects ###Code cg_chars_flat = [char for word in cg_words_flat for char in word] smg_chars_flat = [char for word in smg_words_flat for char in word] cg_chars_Text = Text(cg_chars_flat) smg_chars_Text = Text(smg_chars_flat) cg_chars_Text ###Output _____no_output_____ ###Markdown 3.2.4 Character n-grams 3.2.4.1 Bigrams `Text` objects ###Code cg_char_bigrams = [] smg_char_bigrams = [] for word in cg_words_flat: cg_char_bigrams.append(list(ngrams(word, 2, pad_left=True, pad_right=True, left_pad_symbol='_', right_pad_symbol='_'))) for word in smg_words_flat: smg_char_bigrams.append(list(ngrams(word, 2, pad_left=True, pad_right=True, left_pad_symbol='_', right_pad_symbol='_'))) cg_char_bigrams_flat_tuples = [bigram for bigram_list in cg_char_bigrams for bigram in bigram_list] smg_char_bigrams_flat_tuples = [bigram for bigram_list in smg_char_bigrams for bigram in bigram_list] cg_char_bigrams_flat = ['%s%s' % bigram_tuple for bigram_tuple in cg_char_bigrams_flat_tuples] smg_char_bigrams_flat = ['%s%s' % bigram_tuple for bigram_tuple in smg_char_bigrams_flat_tuples] cg_char_bigrams_Text = Text(cg_char_bigrams_flat) smg_char_bigrams_Text = Text(smg_char_bigrams_flat) cg_char_bigrams_Text ###Output _____no_output_____ ###Markdown 3.2.4.2 Trigrams `Text` objects ###Code import warnings cg_char_trigrams = [] smg_char_trigrams = [] with warnings.catch_warnings(): warnings.filterwarnings('ignore',category=DeprecationWarning) for word in cg_words_flat: if len(word) > 1: # Trigram features for 1-letter words are useless and encoded by other features I use. cg_char_trigrams.append(list(ngrams(word, 3, pad_left=True, pad_right=True, left_pad_symbol='_', right_pad_symbol='_'))) for word in smg_words_flat: if len(word) > 1: smg_char_trigrams.append(list(ngrams(word, 3, pad_left=True, pad_right=True, left_pad_symbol='_', right_pad_symbol='_'))) # Removing redundant trigrams: cg_char_trigrams = [trigram_list[1:-1] for trigram_list in cg_char_trigrams] smg_char_trigrams = [trigram_list[1:-1] for trigram_list in smg_char_trigrams] cg_char_trigrams_flat_tuples = [trigram for trigram_list in cg_char_trigrams for trigram in trigram_list] smg_char_trigrams_flat_tuples = [trigram for trigram_list in smg_char_trigrams for trigram in trigram_list] cg_char_trigrams_flat = ['%s%s%s' % trigram_tuple for trigram_tuple in cg_char_trigrams_flat_tuples] smg_char_trigrams_flat = ['%s%s%s' % trigram_tuple for trigram_tuple in smg_char_trigrams_flat_tuples] cg_char_trigrams_Text = Text(cg_char_trigrams_flat) smg_char_trigrams_Text = Text(smg_char_trigrams_flat) cg_char_trigrams_Text ###Output _____no_output_____ ###Markdown 4. Analysis 4.1. Corpus size ###Code print('Number of CG sentences:', len(cg_sents_clean)) print('Number of SMG sentences:', len(smg_sents_clean)) print('Number of words in CG data:', len(cg_words_flat)) print('Number of words in SMG data:', len(smg_words_flat)) ###Output Number of words in CG data: 7026 Number of words in SMG data: 7100 ###Markdown 4.2 Most frequent words and characters 4.2.1 Most frequent words 4.2.1.1 CG ###Code cg_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.2.1.2 SMG ###Code smg_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.2.2 Most frequent characters 4.2.2.1 CG ###Code cg_chars_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.2.2.2 SMG ###Code smg_chars_Text.plot(10) ###Output _____no_output_____ ###Markdown NB: Recall that 'σ' in the most frequent character charts above includes instances of both 'σ' and 'ς'. 4.3 Most frequent word n-grams 4.3.1 Bigrams 4.3.1.1 CG ###Code cg_word_bigrams_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.3.1.2 SMG ###Code smg_word_bigrams_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.4 Most frequent character n-grams 4.4.1 Bigrams 4.4.1.1 CG ###Code cg_char_bigrams_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.4.1.2 SMG ###Code smg_char_bigrams_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.4.2 Trigrams 4.4.2.1 CG ###Code cg_char_trigrams_Text.plot(10) ###Output _____no_output_____ ###Markdown 4.4.2.2 SMG ###Code smg_char_trigrams_Text.plot(10) ###Output _____no_output_____
Neural Networks and Deep Learning/Week 2 - Logistic Regression as a Neural Network.ipynb
###Markdown Binary Clasification Logistic Regression is a statistical model used to give a probability of a certain class or event such as, fail/pass, dead/alive, win/lose.So let's break down this.What is a statistical model?is a mathematical model that use probability and include some assumptions from a sample data that can be applied for a populationThe statistical model also represent an ideal form of data-generating processok, but the what is a mathematical model?Actually a model is a representation of a system (it described), using mathematical concepts and language. ![Logistic Regression](assets/logisticregression.png) ![logistic regression notation](assets/logisticregreationnotation.png) Logistic Regression Logistic Regression is an algorithm for binary clasification (A learning algorithm) ![Logistic Regression](assets/logisticregretion.png) Logistical Regression Cost Function So to train the parameters - **w**- **b**we need to define a **cost function** ![Cost Function](assets/CostFunction.png) So the important here is Lost Function- Is the value that determinate how well you're doing for a single training example Cost Function- Is the value that determinate how well you're doing for a entire training set Gradient Descent So Gradient Descent allow us to minime the cost function, in other words, it improve our prediction for an entire training set.A formal concept. Gradient Descent is a first-order iterative optimization algorithm for finding the minimum of a function. ![Gradient Descent](assets/GradientDescent.png) ![Gradient Descent2](assets/GradientDescent2.png) So we use the **derivate** (slope of a function by a given point) in **gradient descent** to calculate the minimum of a function.In other to calculate we must see first Computation Graph![computation graph](assets/Computergraph.png) Derivates with a Computation GraphSo the key takeaway from this example, is that when **computing derivatives** and computing all of these derivatives, the **most efficient way to do it** is through a **right to left** computation following the direction of the red arrows![derivates with computation graph](assets/Computingderivates.png) Logistic Regression Gradient DescentJust to recap![Logistic Regression Gradient Descent](assets/LogisticregressionRecap.png)![Logistic Regression Derivates](assets/LogisticRegressionDerivates.png) This is algorithm to implement![Logistic Regression Algorithm](assets/LogisticRegressionAlgorithm.png) As we see the algorithm use 2 **for** loops- Training set- Gradient DescentIn fact, **for** loops are not so efficient when we implement it, with a huge amount of data VectorizationVectorization is the **art** of **get rid** of **for** loopsWe use vectorization because it's **much faster** than the **for** loop, an example below ###Code import numpy as np import time a = np.random.rand(int(1e6)) b = np.random.rand(int(1e6)) tic = time.time() c = np.dot(a,b) toc = time.time() print("Number of values " + str(c)) print("{} {} ms \n".format("Vectorized version", str(1000*(toc-tic)))) c = 0 tic = time.time() for i in range(int(1e6)): c += a[i] * b[i] toc = time.time() print("Number of values " + str(c)) print("{} {} ms".format("For loop", str(1000*(toc-tic)))) ###Output Number of values 249743.245193857 Vectorized version 1.4030933380126953 ms Number of values 249743.24519385604 For loop 464.6720886230469 ms
notebook/DownloadEnwikiDump.ipynb
###Markdown Download Files=== ###Code import requests import json import os git_root_dir = !git rev-parse --show-toplevel git_root_dir = git_root_dir[0] git_root_dir res = requests.get("https://dumps.wikimedia.org/enwiki/20210101/dumpstatus.json") dump_status = json.loads(res.text) dump_status['jobs'].keys() jobs = dump_status['jobs'] jobs.keys() metahistorybz2dump = jobs['metahistorybz2dump'] metahistorybz2dump.keys() len(metahistorybz2dump['files']) (35*60)*671 / 60 / 60 / 24 metahistorybz2dump['updated'] metahistorybz2dump['status'] history1_files = [] for file in metahistorybz2dump['files']: if "history1.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history1_files.append(file_url) len(history1_files) history1_files def generate_download_script(file_list, output_filepath, url_base = "https://dumps.wikimedia.org"): target_dir = "/export/scratch2/wiki_data/enwiki-20210101-pages-meta-history-bz2" with open(output_filepath, 'w') as outfile: outfile.write("#!/bin/bash\n") outfile.write("# This script autogenerated by DownloadEnwikiDump.ipynb\n\n") outfile.write('echo "Starting download." && \\\n') for file_url in file_list: full_url = url_base + file_url outfile.write(f'echo "Downloading \'{file_url}\'." && \\\n') outfile.write(f'wget --no-check-certificate -nc -O {target_dir}/{os.path.basename(file_url)} "{full_url}" && \\\n') outfile.write('echo "Successful." && exit 0\n') outfile.write('echo "Error downloading." && exit 1\n\n') output_filepath = os.path.join(git_root_dir, "scripts", "history1_20210101_download.sh") generate_download_script(history1_files, output_filepath) [history1_files[-4]] os.path.basename(history1_files[-4]) history2_files = [] for file in metahistorybz2dump['files']: if "history2.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history2_files.append(file_url) len(history2_files) output_filepath = os.path.join(git_root_dir, "scripts", "history2_20200101_download.sh") generate_download_script(history2_files, output_filepath) history3_files = [] for file in metahistorybz2dump['files']: if "history3.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history3_files.append(file_url) len(history3_files) output_filepath = os.path.join(git_root_dir, "scripts", "history3_20200101_download.sh") generate_download_script(history3_files, output_filepath) history_num = 4 history_files = [] for file in metahistorybz2dump['files']: if f"history{history_num}.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history_files.append(file_url) print(len(history_files)) output_filepath = os.path.join(git_root_dir, "scripts", f"history{history_num}_20200101_download.sh") generate_download_script(history_files, output_filepath) history_nums = [6, 7, 8, 9, 10] history_files = [] for file in metahistorybz2dump['files']: for history_num in history_nums: if f"history{history_num}.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history_files.append(file_url) print(len(history_files)) output_filepath = os.path.join(git_root_dir, "scripts", f"history6to10_20200101_download.sh") generate_download_script(history_files, output_filepath) history_nums = list(range(11, 21)) print(history_nums) history_files = [] for file in metahistorybz2dump['files']: for history_num in history_nums: if f"history{history_num}.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history_files.append(file_url) print(len(history_files)) output_filepath = os.path.join(git_root_dir, "scripts", f"history11to20_20200101_download.sh") generate_download_script(history_files, output_filepath) history_nums = list(range(21, 28)) print(history_nums) history_files = [] for file in metahistorybz2dump['files']: for history_num in history_nums: if f"history{history_num}.xml" in file: file_url = metahistorybz2dump['files'][file]['url'] history_files.append(file_url) print(len(history_files)) output_filepath = os.path.join(git_root_dir, "scripts", f"history21to27_20200101_download.sh") generate_download_script(history_files, output_filepath) ###Output [21, 22, 23, 24, 25, 26, 27] 202 ###Markdown Stub Meta HistoryGenerate scripts for downloading stub meta history. ###Code xmlstubsdump = jobs['xmlstubsdump'] history_files = [] for file in xmlstubsdump['files']: if "stub-meta-history" in file: file_url = xmlstubsdump['files'][file]['url'] history_files.append(file_url) print(len(history_files)) def generate_stub_download_script(file_list, output_filepath, url_base = "https://dumps.wikimedia.org"): target_dir = "/export/scratch2/wiki_data/enwiki-20210101-stub-meta-history-gz" with open(output_filepath, 'w') as outfile: outfile.write("#!/bin/bash\n") outfile.write("# This script autogenerated by DownloadEnwikiDump.ipynb\n\n") outfile.write('echo "Starting download." && \\\n') for file_url in file_list: full_url = url_base + file_url outfile.write(f'echo "Downloading \'{file_url}\'." && \\\n') outfile.write(f'wget --no-check-certificate -nc -O {target_dir}/{os.path.basename(file_url)} "{full_url}" && \\\n') outfile.write('echo "Successful." && exit 0\n') outfile.write('echo "Error downloading." && exit 1\n\n') output_filepath = os.path.join(git_root_dir, "scripts", f"stub_history_20210101_download.sh") generate_stub_download_script(history_files, output_filepath) ###Output _____no_output_____
Chapter 3/Section 3.5.ipynb
###Markdown [3.5 组合语言的解释器](http://www-inst.eecs.berkeley.edu/~cs61a/sp12/book/interpretation.htmlinterpreters-for-languages-with-combination)运行在任何现代计算机上的软件都以多种编程语言写成。其中有物理语言,例如用于特定计算机的机器语言。这些语言涉及到基于独立储存位和原始机器指令的数据表示和控制。机器语言的程序员涉及到使用提供的硬件,为资源有限的计算构建系统和功能的高效实现。高阶语言构建在机器语言之上,隐藏了表示为位集的数据,以及表示为原始指令序列的程序的细节。这些语言拥有例如过程定义的组合和抽象的手段,它们适用于组织大规模的软件系统。元语言抽象 -- 建立了新的语言 -- 并在所有工程设计分支中起到重要作用。它对于计算机编程尤其重要,因为我们不仅仅可以在编程中构想出新的语言,我们也能够通过构建解释器来实现它们。编程语言的解释器是一个函数,它在语言的表达式上调用,执行求解表达式所需的操作。我们现在已经开始了技术之旅,通过这种技术,编程语言可以建立在其它语言之上。我们首先会为计算器定义解释器,它是一种受限的语言,和 Python 调用表达式具有相同的语法。我们之后会从零开始开发 Scheme 和 Logo 语言的解释器,它们都是 Lisp 的方言,Lisp 是现在仍旧广泛使用的第二老的语言。我们所创建的解释器,在某种意义上,会让我们使用 Logo 编写完全通用的程序。为了这样做,它会实现我们已经在这门课中开发的求值环境模型。 3.5.1 计算器我们的第一种新语言叫做计算器,一种用于加减乘除的算术运算的表达式语言。计算器拥有 Python 调用表达式的语法,但是它的运算符对于所接受的参数数量更加灵活。例如,计算器运算符`mul`和`add`可接受任何数量的参数: ###Code calc> add(1, 2, 3, 4) 10 calc> mul() 1 ###Output _____no_output_____ ###Markdown `sub`运算符拥有两种行为:传入一个运算符,它会对运算符取反。传入至少两个,它会从第一个参数中减掉剩余的参数。`div`运算符拥有 Python 的`operator.truediv`的语义,只接受两个参数。 ###Code calc> sub(10, 1, 2, 3) 4 calc> sub(3) -3 calc> div(15, 12) 1.25 ###Output _____no_output_____ ###Markdown 就像 Python 中那样,调用表达式的嵌套提供了计算器语言中的组合手段。为了精简符号,我们使用运算符的标准符号来代替名称: ###Code calc> sub(100, mul(7, add(8, div(-12, -3)))) 16.0 calc> -(100, *(7, +(8, /(-12, -3)))) 16.0 ###Output _____no_output_____ ###Markdown 我们会使用 Python 实现计算器解释器。也就是说,我们会编写 Python 程序来接受字符串作为输入,并返回求值结果。如果输入是符合要求的计算器表达式,结果为字符串,反之会产生合适的异常。计算器语言解释器的核心是叫做`calc_eval`的递归函数,它会求解树形表达式对象。**表达式树。**到目前为止,我们在描述求值过程中所引用的表达式树,还是概念上的实体。我们从没有显式将表达式树表示为程序中的数据。为了编写解释器,我们必须将表达式当做数据操作。在这一章中,许多我们之前介绍过的概念都会最终以代码实现。计算器中的基本表达式只是一个数值,类型为`int`或`float`。所有复合表达式都是调用表达式。调用表达式表示为拥有两个属性实例的`Exp`类。计算器的`operator`总是字符串:算数运算符的名称或符号。`operands`要么是基本表达式,要么是`Exp`的实例本身。 ###Code class Exp(object): """A call expression in Calculator.""" def __init__(self, operator, operands): self.operator = operator self.operands = operands def __repr__(self): return 'Exp({0}, {1})'.format(repr(self.operator), repr(self.operands)) def __str__(self): operand_strs = ', '.join(map(str, self.operands)) return '{0}({1})'.format(self.operator, operand_strs) ###Output _____no_output_____ ###Markdown `Exp`实例定义了两个字符串方法。`__repr__`方法返回 Python 表达式,而`__str__`方法返回计算器表达式。 ###Code Exp('add', [1, 2]) str(Exp('add', [1, 2])) Exp('add', [1, Exp('mul', [2, 3, 4])]) str(Exp('add', [1, Exp('mul', [2, 3, 4])])) ###Output _____no_output_____ ###Markdown 最后的例子演示了`Exp`类如何通过包含作为`operands`元素的`Exp`的实例,来表示表达式树中的层次结构。 ###Code def calc_eval(exp): """Evaluate a Calculator expression.""" if type(exp) in (int, float): return exp elif type(exp) == Exp: arguments = list(map(calc_eval, exp.operands)) return calc_apply(exp.operator, arguments) ###Output _____no_output_____ ###Markdown **求值。**`calc_eval`函数接受表达式作为参数,并返回它的值。它根据表达式的形式为表达式分类,并且指导它的求值。对于计算器来说,表达式的两种句法形式是数值或调用表达式,后者是`Exp`的实例。数值是自求值的,它们可以直接从`calc_eval`中返回。调用表达式需要使用函数。调用表达式首先通过将`calc_eval`函数递归映射到操作数的列表,计算出参数列表来求值。之后,在第二个函数`calc_apply`中,运算符会作用于这些参数上。计算器语言足够简单,我们可以轻易地在单一函数中表达每个运算符的使用逻辑。在`calc_apply`中,每种条件子句对应一个运算符。 ###Code from operator import mul from functools import reduce def calc_apply(operator, args): """Apply the named operator to a list of args.""" if operator in ('add', '+'): return sum(args) if operator in ('sub', '-'): if len(args) == 0: raise TypeError(operator + ' requires at least 1 argument') if len(args) == 1: return -args[0] return sum(args[:1] + [-arg for arg in args[1:]]) if operator in ('mul', '*'): return reduce(mul, args, 1) if operator in ('div', '/'): if len(args) != 2: raise TypeError(operator + ' requires exactly 2 arguments') numer, denom = args return numer/denom ###Output _____no_output_____ ###Markdown 上面,每个语句组计算了不同运算符的结果,或者当参数错误时产生合适的`TypeError`。`calc_apply`函数可以直接调用,但是必须传入值的列表作为参数,而不是运算符表达式的列表。 ###Code calc_apply('+', [1, 2, 3]) calc_apply('-', [10, 1, 2, 3]) calc_apply('*', []) calc_apply('/', [40, 5]) ###Output _____no_output_____ ###Markdown `calc_eval`的作用是,执行合适的`calc_apply`调用,通过首先计算操作数子表达式的值,之后将它们作为参数传入`calc_apply`。于是,`calc_eval`可以接受嵌套表达式。 ###Code e = Exp('add', [2, Exp('mul', [4, 6])]) str(e) calc_eval(e) ###Output _____no_output_____ ###Markdown `calc_eval`的结构是个类型(表达式的形式)分发的例子。第一种表达式是数值,不需要任何的额外求值步骤。通常,基本表达式不需要任何额外的求值步骤,这叫做自求值。计算器语言中唯一的自求值表达式就是数值,但是在通用语言中可能也包括字符串、布尔值,以及其它。**“读取-求值-打印”循环。**和解释器交互的典型方式是“读取-求值-打印”循环(REPL),它是一种交互模式,读取表达式、对其求值,之后为用户打印出结果。Python 交互式会话就是这种循环的例子。REPL 的实现与所使用的解释器无关。下面的`read_eval_print_loop`函数使用内建的`input`函数,从用户接受一行文本作为输入。它使用语言特定的`calc_parse`函数构建表达式树。`calc_parse`在随后的解析一节中定义。最后,它打印出对由`calc_parse`返回的表达式树调用`calc_eval`的结果。 ###Code def read_eval_print_loop(): """Run a read-eval-print loop for calculator.""" while True: expression_tree = calc_parse(input('calc> ')) print(calc_eval(expression_tree)) ###Output _____no_output_____ ###Markdown `read_eval_print_loop`的这个版本包含所有交互式界面的必要组件。一个样例会话可能像这样: ###Code calc> mul(1, 2, 3) 6 calc> add() 0 calc> add(2, div(4, 8)) 2.5 ###Output _____no_output_____ ###Markdown 这个循环没有实现终端或者错误处理机制。我们可以通过向用户报告错误来改进这个界面。我们也可以允许用户通过发射键盘中断信号(`Control-C`),或文件末尾信号(`Control-D`)来退出循环。为了实现这些改进,我们将原始的`while`语句组放在`try`语句中。第一个`except`子句处理了由`calc_parse`产生的`SyntaxError`异常,也处理了由`calc_eval`产生的`TypeError`和`ZeroDivisionError`异常。 ###Code def read_eval_print_loop(): """Run a read-eval-print loop for calculator.""" while True: try: expression_tree = calc_parse(input('calc> ')) print(calc_eval(expression_tree)) except (SyntaxError, TypeError, ZeroDivisionError) as err: print(type(err).__name__ + ':', err) except (KeyboardInterrupt, EOFError): # <Control>-D, etc. print('Calculation completed.') return ###Output _____no_output_____ ###Markdown 这个循环实现报告错误而不退出循环。发生错误时不退出程序,而是在错误消息之后重新开始循环可以让用户回顾他们的表达式。通过导入`readline`模块,用户甚至可以使用上箭头或`Control-P`来回忆他们之前的输入。最终的结果提供了错误信息报告的界面: ###Code calc> add SyntaxError: expected ( after add calc> div(5) TypeError: div requires exactly 2 arguments calc> div(1, 0) ZeroDivisionError: division by zero calc> ^DCalculation completed. ###Output _____no_output_____ ###Markdown 在我们将解释器推广到计算器之外的语言时,我们会看到,`read_eval_print_loop`由解析函数、求值函数,和由`try`语句处理的异常类型参数化。除了这些修改之外,任何 REPL 都可以使用相同的结构来实现。 3.5.2 解析解析是从原始文本输入生成表达式树的过程。解释这些表达式树是求值函数的任务,但是解析器必须提供符合格式的表达式树给求值器。解析器实际上由两个组件组成,词法分析器和语法分析器。首先,词法分析器将输入字符串拆成标记(token),它们是语言的最小语法单元,就像名称和符号那样。其次,语法分析器从这个标记序列中构建表达式树。 ###Code def calc_parse(line): """Parse a line of calculator input and return an expression tree.""" tokens = tokenize(line) expression_tree = analyze(tokens) if len(tokens) > 0: raise SyntaxError('Extra token(s): ' + ' '.join(tokens)) return expression_tree ###Output _____no_output_____ ###Markdown 标记序列由叫做`tokenize`的词法分析器产生,并被叫做`analyze`语法分析器使用。这里,我们定义了`calc_parse`,它只接受符合格式的计算器表达式。一些语言的解析器为接受以换行符、分号或空格分隔的多种表达式而设计。我们在引入 Logo 语言之前会推迟实现这种复杂性。**词法分析。**用于将字符串解释为标记序列的组件叫做分词器(tokenizer ),或者词法分析器。在我们的视线中,分词器是个叫做`tokenize`的函数。计算器语言由包含数值、运算符名称和运算符类型的符号(比如`+`)组成。这些符号总是由两种分隔符划分:逗号和圆括号。每个符号本身都是标记,就像每个逗号和圆括号那样。标记可以通过向输入字符串添加空格,之后在每个空格处分割字符串来分开。 ###Code def tokenize(line): """Convert a string into a list of tokens.""" spaced = line.replace('(',' ( ').replace(')',' ) ').replace(',', ' , ') return spaced.split() ###Output _____no_output_____ ###Markdown 对符合格式的计算器表达式分词不会损坏名称,但是会分开所有符号和分隔符。 ###Code tokenize('add(2, mul(4, 6))') ###Output _____no_output_____ ###Markdown 拥有更加复合语法的语言可能需要更复杂的分词器。特别是,许多分析器会解析每种返回标记的语法类型。例如,计算机中的标记类型可能是运算符、名称、数值或分隔符。这个分类可以简化标记序列的解析。**语法分析。**将标记序列解释为表达式树的组件叫做语法分析器。在我们的实现中,语法分析由叫做`analyze`的递归函数完成。它是递归的,因为分析标记序列经常涉及到分析这些表达式树中的标记子序列,它本身作为更大的表达式树的子分支(比如操作数)。递归会生成由求值器使用的层次结构。`analyze`函数接受标记列表,以符合格式的表达式开始。它会分析第一个标记,将表示数值的字符串强制转换为数字的值。之后要考虑计算机中的两个合法表达式类型。数字标记本身就是完整的基本表达式树。复合表达式以运算符开始,之后是操作数表达式的列表,由圆括号分隔。我们以一个不检查语法错误的实现开始。 ###Code def analyze(tokens): """Create a tree of nested lists from a sequence of tokens.""" token = analyze_token(tokens.pop(0)) if type(token) in (int, float): return token else: tokens.pop(0) # Remove ( return Exp(token, analyze_operands(tokens)) def analyze_operands(tokens): """Read a list of comma-separated operands.""" operands = [] while tokens[0] != ')': if operands: tokens.pop(0) # Remove , operands.append(analyze(tokens)) tokens.pop(0) # Remove ) return operands ###Output _____no_output_____ ###Markdown 最后,我们需要实现`analyze_token`。`analyze_token`函数将数值文本转换为数值。我们并不自己实现这个逻辑,而是依靠内建的 Python 类型转换,使用`int`和`float`构造器来将标记转换为这种类型。 ###Code def analyze_token(token): """Return the value of token if it can be analyzed as a number, or token.""" try: return int(token) except (TypeError, ValueError): try: return float(token) except (TypeError, ValueError): return token ###Output _____no_output_____ ###Markdown 我们的`analyze`实现就完成了。它能够正确将符合格式的计算器表达式解析为表达式树。这些树由`str`函数转换回计算器表达式。 ###Code expression = 'add(2, mul(4, 6))' analyze(tokenize(expression)) str(analyze(tokenize(expression))) ###Output _____no_output_____ ###Markdown `analyse`函数只会返回符合格式的表达式树,并且它必须检测输入中的语法错误。特别是,它必须检测表达式是否完整、正确分隔,以及只含有已知的运算符。下面的修订版本确保了语法分析的每一步都找到了预期的标记。 ###Code known_operators = ['add', 'sub', 'mul', 'div', '+', '-', '*', '/'] def analyze(tokens): """Create a tree of nested lists from a sequence of tokens.""" assert_non_empty(tokens) token = analyze_token(tokens.pop(0)) if type(token) in (int, float): return token if token in known_operators: if len(tokens) == 0 or tokens.pop(0) != '(': raise SyntaxError('expected ( after ' + token) return Exp(token, analyze_operands(tokens)) else: raise SyntaxError('unexpected ' + token) def analyze_operands(tokens): """Analyze a sequence of comma-separated operands.""" assert_non_empty(tokens) operands = [] while tokens[0] != ')': if operands and tokens.pop(0) != ',': raise SyntaxError('expected ,') operands.append(analyze(tokens)) assert_non_empty(tokens) tokens.pop(0) # Remove ) return elements def assert_non_empty(tokens): """Raise an exception if tokens is empty.""" if len(tokens) == 0: raise SyntaxError('unexpected end of line') ###Output _____no_output_____
cop/r/ECM.ipynb
###Markdown INF285/ILI285 Computación Científica COP-R Ecuación Cuadrática Matricial Librerías ###Code import numpy as np import scipy.sparse.linalg as spla ###Output _____no_output_____ ###Markdown Pregunta La tradicional ecuación cuadrática $\lambda\,x^2+\theta\,x+\gamma=0$ se estudió profundamente al comienzo del curso en el tema de errores de cancelación, esta ecuación se obtiene haciendo una manipulación algebraica conveniente, particularmente la solución es la siguiente:\begin{equation}x_{\pm}=\dfrac{-\theta\pm\sqrt{\theta^2-4\,\lambda\,\gamma}}{2\,\lambda}\end{equation}Uno podría extender la ecuación cuadrática de la siguiente forma:\begin{equation}\Lambda\,X^2+\Theta\,X+\Gamma=\underline{0}\end{equation}donde $\Lambda, \Theta, \Gamma$ y $X \in \mathbb{R}^{n\times n}$, $\underline{0}$ corresponde a la matriz nula de $n \times n$ y $X^2=X\,X$. Una posible simplificación puede ser:\begin{equation}X^2+\Lambda^{-1}\,\Theta\,X+\Lambda^{-1}\,\Gamma=\underline{0}\end{equation}Este caso en general es un poco más complejo de lo esperado, por lo cual estudiaremos la siguiente ecuación:\begin{equation}(X+B)^2=C+CX+X^2+D\end{equation}Los archivos de datos se obtienen desde el siguiente link: https://github.com/sct-utfsm/INF-285/tree/master/cop/r/data/ 1. Considerando en este caso que $B=D=\underline{0}$ y $C=$C1.npy, obtenga la norma infinito de $X$, es decir $\|X\|_\infty$.2. Considerando en este caso que $C=\underline{0}$, $B=$B2.npy y $D=B^2$, obtenga la norma infinito de $X$, es decir $\|X\|_\infty$.3. Considerando en este caso que $B=$B3.npy, $C=$C3.npy y $D=\underline{0}$, obtenga la norma infinito de $X$, es decir $\|X\|_\infty$. Desarrollo ###Code # Loading data C1 = np.load('data/C1.npy') B2 = np.load('data/B2.npy') B3 = np.load('data/B3.npy') C3 = np.load('data/C3.npy') n = C1.shape[0] # 1. (X+B)^2=C+C X+X^2 + D # X^2 = C+C X+X^2 # 0 = C+C X # -C = C X # -C^{-1}C=X # -I=X X = -np.eye(n) print(np.linalg.norm(X,np.inf)) # 2. (X+B)^2=C+C X+X^2 + D # (X+B)^2=X^2 + B^2 # X^2 + B X + X B + B^2 = X^2 + B^2 # B X + X B = 0 X = np.zeros((n,n)) print(np.linalg.norm(X,np.inf)) # 3. (X+B)^2=C+C X+X^2 + D # (X+B)^2=C+C X+X^2 # X^2+B X +X B + B^2=C+C X+X^2 # B X +X B + B^2=C+C X # (B - C) X +X B = C - B^2 def compute_matrix_vector_product(x,B,C,n): X = np.reshape(x,(n,n)) out = np.dot(B-C,X)+np.dot(X,B) return out.flatten() afun = spla.LinearOperator((n**2, n**2), matvec = lambda x: compute_matrix_vector_product(x,B3,C3,n)) x, exitCode = spla.gmres(afun, (C3-np.dot(B3,B3)).flatten(), tol=1e-10) X_GMRes = np.reshape(x,(n,n)) print(np.linalg.norm(X_GMRes,np.inf)) ###Output 1.0 0.0 29.40492924190162
how_to_ac3airborne/datasets/polar/gps_ins.ipynb
###Markdown GPS and INSPosition and orientation of Polar 5 and Polar 6 are recorded by an on-board GPS sensor and the internal navigation system (INS). The following example presents the variables recored by these instruments. Data access* Data accessTo analyse the data they first have to be loaded by importing the (AC)³ airborne meta data catalogue. To do so the ac3airborne package has to be installed. More information on how to do that and about the catalog can be found [here](https://github.com/igmk/ac3airborne-intakeac3airborne-intake-catalogue). Since some of the data, like the preliminary data of the HALO-(AC)3 campaign, is stored on the (AC)3 nextcloud server, username and password as credentials ([registration](https://cloud.ac3-tr.de/index.php/login)) are required and need to be loaded from environment variables. ###Code import os ac3cloud_username = os.environ['AC3_USER'] ac3cloud_password = os.environ['AC3_PASSWORD'] import ac3airborne ###Output _____no_output_____ ###Markdown Get data All flights of Polar 5: ###Code cat = ac3airborne.get_intake_catalog() meta = ac3airborne.get_flight_segments() flights_p5 = {} for campaign in ['ACLOUD', 'AFLUX', 'MOSAiC-ACA','HALO-AC3']: flights_p5.update(meta[campaign]['P5']) ###Output _____no_output_____ ###Markdown All flights of Polar 6: ###Code flights_p6 = {} for campaign in ['ACLOUD', 'PAMARCMiP','HALO-AC3']: flights_p6.update(meta[campaign]['P6']) ###Output _____no_output_____ ###Markdown ```{note}Have a look at the attributes of the xarray dataset `ds_gps_ins` for all relevant information on the dataset, such as author, contact, or citation information.``` ###Code ds_gps_ins = cat['AFLUX']['P5']['GPS_INS']['AFLUX_P5_RF10'].to_dask() ds_gps_ins ###Output Invalid MIT-MAGIC-COOKIE-1 key ###Markdown The dataset `ds_gps_ins` includes the aircraft's position (`lon`, `lat`, `alt`), attitude (`roll`, `pitch`, `heading`), and the ground speed, vertical speed and true air speed (`gs`, `vs`, `tas`). Load flight phase informationPolar 5 flights are divided into segments to easily access start and end times of flight patterns. For more information have a look at the respective [github](https://github.com/igmk/flight-phase-separation) repository.At first we want to load the flight segments of (AC)³ airborne ###Code meta = ac3airborne.get_flight_segments() # this happened before, but it doesn't hurt. ###Output _____no_output_____ ###Markdown The following command lists all flight segments into the dictionary `segments`: ###Code segments = {s.get("segment_id"): {**s, "flight_id": flight["flight_id"]} for campaign in meta.values() for platform in campaign.values() for flight in platform.values() for s in flight["segments"] } ###Output _____no_output_____ ###Markdown In this example, we want to look at a racetrack pattern during `ACLOUD_P5_RF10`. ###Code seg = segments["AFLUX_P5_RF10_rt01"] ###Output _____no_output_____ ###Markdown Using the start and end times of the segment, we slice the data to this flight section. ###Code ds_gps_ins_rt = ds_gps_ins.sel(time=slice(seg["start"], seg["end"])) ###Output _____no_output_____ ###Markdown Plots ###Code %matplotlib inline import matplotlib.pyplot as plt import matplotlib.dates as mdates import matplotlib.colors as mcolors import numpy as np import ipyleaflet from simplification.cutil import simplify_coords_idx plt.style.use("../../mplstyle/book") ###Output _____no_output_____ ###Markdown Plot all flights ###Code def simplify_dataset(ds, tolerance): indices_to_take = simplify_coords_idx(np.stack([ds.lat.values, ds.lon.values], axis=1), tolerance) return ds.isel(time=indices_to_take) # define colors for the flight tracks colors = [mcolors.to_hex(c) for c in plt.cm.inferno(np.linspace(0, 1, len(flights_p5)))] m = ipyleaflet.Map(basemap=ipyleaflet.basemaps.Esri.NatGeoWorldMap, center=(80., 6), zoom=3) for (flight_id, flight),color in zip(flights_p5.items(),colors): mission = flight['mission'] # read gps dataset of flight if mission == 'HALO-AC3': ds = cat[mission]['P5']['GPS_INS'][flight_id](user=ac3cloud_username,password=ac3cloud_password).to_dask() else: ds = cat[mission]['P5']['GPS_INS'][flight_id].to_dask() # slice to takeoff and landing times ds = ds.sel(time=slice(meta[mission]['P5'][flight_id]['takeoff'], meta[mission]['P5'][flight_id]['landing'])) # reduce dataset for plotting ds_reduced = simplify_dataset(ds, tolerance=1e-5) track = ipyleaflet.Polyline( locations=np.stack([ds_reduced.lat.values, ds_reduced.lon.values], axis=1).tolist(), color=color, fill=False, weight=2, name=flight_id) m.add_layer(track) m.add_control(ipyleaflet.ScaleControl(position='bottomleft')) m.add_control(ipyleaflet.LegendControl(dict(zip(flights_p5.keys(), colors)), name="Flights", position="bottomright")) m.add_control(ipyleaflet.LayersControl(position='topright')) m.add_control(ipyleaflet.FullScreenControl()) display(m) ###Output _____no_output_____ ###Markdown Plot time series of one flight ###Code fig, ax = plt.subplots(9, 1, sharex=True) kwargs = dict(s=1, linewidths=0, color='k') ax[0].scatter(ds_gps_ins.time, ds_gps_ins['alt'], **kwargs) ax[0].set_ylabel('alt [m]') ax[1].scatter(ds_gps_ins.time, ds_gps_ins['lon'], **kwargs) ax[1].set_ylabel('lon [°E]') ax[2].scatter(ds_gps_ins.time, ds_gps_ins['lat'], **kwargs) ax[2].set_ylabel('lat [°N]') ax[3].scatter(ds_gps_ins.time, ds_gps_ins['roll'], **kwargs) ax[3].set_ylabel('roll [°]') ax[4].scatter(ds_gps_ins.time, ds_gps_ins['pitch'], **kwargs) ax[4].set_ylabel('pitch [°]') ax[5].scatter(ds_gps_ins.time, ds_gps_ins['heading'], **kwargs) ax[5].set_ylim(-180, 180) ax[5].set_ylabel('heading [°]') ax[6].scatter(ds_gps_ins.time, ds_gps_ins['gs'], **kwargs) ax[6].set_ylabel('gs [kts]') ax[7].scatter(ds_gps_ins.time, ds_gps_ins['vs'], **kwargs) ax[7].set_ylabel('vs [m/s]') ax[8].scatter(ds_gps_ins.time, ds_gps_ins['tas'], **kwargs) ax[8].set_ylabel('tas [m/s]') ax[-1].xaxis.set_major_formatter(mdates.DateFormatter('%H:%M')) plt.show() ###Output _____no_output_____ ###Markdown Plot time series of racetrack pattern ###Code fig, ax = plt.subplots(9, 1, sharex=True) kwargs = dict(s=1, linewidths=0, color='k') ax[0].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['alt'], **kwargs) ax[0].set_ylabel('alt [m]') ax[1].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['lon'], **kwargs) ax[1].set_ylabel('lon [°E]') ax[2].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['lat'], **kwargs) ax[2].set_ylabel('lat [°N]') ax[3].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['roll'], **kwargs) ax[3].set_ylabel('roll [°]') ax[4].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['pitch'], **kwargs) ax[4].set_ylabel('pitch [°]') ax[5].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['heading'], **kwargs) ax[5].set_ylim(-180, 180) ax[5].set_ylabel('heading [°]') ax[6].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['gs'], **kwargs) ax[6].set_ylabel('gs [kts]') ax[7].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['vs'], **kwargs) ax[7].set_ylabel('vs [m/s]') ax[8].scatter(ds_gps_ins_rt.time, ds_gps_ins_rt['tas'], **kwargs) ax[8].set_ylabel('tas [m/s]') ax[-1].xaxis.set_major_formatter(mdates.DateFormatter('%H:%M')) plt.show() ###Output _____no_output_____
notebooks/si_04_variance_for_preferred_number_of_clusters.ipynb
###Markdown Variance for Preferred Number of ClustersIn this notebook, we conduct the sensitivity analysis for our `infomap` input parameter, the preferred number of clusters,as reported in the SI.In short, we investigate the question:How do the clusterings differ if we vary the preferred number of clusters? Preparations ###Code from cdlib import evaluation, readwrite import pandas as pd import matplotlib.pyplot as plt import seaborn as sns sns.set_style("darkgrid") plt.rcParams['figure.figsize'] = (12,9) plt.rcParams['font.size'] = 16 def make_dfs(dataset, preferred_cluster_sizes, years, config_strs): dfs = [] for year in years: config_str = config_strs[0] ref_clustering = readwrite.read_community_json( f'../../legal-networks-data/{dataset.lower()}/11_cluster_results/{year}_0-0_1-0_-1_a-infomap_n100_m1-0_s0_c1000.json' ) clusterings = [ readwrite.read_community_json( f'../../legal-networks-data/{dataset.lower()}/11_cluster_results/{year}_{config_str}.json' ) for config_str in config_strs ] df = pd.DataFrame( { 'Preferred Clusters': p, 'NMI': evaluation.normalized_mutual_information(ref_clustering, c).score, 'Rand': evaluation.adjusted_rand_index(ref_clustering, c).score, "Year": year } for p, c in zip(preferred_cluster_sizes, clusterings) ) dfs.append(df) return dfs def make_boxplot(dfs, pivot_col, y_label, save_path=None): pd.concat(dfs).pivot(columns='Preferred Clusters', index='Year', values=pivot_col ).boxplot(notch=0, sym="",color=dict(boxes='k', whiskers='k', medians='r', caps='k')) plt.ylabel(y_label) plt.xlabel('Preferred Number of Clusters') labels = plt.gca().get_xticklabels() labels[0] = 'auto' plt.gca().set_xticklabels(labels) plt.tight_layout() if save_path is not None: plt.savefig(save_path) ###Output _____no_output_____ ###Markdown Computing the statistics US ###Code dataset = 'us' preferred_cluster_sizes = list(range(0,150+1,10)) + [200] years = range(1994,2018+1) config_strs = [ f'0-0_1-0_-1_a-infomap_n{preferred_cluster_size}_m1-0_s0_c1000' if preferred_cluster_size else '0-0_1-0_-1_a-infomap_m1-0_s0_c1000' for preferred_cluster_size in preferred_cluster_sizes ] dfs = make_dfs(dataset, preferred_cluster_sizes, years, config_strs) make_boxplot(dfs, 'NMI', 'Normalized Mutual Information', '../graphics/preferred_number_of_modules_nmi_us.pdf') make_boxplot(dfs, 'Rand', 'Adjusted Rand Index', '../graphics/preferred_number_of_modules_rand_us.pdf') ###Output _____no_output_____ ###Markdown DE ###Code dataset = 'de' preferred_cluster_sizes = list(range(0,150+1,10)) + [200] years = [f'{year}-01-01' for year in range(1994,2018+1)] config_strs = [ f'0-0_1-0_-1_a-infomap_n{preferred_cluster_size}_m1-0_s0_c1000' if preferred_cluster_size else '0-0_1-0_-1_a-infomap_m1-0_s0_c1000' for preferred_cluster_size in preferred_cluster_sizes ] dfs = make_dfs(dataset, preferred_cluster_sizes, years, config_strs) make_boxplot(dfs, 'NMI', 'Normalized Mutual Information', '../graphics/preferred_number_of_modules_nmi_de.pdf') make_boxplot(dfs, 'Rand', 'Adjusted Rand Index', '../graphics/preferred_number_of_modules_rand_de.pdf') ###Output _____no_output_____
lab00_aprendendo_python.ipynb
###Markdown Lab 0: Introdução ao Google Colab Esse Notebook estará disponível em 1. Jupyter notebooksThis webpage is called a Jupyter notebook. A notebook is a place to write programs and view their results. 1.1. Text cellsIn a notebook, each rectangle containing text or code is called a *cell*.Text cells (like this one) can be edited by double-clicking on them. They're written in a simple format called [Markdown](http://daringfireball.net/projects/markdown/syntax) to add formatting and section headings. You don't need to learn Markdown, but you might want to.After you edit a text cell, select the "run cell" button at the top that looks like ▶| to confirm any changes. **Question 1.1.1.** This paragraph is in its own text cell. Try editing it so that **this** sentence is the last sentence in the paragraph, and then select the "run cell" ▶| button on the top. This sentence, for example, should be deleted. So should this one. ###Code print("Hello, World!") ###Output _____no_output_____ ###Markdown And this one: ###Code print("\N{WAVING HAND SIGN}, \N{EARTH GLOBE ASIA-AUSTRALIA}!") ###Output _____no_output_____ ###Markdown The fundamental building block of Python code is an expression. Cells can contain multiple lines with multiple expressions. When you run a cell, the lines of code are executed in the order in which they appear. Every `print` expression prints a line. Run the next cell and notice the order of the output. ###Code print("First this line is printed,") print("and then this one.") ###Output _____no_output_____ ###Markdown **Question 1.2.1.** Change the cell above so that it prints out: First this line, then the whole 🌏, and then this one.*Hint:* If you're stuck on how to print the Earth symbol, try looking at the print expressions above. 1.3. Writing Jupyter notebooksYou can use Jupyter notebooks for your own projects or documents. They are among the world's most popular programming environments for data science. When you make your own notebook, you'll need to create your own cells for text and code.To add a cell, select the + button in the menu bar. A new cell starts out as text. You can change it to a code cell by selecting it so that it's highlighted, then selecting the drop-down box next to the restart (⟳) button in the menu bar, and choosing Code instead of Markdown.**Question 1.3.1.** Add a code cell below this one. Write code in it that prints out: A whole new cell! ♪🌏♪(That musical note symbol is like the Earth symbol. Its long-form name is `\N{EIGHTH NOTE}`.)Run your cell to verify that it works. 1.4. ErrorsPython is a language, and like natural human languages, it has rules. It differs from natural language in two important ways:1. The rules are *simple*. You can learn most of them in a few weeks and gain reasonable proficiency with the language in a semester.2. The rules are *rigid*. If you're proficient in a natural language, you can understand a non-proficient speaker, glossing over small mistakes. A computer running Python code is not smart enough to do that.Whenever you write code, you'll make mistakes. When you run a code cell that has errors, Python will sometimes produce error messages to tell you what you did wrong.Errors are okay; even experienced programmers make many errors. When you make an error, you just have to find the source of the problem, fix it, and move on.We have made an error in the next cell. Run it and see what happens. ###Code print("This line is missing something." ###Output _____no_output_____ ###Markdown You should see something like this (minus our annotations):The last line of the error output attempts to tell you what went wrong. The *syntax* of a language is its structure, and this `SyntaxError` tells you that you have created an illegal structure. "`EOF`" means "end of file," so the message is saying Python expected you to write something more (in this case, a right parenthesis) before finishing the cell.There's a lot of terminology in programming languages. You'll learn as you go. If you are ever having trouble understanding an error message, search the discussion forum. If you don't find an answer, post a question about the error yourself.Try to fix the code above so that you can run the cell and see the intended message instead of an error. 1.5. The KernelThe kernel is a program that executes the code inside your notebook and outputs the results. In the top right of your window, you can see a circle that indicates the status of your kernel. If the circle is empty (⚪), the kernel is idle and ready to execute code. If the circle is filled in (⚫), the kernel is busy running some code. You may run into problems where your kernel is stuck for an excessive amount of time, your notebook is very slow and unresponsive, or your kernel loses its connection. If this happens, try the following steps:1. At the top of your screen, select **Kernel**, then **Interrupt**.2. If that doesn't help, select **Kernel**, then **Restart**. If you do this, you will have to run your code cells from the start of your notebook up until where you paused your work.3. If that doesn't help, restart your server. First, save your work by selecting **File** at the top left of your screen, then **Save and Checkpoint**. Next, select **Control Panel** at the top right. Choose **Stop My Server** to shut it down, then **My Server** to start it back up. Then, navigate back to the notebook you were working on. Submitting Assignments and the Grader In the course, you may have to submit assignments for grading. To submit assignments, you can navigate to the right side of the toolbar and select the button labeled `Submit` as shown below. Please make sure to save your notebook before submitting. To grade assignments in the course, there will be a grader called Gofer Grader. When you submit assignments, the grader goes through your assignment and grades it. At the bottom of an assignment, under the Submission header, you may see some lines of code that look like this: These lines of code import the grader and run all of the tests in the notebook. You can see what the tests are in the notebook by identifying the code cells with the format `check(test/...)`. These tests/checks are placed after a question usually, in order to see if you have gotten the question correct. If the question is correct, you should see something that looks similar to this:If you have gotten the question wrong and therefore "failed" the check, you will see something like this:If you get this result, make sure to go back to your code that was being tested and change it before submitting! 1.6. Completing a labAll assignments in the course will be distributed as notebooks like this one. At the top of each assignment, you'll see a cell like the one below that imports autograder tests. Run it to import the autograder tests. ###Code # Don't change this cell, just run it # Import autograder tests from gofer.ok import check ###Output _____no_output_____ ###Markdown When you finish a question, you need to check your answer by running the check command below. It's OK to grade multiple times; Gofer will only try to grade your final submission for each question. There are no hidden autograder tests. If you pass all the given autograder tests for a question, you will receive full credit for that question. ###Code check("tests/q0.py") ###Output _____no_output_____ ###Markdown The notebook resides on a server that is run by the course staff, and so we have access to it as well. Once you're finished with a lab, use the File menu within the notebook page (below the Jupyter logo) to "Save and Checkpoint" and you're done. You may also check your notebook in its entirety with the following command. ###Code import glob from gofer.ok import check_all display(check_all(glob.glob('tests/q*.py'))) ###Output _____no_output_____
8_automated_hyperparameter_search.ipynb
###Markdown Búsqueda de HiperparámetrosLas redes neuronales tienen decenas de hiperparámetros que afectan su arquitectura y proceso de entrenamiento. Más aún, el desempeño final del modelo está condicionado a encontar un conjunto de valores para dichos hiperparámetros exitosos, para una inicialización aleatoria de los pesos dada. Por ello, la exploración de hiperparámetros se vuelve una de las partes más tediosas y críticas del entrenamiento de redes neuronales. Para obtener resultados que sean correctos, significativos y reproducibles es necesario planificar y sistemizar este proceso de búsqueda. > hyper-parameter optimization should be regarded as a formal outer loop in the learning processFormalmente, este proceso se puede describir como la minimización de la función de pérdida (o maximizar la performance) como si fuera una función de *caja negra* que toma como parámetros los valores de los hiperparámetros:$$ f(\theta) = loss_\theta(y, \hat{y}) $$$$ \theta^* = argmin_\theta f(\theta) $$donde $\theta$ es el conjunto de hiperparámetros del modelo, $loss$ es la pérdida generada entre las etiquetas verdaderas $y$ y las etiquetas generadas por el modelo $\hat{y}$, y $f$ es la función objetivo de la minimización.Las estrategias principales para la exploración del espacio de hiperparámetros son:* Búsqueda manual, donde un humano define los valores de cada hiperparámetro.* Búsqueda por grilla o *grid search*, donde se define un conjunto de valores posibles que puede tomar cada hiperparámetro, y se realiza un experimento por cada combinación posible.* Búsqueda aleatoria o *random search*, donde se define un rango de valores posibles para cada hiperparámetro, y se elige al azar un valor del rango para cada experimento.* Búsqueda automátizada, *automated search* o *model-based search*, que es igual a la búsqueda aleatoria pero la selección del valor de cada hiperparámetro está condicionado por los resultados de experimentos anteriores. Para más información ver el paper [*Algorithms for Hyper-Parameter Optimization*](https://proceedings.neurips.cc/paper/2011/file/86e8f7ab32cfd12577bc2619bc635690-Paper.pdf)En la siguiente imagen, tomada del paper [*Random Search for Hyper-Parameter Optimization*](https://www.jmlr.org/papers/volume13/bergstra12a/bergstra12a.pdf), se muestra el impacto de las primeras dos estrategias para un hiperparámetro con alta influencia en el desempeño del modelo final, y otro que sin influencia. No solo require muchas evaluaciones para lograr cobertura, sino que las combinaciones en dónde sólo se varían hiperparámetros no relevantes no recolectan información nueva. El éxito de la búsqueda por grilla depende de que el nivel de granularidad de la grilla cubra adecuadamente los valores relevantes, que son desconocidos a priori.![Comparación de las exploraciones entre grid search y random search](https://res.cloudinary.com/dyd911kmh/image/upload/f_auto,q_auto:best/v1531340388/grid_vs_random_jltknd.png) Para solucionar todos estos problemas, es que se utiliza la **exploración bayesiana**. Este método modela la loss como un Gaussian process, y tiene en cuenta los resultados de los experimentos anteriores para ir construyendo una distribución de probabilidad de la pérdida dados los hiperparámetros:$$ P(loss | \theta)$$Para elegir una nueva combinación de hiperparámetros a probar dados los experimentos previos, el algoritmo utiliza una *surrogate function* para aproximar el comportamiento de la pérdida y una *selection function* basada en la mejora esperada. A grandes rasgos, el algoritmo sigue los siguientes pasos: 1. Encontrar el mejor conjunto de hiperparámetros que maximize la mejora esperada (EI), estimada a través de la *surrogate function*. 2. Calcular la performance del modelo con la combinación de hiperparámetros elegida. Esto corresponde a evaluar la función objetivo. 3. Actualizar la forma de la *surrogate function* utilizando el teorema de Bayes para que se ajuste mejor a la verdadera distribución $ P(loss | \theta)$Afortunadamente, muchos algoritmos de búsqueda están implementados y funcionan como cajas negras. Veremos un ejemplo utilizando la librería hyperopt ###Code # If running in colab, you need to update gensim # !pip install --upgrade gensim import csv import functools import gzip import numpy as np import pandas as pd import torch import torch.nn as nn import torch.nn.functional as F import tempfile import seaborn from gensim import corpora from gensim.models import KeyedVectors from gensim.parsing import preprocessing from gensim.scripts.glove2word2vec import glove2word2vec from sklearn import metrics from sklearn.model_selection import train_test_split from torch.utils.data import Dataset, DataLoader, IterableDataset from tqdm.notebook import tqdm, trange # Ensure version 4.X import gensim gensim.__version__ ###Output _____no_output_____ ###Markdown Parte 1: Preprocesamiento del textoPrimero leeremos el dataset como se explica en la notebook 5_cnns.ipynb. ###Code # If necessary, download data # %%bash # mkdir data # curl -L https://cs.famaf.unc.edu.ar/\~ccardellino/resources/diplodatos/glove.6B.50d.txt.gz -o ./data/glove.6B.50d.txt.gz # curl -L https://cs.famaf.unc.edu.ar/\~ccardellino/resources/diplodatos/imdb_reviews.csv.gz -o ./data/imdb_reviews.csv.gz class IMDBReviewsDataset(Dataset): def __init__(self, dataset, transform=None): self.dataset = dataset self.transform = transform def __len__(self): return self.dataset.shape[0] def __getitem__(self, item): if torch.is_tensor(item): item = item.to_list() item = { "data": self.dataset.loc[item, "review"], "target": self.dataset.loc[item, "sentiment"] } if self.transform: item = self.transform(item) return item class RawDataProcessor: def __init__(self, dataset, ignore_header=True, filters=None, vocab_size=50000): if filters: self.filters = filters else: self.filters = [ lambda s: s.lower(), preprocessing.strip_tags, preprocessing.strip_punctuation, preprocessing.strip_multiple_whitespaces, preprocessing.strip_numeric, preprocessing.remove_stopwords, preprocessing.strip_short, ] # Create dictionary based on all the reviews (with corresponding preprocessing) self.dictionary = corpora.Dictionary( dataset["review"].map(self._preprocess_string).tolist() ) # Filter the dictionary and compactify it (make the indices continous) self.dictionary.filter_extremes(no_below=2, no_above=1, keep_n=vocab_size) self.dictionary.compactify() # Add a couple of special tokens self.dictionary.patch_with_special_tokens({ "[PAD]": 0, "[UNK]": 1 }) self.idx_to_target = sorted(dataset["sentiment"].unique()) self.target_to_idx = {t: i for i, t in enumerate(self.idx_to_target)} def _preprocess_string(self, string): return preprocessing.preprocess_string(string, filters=self.filters) def _sentence_to_indices(self, sentence): return self.dictionary.doc2idx(sentence, unknown_word_index=1) def encode_data(self, data): return self._sentence_to_indices(self._preprocess_string(data)) def encode_target(self, target): return self.target_to_idx[target] def __call__(self, item): if isinstance(item["data"], str): data = self.encode_data(item["data"]) else: data = [self.encode_data(d) for d in item["data"]] if isinstance(item["target"], str): target = self.encode_target(item["target"]) else: target = [self.encode_target(t) for t in item["target"]] return { "data": data, "target": target, "sentence": item["data"] } ###Output _____no_output_____ ###Markdown Separando el conjunto de validación o *dev*En deep learning, es **MUY** importante utilizar un conjunto de validación durante la búsqueda de hiperparámetros, que puede ser tomado de la partición de entrenamiento. Esto es independiente de la estrategia de búsqueda que se utilice.De esta manera, se previene el overfitting indirecto y se cuenta con una partición de datos nunca antes vista para poder evaluar la generalización real del modelo a datos no vistos. ###Code dataset = pd.read_csv("./data/imdb_reviews.csv.gz") preprocess = RawDataProcessor(dataset) train_indices, test_indices = train_test_split(dataset.index, test_size=0.2, random_state=42) train_indices, dev_indices = train_test_split(train_indices, test_size=0.2, random_state=42) train_dataset = IMDBReviewsDataset(dataset.loc[train_indices].reset_index(drop=True), transform=preprocess) dev_dataset = IMDBReviewsDataset(dataset.loc[dev_indices].reset_index(drop=True), transform=preprocess) # We won't use test_dataset until the end! test_dataset = IMDBReviewsDataset(dataset.loc[test_indices].reset_index(drop=True), transform=preprocess) class PadSequences: def __init__(self, pad_value=0, max_length=100): self.pad_value = pad_value self.max_length = max_length def __call__(self, items): data, target = list(zip(*[(item["data"], item["target"]) for item in items])) seq_lengths = [len(d) for d in data] max_length = self.max_length seq_lengths = [min(self.max_length, l) for l in seq_lengths] data = [d[:l] + [self.pad_value] * (max_length - l) for d, l in zip(data, seq_lengths)] return { "data": torch.LongTensor(data), "target": torch.FloatTensor(target) } ###Output _____no_output_____ ###Markdown Parte 2: Esqueleto de la red neuronalDefinimos el modelo a entrenar. ###Code import torch import torch.nn as nn class ImdbLSTM(nn.Module): def __init__(self, pretrained_embeddings_path, dictionary, embedding_size, hidden_layer=32, num_layers=1, dropout=0., bias=True, bidirectional=False, freeze_embedings=True): super(ImdbLSTM, self).__init__() output_size = 1 # Create the Embeddings layer and add pre-trained weights embeddings_matrix = torch.randn(len(dictionary), embedding_size) embeddings_matrix[0] = torch.zeros(embedding_size) with gzip.open(pretrained_embeddings_path, "rt") as fh: for line in fh: word, vector = line.strip().split(None, 1) if word in dictionary.token2id: embeddings_matrix[dictionary.token2id[word]] =\ torch.FloatTensor([float(n) for n in vector.split()]) self.embedding_config = {'freeze': freeze_embedings, 'padding_idx': 0} self.embeddings = nn.Embedding.from_pretrained( embeddings_matrix, **self.embedding_config) # Set our LSTM parameters self.lstm_config = {'input_size': embedding_size, 'hidden_size': hidden_layer, 'num_layers': num_layers, 'bias': bias, 'batch_first': True, 'dropout': dropout if num_layers > 1 else 0.0, 'bidirectional': bidirectional} # Set our fully connected layer parameters self.linear_config = {'in_features': hidden_layer, 'out_features': output_size, 'bias': bias} # Instanciate the layers self.lstm = nn.LSTM(**self.lstm_config) self.droupout_layer = nn.Dropout(dropout) self.classification_layer = nn.Linear(**self.linear_config) self.activation = nn.Sigmoid() def forward(self, inputs): emb = self.embeddings(inputs) lstm_out, _ = self.lstm(emb) # Take last state of lstm, which is a representation of # the entire text lstm_out = lstm_out[:, -1, :].squeeze() lstm_out = self.droupout_layer(lstm_out) predictions = self.activation(self.classification_layer(lstm_out)) return predictions ###Output _____no_output_____ ###Markdown Encapsularemos el algoritmo de entrenamiento dentro de una función parametrizable. La función debería devolver los resultados obtenidos. ###Code # Some default values EPOCHS = 2 MAX_SEQUENCE_LEN = 100 import torch.optim as optim def train_imbd_model(train_dataset, dev_dataset, pretrained_embeddings_path, dictionary, embedding_size, batch_size=128, max_sequence_len=MAX_SEQUENCE_LEN, hidden_layer=32, dropout=0., epochs=EPOCHS, lr=0.001, optimizer_class=optim.Adam, verbose=False): if verbose: print_fn = print else: print_fn = lambda *x: None # We define again the data loaders since this code could run in # parallel pad_sequeces = PadSequences(max_length=max_sequence_len) train_loader = DataLoader(train_dataset, batch_size=batch_size, shuffle=True, collate_fn=pad_sequeces, drop_last=False) dev_loader = DataLoader(dev_dataset, batch_size=batch_size, shuffle=False, collate_fn=pad_sequeces, drop_last=False) # We are not going to explore all hyperparameters, only this ones. model = ImdbLSTM(pretrained_embeddings_path, dictionary, embedding_size, hidden_layer=hidden_layer, dropout=dropout) loss_function = nn.BCELoss() optimizer = optimizer_class(model.parameters(), lr) history = { 'train_loss': [], 'test_loss': [], 'test_avp': [] } for epoch in range(epochs): model.train() running_loss = [] print_fn("Epoch", epoch) for idx, batch in enumerate(train_loader): optimizer.zero_grad() output = model(batch["data"]) loss_value = loss_function(output.squeeze(), batch["target"]) loss_value.backward() optimizer.step() running_loss.append(loss_value.item()) train_loss = sum(running_loss) / len(running_loss) print_fn("\t Final train_loss", train_loss) history['train_loss'].append(train_loss) model.eval() running_loss = [] targets = [] predictions = [] for batch in dev_loader: output = model(batch["data"]) running_loss.append( loss_function(output.squeeze(), batch["target"]).item() ) targets.extend(batch["target"].numpy()) # Round up model output to get the predictions. # What would happen if you change the activation to tanh? predictions.extend(output.squeeze().round().detach().numpy()) test_loss = sum(running_loss) / len(running_loss) avp = metrics.average_precision_score(targets, predictions) print_fn("\t Final test_loss", test_loss) print_fn("\t Final test_avp", avp) history['test_loss'].append(test_loss) history['test_avp'].append(avp) return history history = train_imbd_model( train_dataset, dev_dataset, pretrained_embeddings_path="./data/glove.6B.50d.txt.gz", dictionary=preprocess.dictionary, embedding_size=50, verbose=True) ###Output Epoch 0 Final train_loss 0.6751024463176727 Final test_loss 0.61183714488196 Final test_avp 0.6281198166028745 Epoch 1 Final train_loss 0.633317717552185 Final test_loss 0.6387771065272982 Final test_avp 0.606606246189802 ###Markdown Utilizando hyperoptPara utilizar alguno de los algoritmos de hyperopt, es necesario definir una función objetivo que será minimizada. Esta función recibe un objeto con los valores para los hiperparámetros de cada experimento, y debe devolver una única métrica (o un diccionario con la clave `key` asociada a dicha métrica). En nuestro caso, utilizaremos el *average precision score* obtenido en el conjunto de validación.Les recomendamos consultar el [Tutorial oficial](https://github.com/hyperopt/hyperopt/wiki/FMin) para más detalles. ###Code from hyperopt import STATUS_OK # define an objective function def objective_fn(args): print("Exploring config:", args) # These references a train_dataset and dev_dataset are # taken from the globa context! history = train_imbd_model( train_dataset, dev_dataset, pretrained_embeddings_path="./data/glove.6B.50d.txt.gz", dictionary=preprocess.dictionary, embedding_size=50, **args) # This is the value that will be minimized! history['loss'] = history['test_avp'][-1] * -1 # These are required keys history['status'] = STATUS_OK return history from hyperopt import hp, fmin, tpe, Trials # define a search space space = { 'lr': hp.loguniform('lr', numpy.log(0.0001), numpy.log(0.005)), # see appendix 'optimizer_class': hp.choice( 'optimizer_class', [optim.Adam, optim.RMSprop]), 'dropout': hp.uniform('dropout', 0, 0.5) } # define the Trials object, which will allow us to store # information from every experiment. trials = Trials() # minimize the objective over the space best = fmin(objective_fn, space, algo=tpe.suggest, max_evals=10, trials=trials) print("Best hyperparameters:") print(best) # We can see the results of each experiment with the trials object. trials.results ###Output _____no_output_____ ###Markdown Recomendaciones finales* Es recomendable utilizar un parámetro de *paciencia*, que corta el ciclo de entrenamiento cuando no detecta mejoras en el desempeño sobre el conjunto de validación por n cantidad de épocas. Esto ayudaría a evitar que el modelo sobreajuste.* Realizar una búsqueda de grilla previa para determinar los valores para el optimizador, learning rate, batch size y número de épocas mínimas de entrenamiento, ya que estos son hiperparámetros muy determinantes.* No es necesario realizar la búsqueda de hiperparámetros sobre el conjunto de datos entero, ni entrenar el clasificador durante todas las epocas hasta que comienza a diverger. Se puede utilizar para encontrar los espacios más prometedores de valores posibles, y luego realizar una segunda búsqueda con con menos iteraciones pero con el proceso de entrenamiento completo.* No realizar la búsqueda utilizando notebooks, sino scripts.* Combinar hyperopt con mlflow para un registro de los resultados ordenado.* Modificar el bucle de entrenamiento para guardar el último modelo con las mejores métricas en el conjunto de validación. ###Code ###Output _____no_output_____ ###Markdown Apéndice: hp.loguniformSegún la documentación oficial, la distribución `hp.loguniform`:* Returns a value drawn according to exp(uniform(low, high)) so that the logarithm of the return value is uniformly distributed.* When optimizing, this variable is constrained to the interval [exp(low), exp(high)].Supongamos que queremos que nuestros valores de lr se distribuyan logaritmicamente en el intervalo [0.0001, 0.005], entonces los valores de low y high deberían ser: log(0.0001) y log(0.005). Veamos qué distribución de muestras obtenemos. ###Code import numpy import seaborn low = numpy.log(0.0001) high = numpy.log(0.005) sample_size = 1000 sample = numpy.exp(numpy.random.uniform(low, high, size=sample_size)) seaborn.displot(sample) sample.max(), sample.min() ###Output _____no_output_____
analyses/Correlations.ipynb
###Markdown Correlations This file contains all of the correlations that we want to calculate. This means that we need at least 2 columns to create a result.Most of these correlations have to do with coop salary, or term average. We use these as metrics of student success because they're numeric, making them easier to process. Other properties are very subjective. Salary and grades are not the most indicative of how successful a student is, but with the existing data, it's the best indication we have. ###Code from IPython.display import display import pandas as pd import numpy as np import matplotlib.pyplot as plt import math import json from collections import defaultdict from Bucket import Bucket from Distribution import Distribution from GradeSalaryHistory import GradeSalaryHistory pd.set_option('display.max_columns', None) plt.style.use('ggplot') # Show matplotlib plots in this notebook %matplotlib inline # Setting plot parameters from pylab import rcParams params = { 'figure.figsize': (8, 8), 'legend.fontsize': 15 } rcParams.update(params) def isnan(a): return a != a def isempty(a): return isnan(a) or not a def isfloat(value): try: float(value) return not isnan(value) except ValueError: return False # Where to write the buckets BUCKET_DIR = '../private/buckets/' GRADE_SALARY_FILENAME = '../private/grade-vs-salary-temp.json' df = pd.read_csv('../private/results-04-10.csv') # TODO: Write the response file COOP = ['1', '2', '3', '4', '5', '6'] TERM = ['1a', '1b', '2a', '2b', '3a', '3b', '4a'] MULTI_VAL_COL = ['ethnicity', 'fav_lang', 'preferred_tech_discipline', 'text_editor'] SALARY_COL = ['coop_salary_' + i + '.csv' for i in COOP] GRADE_COL = ['term_avg_' + i + '.csv' for i in TERM] def write_buckets(df, col_name): """Creates a bucket for each value and then writes the values into a file.""" buckets, aggregate_values = correlate_columns(df, col_name) with open(BUCKET_DIR + col_name + '_buckets.json', 'w') as f: result = {} for i in buckets: result[i] = buckets[i].summary() for i in aggregate_values: result[i] = aggregate_values[i] f.write(json.dumps(result, indent=True)) def correlate_columns(df, col_name): """Generates a bucket for unique column value.""" # Get unique values (some rows have multiple values) col_values = np.array([]) for i in df[col_name]: if isfloat(i): # Floor any data i = math.floor(float(i)) if isnan(i): # Don't want to include any NaN print('Skipping ', i) continue # Append separate values if it's a comma separated value val = str(i) if ',' in val and col_name in MULTI_VAL_COL: col_values = np.append(col_values, list(map(str.strip, val.split(',')))) else: col_values = np.append(col_values, val) col_values = np.unique(col_values) # Create a bucket for each column value buckets = {} aggregate_values = { 'first_half_grades': defaultdict(list), 'second_half_grades': defaultdict(list), 'first_half_salaries': defaultdict(list), 'second_half_salaries': defaultdict(list) } for col_val in col_values: # Initialize distributions salaries = {} grades = {} for i in TERM: grades[i] = np.array([]) for i in COOP: salaries[i] = np.array([]) # Iterate through all rows for i in range(0, df.shape[0]): val = df[col_name][i] if isfloat(val): val = math.floor(float(val)) # Filter any NaN or non matching values if isnan(val): continue if col_name in MULTI_VAL_COL: if str(col_val) not in map(str.strip, str(val).split(',')): continue else: if str(col_val) != str(val): continue first_half_grade = 0 second_half_grade = 0 first_half_salary = 0 second_half_salary = 0 # Add grades fhgc = 0 # first half grade count shgc = 0 # second half grade count fhsc = 0 shsc = 0 for j in range(0, len(TERM)): t = TERM[j] avg = df['term_avg_' + t][i] if avg == 'exchange': continue if isnan(avg): continue grades[t] = np.append(grades[t], avg) if j < 4: first_half_grade += float(avg) fhgc += 1 else: second_half_grade += float(avg) shgc += 1 # Add salary for j in range(0, len(COOP)): c = COOP[j] salary = df['coop_salary_' + c][i] if type(salary) == float and math.isnan(salary): continue if type(salary) == str and not salary: continue if isempty(salary): # Not sure if this is needed continue if type(salary) == str and ',' in salary: salary = salary.replace(',', '') salaries[c] = np.append(salaries[c], salary) if j < 3: first_half_salary += float(salary) fhsc += 1 else: second_half_salary += float(salary) shsc += 1 if col_val == 'Prefer not to say': # Only one person, don't show their data. continue if fhgc > 0: aggregate_values['first_half_grades'][col_val].append(first_half_grade / fhgc) if shgc > 0: aggregate_values['second_half_grades'][col_val].append(second_half_grade / shgc) if fhsc > 0: aggregate_values['first_half_salaries'][col_val].append(first_half_salary / fhsc) if shsc > 0: aggregate_values['second_half_salaries'][col_val].append(second_half_salary / shsc) # Create the bucket buckets[col_val] = Bucket(col_name, col_val, [Distribution(grades[i].astype(float)) for i in TERM], [Distribution(salaries[i].astype(float)) for i in COOP]) return buckets, aggregate_values # From https://github.com/se2018/class-profile/tree/master/analyses to_correlate = [ 'gender', 'ethnicity', 'family_income', 'work_os', 'phone', 'soft_eng_rating', 'se_friendships', 'is_international', 'parents_edu', 'parents_technical', 'admission_avg', 'code_start_age', 'fav_lang', 'num_hackathons', 'side_proj', 'exercise', 'cooking', 'sleep_time', 'preferred_tech_discipline', 'text_editor' ] for i in to_correlate: write_buckets(df, i) ###Output ('Skipping ', nan) ('Skipping ', nan) ('Skipping ', nan) ('Skipping ', nan) ('Skipping ', nan) ('Skipping ', nan) ###Markdown Next section is to get the correlation between grades and coop jobs. ###Code def create_history(df): """Creates an entry for each coop term about the grades and salaries leading up to it.""" result = {} for i, term in enumerate(COOP): result[term] = defaultdict(list) for row in range(0, df.shape[0]): # Skip any entries that are missing data on the coop if isempty(df['coop_name_' + term][row]) or isempty(df['coop_salary_' + term][row]): print('Empty term for row ', row, ', term ', term, 'skipping entry.') continue # Process a coop term term_avgs = np.array([]) salaries = np.array([]) # Get previous grades for study in range(0, i+1): val = df['term_avg_' + str(TERM[study])][row] if isnan(val) or val == 'exchange': term_avgs = np.append(term_avgs, 0.0) else: term_avgs = np.append(term_avgs, math.floor(float(val))) # Get previous salaries for coop in range(0, i): val = str(df['coop_salary_' + COOP[coop]][row]) val = val.replace(',', '') if isnan(float(val)): salaries = np.append(salaries, 0) else: salaries = np.append(salaries, float((int(float(val)) / 500 * 500))) location = df['coop_loc_' + term][row] if isempty(location): location = '' salary = df['coop_salary_' + term][row] # Estimate to the nearest 500*i salary = float(int(float(str(salary).replace(',', ''))) / 500 * 500) result[term][str(salary)].append(GradeSalaryHistory(term_avgs, salaries, location, salary).summary()) return result with open(GRADE_SALARY_FILENAME, 'w') as f: f.write(json.dumps(create_history(df), indent=True)) ###Output ('Empty term for row ', 5, ', term ', '1', 'skipping entry.') ('Empty term for row ', 30, ', term ', '1', 'skipping entry.') ('Empty term for row ', 45, ', term ', '1', 'skipping entry.') ('Empty term for row ', 55, ', term ', '1', 'skipping entry.') ('Empty term for row ', 60, ', term ', '1', 'skipping entry.') ('Empty term for row ', 67, ', term ', '1', 'skipping entry.') ('Empty term for row ', 93, ', term ', '1', 'skipping entry.') ('Empty term for row ', 7, ', term ', '2', 'skipping entry.') ('Empty term for row ', 45, ', term ', '2', 'skipping entry.') ('Empty term for row ', 64, ', term ', '2', 'skipping entry.') ('Empty term for row ', 82, ', term ', '3', 'skipping entry.') ('Empty term for row ', 91, ', term ', '3', 'skipping entry.') ('Empty term for row ', 100, ', term ', '3', 'skipping entry.') ('Empty term for row ', 8, ', term ', '4', 'skipping entry.') ('Empty term for row ', 73, ', term ', '4', 'skipping entry.') ('Empty term for row ', 107, ', term ', '4', 'skipping entry.') ('Empty term for row ', 112, ', term ', '4', 'skipping entry.') ('Empty term for row ', 26, ', term ', '5', 'skipping entry.') ('Empty term for row ', 77, ', term ', '5', 'skipping entry.') ('Empty term for row ', 86, ', term ', '5', 'skipping entry.') ('Empty term for row ', 58, ', term ', '6', 'skipping entry.') ('Empty term for row ', 64, ', term ', '6', 'skipping entry.') ('Empty term for row ', 74, ', term ', '6', 'skipping entry.') ('Empty term for row ', 87, ', term ', '6', 'skipping entry.') ###Markdown Check for gender vs SE rating ###Code gender_rating = defaultdict(list) for i in range(0, df.shape[0]): gender_rating[df['gender'][i]].append(df['soft_eng_rating'][i]) del gender_rating['Prefer not to say'] # There's only one entry stats = {} stats['mean'] = { 'Male': np.mean(gender_rating['Male']), 'Female': np.mean(gender_rating['Female']) } stats['stddev'] = { 'Male': np.std(gender_rating['Male']), 'Female': np.std(gender_rating['Female']) } gender_rating['stats'] = stats with open('../private/gender_rating.json', 'w') as f: f.write(json.dumps(gender_rating, indent=True, default=str)) ###Output _____no_output_____
notebooks/Milestone5-Data_Analysis.ipynb
###Markdown Milestone 5 - Data Analysis Jarod Research QuestionDoes the release of triple A multiplayer fps titles effect the player numbers of CS:GO? Data Analysis From the SteamChartz dataset, I extracted the rows about Counter-Strike: Global Offensive, the most popular game on Steam, to process the data and wrangle the average players, gained average players from the previous month, and the four-month rolling gain average and the seven-month rolling gain average for every month recorded in the dataset. This data allows us to have a timeline of the average and the gain. All we need to do with the timeline is see if there are any dips in the average players that are not a part of a longer pre-existing trend and compare them to the release dates of triple-A titles like Call of Duty, Battlefield, Halo, Overwatch, and Valorant.From exploring the Dashboard, I can extract that only Overwatch and Valorant had significant losses in players, both of which fit into a similar niche that Counter-Strike: Global Offensive occupies. Valorant more closely competes with Counter-Strike: Global Offensive, as their gameplay is heavily inspired by it; this is further indicated by the roughly 186,000 players that drop Counter-Strike: Global Offensive in the two months leading up to the release of Valorant. While the release of Call of Duty and Battlefield can reduce the player numbers of Counter-Stike: Global Offensive, they do not frequently occur early on and, in general, are not substantial decreases in player numbers compared to the average players. Lastly, Halo didn’t seem to affect the player numbers of Counter-Strike: Global Offensive as the trend didn’t fluctuate when it was released. From these observations, I can conclude that while you can see decreases in players with the release of other games, most of them are insignificant compared to the total players. Furthermore, the Counter-Strike: Global Offensive community is a dedicated and isolated gaming community with few transient players, only losing many players due to competition in their niche(i.e. Valorant). Harshal ###Code import seaborn as sns import matplotlib.pyplot as plt from project_functions2 import load_process_covid ###Output _____no_output_____ ###Markdown Plot ###Code df_covid_sorted = load_process_covid('../data/raw/SteamCharts.csv') df_covid_sorted sns.barplot(x ='Month', y= 'AvgGainPerGame', hue = 'Year', data = df_covid_sorted) ###Output _____no_output_____ ###Markdown Final Research Question Analysis (Harshal - Task 4)Research Question: How has Covid-19 affected the active users of games on steam? I used the data set to find the average gain of players on steam across all games on the platform at the time to see what the AveageGainPerGame on steam is, to show that during the months when covid first started the surge of lockdowns across the world, that more players had joined during this time since everyone was at home and were not allowed to go outside. I processed the raw dataset into specific points in time of the specific years so January, February, March of 2019, to show the pre-covid and January, February, March of 2020 to show during covid. I used a barplot to show the Months and AvgGainPerGame as x and y values for the plot, also groupingthe years 2019 and 2020 for each month so you can compare the differences in both AvgGainPerGame in that specific month.As you can see that during the months in the year 2020, the bars are above the x-axis, meaning they are in positive AvgGain, but when you compare it to 2019bars which mostly seem to be negative or significantly lower than the 2020 bars. This indicates many gains in players on the platform steamwas when covid was at large during its initial months of 2020. Showing that covid has impacted the steam platforms gamer counts due to many peoplefinding new modes of entertainment via video games! Lance ###Code import matplotlib.pylab as plt import seaborn as sns import numpy as np from project_functions3 import load_csgo sns.set_theme(style="ticks", font_scale=1.3, # This scales the fonts slightly higher ) plt.rc("axes.spines", top=False, right=False) ###Output _____no_output_____ ###Markdown Plot ###Code df_csgo = load_csgo('../data/raw/Steamcharts.csv', '../data/raw/Twitch_game_data.csv') df_csgo sns.barplot(x = 'Month', y = 'Avg_players', data = df_csgo, estimator=np.median) sns.barplot(x = 'Month', y = 'Avg_viewers', data = df_csgo, estimator=np.median) ###Output _____no_output_____
documents/presentation-9/script.ipynb
###Markdown Statistical Analysis of Data Environment SettingsAn statistical Analysis of the data captured will be performed.The environment configuration is the following:- A rectangle area is used whose dimension is 3 x 3 meters. - A custom robot similar to an epuck was used.- The robot starts in the middle of the arena.- The robot moves in a random fashion way around the environment avoiding obstacles for 100 robot steps then it is moved into another random location.- The data is not normalized in this experiment.- The robot has 8 sensors that measure the distance between the robot and the walls.- Some noise was introduced in the sensors measurements of the robot using the concept of [lookup tables](https://cyberbotics.com/doc/reference/distancesensor) in the Webots simulator which according to Webots documentation "The first column of the table specifies the input distances, the second column specifies the corresponding desired response values, and the third column indicates the desired standard deviation of the noise. The noise on the return value is computed according to a gaussian random number distribution whose range is calculated as a percent of the response value (two times the standard deviation is often referred to as the signal quality)". The following values were taken: - (0, 0, 0.05) - (10, 10, 0.05) - The simulator runs during 25 hours of simulation (~30 minutes in fast mode). ###Code # Install a pip package in the current Jupyter kernel import sys !{sys.executable} -m pip install scikit-learn !{sys.executable} -m pip install keras import pandas as pd import tensorflow as tf import numpy as np import math from sklearn.ensemble import RandomForestRegressor from keras import models from keras import layers from keras import regularizers import matplotlib.pyplot as plt from keras import optimizers ###Output Requirement already satisfied: scikit-learn in /usr/local/lib/python3.7/site-packages (0.22) Requirement already satisfied: joblib>=0.11 in /usr/local/lib/python3.7/site-packages (from scikit-learn) (0.14.1) Requirement already satisfied: numpy>=1.11.0 in /usr/local/lib/python3.7/site-packages (from scikit-learn) (1.16.1) Requirement already satisfied: scipy>=0.17.0 in /Users/sebastiangerard/Library/Python/3.7/lib/python/site-packages (from scikit-learn) (1.4.1) Requirement already satisfied: keras in /usr/local/lib/python3.7/site-packages (2.3.1) Requirement already satisfied: scipy>=0.14 in /Users/sebastiangerard/Library/Python/3.7/lib/python/site-packages (from keras) (1.4.1) Requirement already satisfied: keras-preprocessing>=1.0.5 in /Users/sebastiangerard/Library/Python/3.7/lib/python/site-packages (from keras) (1.1.0) Requirement already satisfied: numpy>=1.9.1 in /usr/local/lib/python3.7/site-packages (from keras) (1.16.1) Requirement already satisfied: h5py in /Users/sebastiangerard/Library/Python/3.7/lib/python/site-packages (from keras) (2.9.0) Requirement already satisfied: six>=1.9.0 in /usr/local/lib/python3.7/site-packages (from keras) (1.12.0) Requirement already satisfied: keras-applications>=1.0.6 in /Users/sebastiangerard/Library/Python/3.7/lib/python/site-packages (from keras) (1.0.8) Requirement already satisfied: pyyaml in /usr/local/lib/python3.7/site-packages (from keras) (5.2) ###Markdown First Experiment ###Code csv_file = 'robot_info_dataset.csv' df = pd.read_csv(csv_file) df[['x', 'y', 'theta', 'sensor_1', 'sensor_2','sensor_3','sensor_4','sensor_5','sensor_6','sensor_7', 'sensor_8']].head() ###Output _____no_output_____ ###Markdown Data pre-processing The data collected 1125965 samples. ###Code df.shape df = df.sample(frac=1) df = df[:1125965] df.shape ###Output _____no_output_____ ###Markdown The data set contains some null values so they should be deleted from the samples. ###Code df = df.dropna() ###Output _____no_output_____ ###Markdown Input and output variables The data will be split into training, testing and validation sets. 60% of the data will be used for training, 20% for training and 20% of validation. ###Code # train size test_size_percentage = .2 train_size_percentage = .8 ds_size = df.shape[0] train_size = int(train_size_percentage * ds_size) test_size = int(test_size_percentage * ds_size) # shuffle dataset sampled_df = df.sample(frac=1) # separate inputs from outputs inputs = sampled_df[['x', 'y', 'theta']] targets = sampled_df[['sensor_1', 'sensor_2', 'sensor_3', 'sensor_4', 'sensor_5', 'sensor_6', 'sensor_7', 'sensor_8']] # train train_inputs = inputs[:train_size] train_targets = targets[:train_size] # test test_inputs = inputs[train_size:] test_targets = targets[train_size:] inputs.head() ###Output _____no_output_____ ###Markdown Neural Network As input the neural network receives the x, y coordinates and rotation angle $\theta$. The output are the sensor measurements. One model per sensor will be created. ###Code def get_model(): # neural network with a 10-neuron hidden layer model = models.Sequential() model.add(layers.Dense(16, activation='relu', input_shape=(3,))) # model.add(layers.Dropout(0.5)) model.add(layers.Dense(32, activation='relu')) model.add(layers.Dense(64, activation='relu')) model.add(layers.Dense(16, activation='relu')) model.add(layers.Dense(1)) # rmsprop = optimizers.RMSprop(learning_rate=0.01) model.compile(optimizer='rmsprop', loss='mse', metrics=['mae']) return model def k_fold(sensor_number, num_epochs=10, k=5): num_val_samples = len(train_inputs) // k validation_scores = [] histories = [] nmse = [] for i in range(k): print('processing fold #', i) val_data = train_inputs[i * num_val_samples: (i + 1) * num_val_samples] val_targets = train_targets[[sensor_number]][i * num_val_samples: (i + 1) * num_val_samples] partial_train_data = np.concatenate( [train_inputs[:i * num_val_samples], train_inputs[(i + 1) * num_val_samples:]], axis=0) partial_train_targets = np.concatenate( [train_targets[[sensor_number]][:i * num_val_samples], train_targets[[sensor_number]][(i + 1) * num_val_samples:]], axis=0) model = get_model() history = model.fit(partial_train_data, partial_train_targets, validation_data=(val_data, val_targets), epochs=num_epochs, batch_size=64, verbose=1) histories.append(history.history) predictions_targets = model.predict(val_data) nmse.append(np.mean((predictions_targets - val_targets)**2)/np.var(val_targets)) return histories, nmse histories, nmse = k_fold('sensor_3', 50, 3) print("NMSE: ") print(np.mean(nmse)) num_epochs = 50 val_mae_history = [np.mean([x['val_mae'][i] for x in histories]) for i in range(num_epochs)] mae_history = [np.mean([x['mae'][i] for x in histories]) for i in range(num_epochs)] plt.plot(range(3, len(val_mae_history) + 1), val_mae_history[2:], 'ro') plt.plot(range(3, len(mae_history) + 1), mae_history[2:], 'bo') plt.xlabel('Epochs') plt.ylabel('MAE') plt.legend(['test', 'train'], loc='upper left') plt.show() val_loss_history = [np.mean([x['val_loss'][i] for x in histories]) for i in range(num_epochs)] loss_history = [np.mean([x['loss'][i] for x in histories]) for i in range(num_epochs)] plt.plot(range(1, len(val_loss_history) + 1), val_loss_history, 'ro') plt.plot(range(1, len(loss_history) + 1), loss_history, 'bo') plt.xlabel('Epochs') plt.ylabel('LOSS') plt.legend(['test', 'train'], loc='upper left') plt.show() model = get_model() history = model.fit(inputs, targets[['sensor_5']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_5.h5") model = get_model() history = model.fit(inputs, targets[['sensor_6']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_6.h5") model = get_model() history = model.fit(inputs, targets[['sensor_7']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_7.h5") model = get_model() history = model.fit(inputs, targets[['sensor_8']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_8.h5") model = get_model() history = model.fit(inputs, targets[['sensor_1']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_1.h5") model = get_model() history = model.fit(inputs, targets[['sensor_2']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_2.h5") model = get_model() history = model.fit(inputs, targets[['sensor_3']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_3.h5") model = get_model() history = model.fit(inputs, targets[['sensor_4']], epochs=50, batch_size=64, verbose=1) history.history['mae'] model.save("nn_sensor_4.h5") ###Output Epoch 1/50 1125965/1125965 [==============================] - 25s 23us/step - loss: 0.3168 - mae: 0.3743 Epoch 2/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.1257 - mae: 0.2427 Epoch 3/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0994 - mae: 0.2072 Epoch 4/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0871 - mae: 0.1889 Epoch 5/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0799 - mae: 0.1777 Epoch 6/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0756 - mae: 0.1710 Epoch 7/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0720 - mae: 0.1655 Epoch 8/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0695 - mae: 0.1617 Epoch 9/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0678 - mae: 0.1588 Epoch 10/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0658 - mae: 0.1558 Epoch 11/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0640 - mae: 0.1530 Epoch 12/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0632 - mae: 0.1512 Epoch 13/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0622 - mae: 0.1496 Epoch 14/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0608 - mae: 0.1475 Epoch 15/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0601 - mae: 0.1457 Epoch 16/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0593 - mae: 0.1445 Epoch 17/50 1125965/1125965 [==============================] - 26s 24us/step - loss: 0.0588 - mae: 0.1436 Epoch 18/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0579 - mae: 0.1418 Epoch 19/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0577 - mae: 0.1415 Epoch 20/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0571 - mae: 0.1407 Epoch 21/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0566 - mae: 0.1397 Epoch 22/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0562 - mae: 0.1390 Epoch 23/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0556 - mae: 0.1380 Epoch 24/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0552 - mae: 0.1374 Epoch 25/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0545 - mae: 0.1366 Epoch 26/50 1125965/1125965 [==============================] - 27s 24us/step - loss: 0.0543 - mae: 0.1363 Epoch 27/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0538 - mae: 0.1355 Epoch 28/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0535 - mae: 0.1348 Epoch 29/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0530 - mae: 0.1341 Epoch 30/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0526 - mae: 0.1334 Epoch 31/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0524 - mae: 0.1330 Epoch 32/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0520 - mae: 0.1327 Epoch 33/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0517 - mae: 0.1323 Epoch 34/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0514 - mae: 0.1317 Epoch 35/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0513 - mae: 0.1313 Epoch 36/50 1125965/1125965 [==============================] - 27s 24us/step - loss: 0.0511 - mae: 0.1312 Epoch 37/50 1125965/1125965 [==============================] - 29s 26us/step - loss: 0.0507 - mae: 0.1308 Epoch 38/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0505 - mae: 0.1303 Epoch 39/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0503 - mae: 0.1301 Epoch 40/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0500 - mae: 0.1296 Epoch 41/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0497 - mae: 0.1293 Epoch 42/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0495 - mae: 0.1289 Epoch 43/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0491 - mae: 0.1283 Epoch 44/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0490 - mae: 0.1281 Epoch 45/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0488 - mae: 0.1278 Epoch 46/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0486 - mae: 0.1275 Epoch 47/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0486 - mae: 0.1274 Epoch 48/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0484 - mae: 0.1270 Epoch 49/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0485 - mae: 0.1273 Epoch 50/50 1125965/1125965 [==============================] - 26s 23us/step - loss: 0.0480 - mae: 0.1269
Univariate/ML/NYCT/UnivariateML_NYCT_(OCSVM+XGBoost).ipynb
###Markdown Helper Function ###Code def getScaledTrainTextDataset(dataset_scaled, trainRate=0.3): print('Shape: ',dataset_scaled.shape) train_size = int(len(dataset_scaled)*trainRate) test_size = len(dataset_scaled) - train_size print('Trainsize:',train_size, ' - Testsize:',test_size) train_start_index,train_end_index = 0,train_size test_start_index,test_end_index = train_size,len(dataset_scaled) train = dataset_scaled[0:train_size] test = dataset_scaled[train_size:] return train,test # convert an array of values into a dataset matrix def create_XY_lookback_dataset(dataset, look_back=1): dataX, dataY = [], [] for i in range(len(dataset)-look_back-1): a = dataset[i:(i+look_back)] dataX.append(a) dataY.append(dataset[i + look_back]) return numpy.array(dataX), numpy.array(dataY) def create_XY_lookback_dataset_multistepOutput(dataset, look_back=1, look_forward = 1): dataX, dataY = [], [] for i in range(len(dataset)-look_back-1-look_forward): a = dataset[i:(i+look_back), 0] dataX.append(a) dataY.append(dataset[i + look_back:i+look_back+look_forward, 0]) return numpy.array(dataX), numpy.array(dataY) def calculateRMSE(testPredict,trainY,testY,inverse_transform=True, verbose= 1): if inverse_transform: testPredict_inverse = scaler.inverse_transform(testPredict) testY_inverse = scaler.inverse_transform([testY]) else: testPredict_inverse = testPredict testY_inverse = testY # calculate root mean squared error testScore = math.sqrt(mean_squared_error(testY_inverse[0], testPredict_inverse[:,0])) if verbose == 1: print('Test Score: %.2f RMSE' % (testScore)) print('Persistent Model Testscore small:',global_testPredict_small, ' - Persistent Model Testscore big:', global_testPredict_big) return testScore def calculateRMSE_MultipleOutput(trainPredict,testPredict,trainY,testY,inverse_transform=True, verbose= 1): if inverse_transform: trainPredict_inverse = scaler.inverse_transform(trainPredict) trainY_inverse = scaler.inverse_transform(trainY) testPredict_inverse = scaler.inverse_transform(testPredict) testY_inverse = scaler.inverse_transform(testY) else: trainPredict_inverse = trainPredict trainY_inverse = trainY testPredict_inverse = testPredict testY_inverse = testY # calculate root mean squared error trainScore = math.sqrt(mean_squared_error(trainY_inverse[0], trainPredict_inverse[:,0])) testScore = math.sqrt(mean_squared_error(testY_inverse[0], testPredict_inverse[:,0])) if verbose == 1: print('Train Score: %.2f RMSE' % (trainScore)) print('Test Score: %.2f RMSE' % (testScore)) print('Persistent Model Testscore small:',global_testPredict_small, ' - Persistent Model Testscore big:', global_testPredict_big) return trainScore, testScore def plotErrorPrediction(dataset_scaled, testPredict, trainPredict, ShowTestError=True, inverse_transform=True): if inverse_transform: trainPredict_inverse = scaler.inverse_transform(trainPredict) testPredict_inverse = scaler.inverse_transform(testPredict) else: trainPredict_inverse = trainPredict testPredict_inverse = testPredict # shift train predictions for plotting trainPredictPlot = numpy.zeros_like(dataset_scaled) trainPredictPlot[:, :] = numpy.nan trainPredictPlot[look_back:len(trainPredict)+look_back, :] = trainPredict_inverse # shift test predictions for plotting testPredictPlot = numpy.empty_like(dataset_scaled) testPredictPlot[:, :] = numpy.nan testPredictPlot[len(trainPredict)+(look_back*2)+1:len(dataset)-1, :] = testPredict_inverse error_test = dataset-testPredictPlot error_test[np.isnan(error_test)] = 0. # error_test[error_training<2000] = 0. error_test = np.abs(error_test) import seaborn as sns sns.distplot(error_test[error_test!=0]) print(pandas.DataFrame(error_test[error_test!=0]).describe()) return error_test, trainPredictPlot,testPredictPlot def plotErrorPredictionValidation(dataset_scaled, validationPredict, testPredict, trainPredict, ShowTestError=True, inverse_transform=True): if inverse_transform: trainPredict_inverse = scaler.inverse_transform(trainPredict) testPredict_inverse = scaler.inverse_transform(testPredict) else: trainPredict_inverse = trainPredict testPredict_inverse = testPredict # shift train predictions for plotting trainPredictPlot = numpy.zeros_like(dataset_scaled) trainPredictPlot[:, :] = numpy.nan trainPredictPlot[look_back:len(trainPredict)+look_back, :] = trainPredict_inverse # shift test predictions for plotting testPredictPlot = numpy.empty_like(dataset_scaled) testPredictPlot[:, :] = numpy.nan testPredictPlot[len(trainPredict)+(look_back*2)+1+len(validationPredict):len(dataset)-1, :] = testPredict_inverse error_test = dataset-testPredictPlot error_test[np.isnan(error_test)] = 0. # error_test[error_training<2000] = 0. error_test = np.abs(error_test) import seaborn as sns sns.distplot(error_test[error_test!=0]) print(pandas.DataFrame(error_test[error_test!=0]).describe()) return error_test, trainPredictPlot,testPredictPlot ###Output _____no_output_____ ###Markdown OCSVM ###Code import warnings from sklearn.cluster import KMeans from statsmodels.tsa.ar_model import AR from sklearn.metrics import mean_squared_error from math import sqrt from pandas import read_csv import pandas as pd from pandas import DataFrame from pandas import concat import numpy as np from matplotlib import pyplot from sklearn.svm import OneClassSVM from sklearn import preprocessing import sys class OneClassSVM_AnomalyDetection: def __init__(self,path, window_width, nu, train_rate): self.df = read_csv(path, header=0, index_col=0, parse_dates=True,squeeze=True) self.df = self.df.reset_index(drop=True) self.df.rename(columns={'anomaly':'is_anomaly'}, inplace=True) self.nu = nu self.window_width = window_width series = pd.DataFrame(self.df.iloc[:,0].values) self.values = DataFrame(series.values) self.dataframe = concat([self.values.shift(1), self.values], axis=1) self.dataframe.columns = ['t', 't+1'] self.train_size = int(len(self.values) * train_rate) # train_labeled, test_labeled = self.dataframe.values[1:self.train_size], self.dataframe.values[self.train_size:] # self.train_X, self.train_y = train_labeled[:,0], train_labeled[:,1] # self.test_X, self.test_y = test_labeled[:,0], test_labeled[:,1] # self.create_persistence() # X = series.values # self.train, self.test = X[1:self.train_size], X[self.train_size:] def __build_sets(self): train_labeled, test_labeled = self.dataframe.values[1:self.train_size], self.dataframe.values[self.train_size:] self.train_X, self.train_y = train_labeled[:,0], train_labeled[:,1] self.test_X, self.test_y = test_labeled[:,0], test_labeled[:,1] X = self.dataframe.iloc[:,1].values self.train, self.test = X[1:self.train_size], X[self.train_size:] def standardize_dataframe(self): X = self.dataframe.values self.scalar = preprocessing.StandardScaler().fit(X) X = self.scalar.transform(X) self.dataframe = pd.DataFrame(X) def inverse_standardize_dataframe(self): X = self.dataframe.values X = self.scalar.inverse_transform(X) self.dataframe = pd.DataFrame(X) def model_persistence(self, x): return x def create_persistence(self): rmse = sqrt(mean_squared_error(self.dataframe['t'].iloc[self.train_size:], self.dataframe['t+1'].iloc[self.train_size::])) # print('Persistent Model RMSE: %.3f' % rmse) def fit(self): self.create_persistence() self.standardize_dataframe() self.__build_sets() self.compute_anomalyScores() self.inverse_standardize_dataframe() def getWindowedVectors(self, X): vectors = [] for i,_ in enumerate(X[:-self.window_width+1]): vectors.append(X[i:i+self.window_width]) return vectors def compute_anomalyScores(self): self.errors = np.zeros_like(self.test) # compute anomalies warnings.filterwarnings("ignore") # history = self.getWindowedVectors(self.train) for i,_ in enumerate(self.test[:-self.window_width+1]): sys.stdout.write('\r'+str(i)+':'+str(len(self.test) - self.window_width)) window = self.test[i:i+self.window_width] window2D = np.zeros((len(window),2)) window2D[:,1] = window clf=OneClassSVM(nu=self.nu) clf.fit(window2D) error = clf.decision_function(window2D) error[error>0] = 0 self.errors[i:i+self.window_width] += error*-10 # normalize anomaly score self.errors[:-self.window_width+1] /= self.window_width for i,error in enumerate(self.test[-self.window_width+1:]): self.errors[-self.window_width + 1 + i] /=self.window_width-(i+1) # self.errors_original = self.errors # scalar = preprocessing.MinMaxScaler((0,1)).fit(self.errors.reshape(-1,1)) # self.errors = scalar.transform(self.errors.reshape(-1,1))*10 def plot(self): # plot predicted error pyplot.figure(figsize=(50,5)) pyplot.plot(self.test) # pyplot.plot(self.predictions, color='red') pyplot.plot(self.errors, color = 'red', linewidth=0.5) pyplot.show() def get_roc_auc(self, plot=True, verbose=True): # get the predicted errors of the anomaly points indices = self.df[self.df['is_anomaly']==1].index >self.train_size true_anomaly_predicted_errors = self.errors[self.df[self.df['is_anomaly']==1].index[indices] - self.train_size ] if len(true_anomaly_predicted_errors) == 0: return np.nan # sort them true_anomaly_predicted_errors = np.sort(true_anomaly_predicted_errors,axis=0).reshape(-1) true_anomaly_predicted_errors_extended = np.r_[np.linspace(0,true_anomaly_predicted_errors[0],40)[:-1],true_anomaly_predicted_errors] true_anomaly_predicted_errors_extended = np.r_[true_anomaly_predicted_errors_extended, true_anomaly_predicted_errors_extended[-1] + np.mean(true_anomaly_predicted_errors_extended)] # now iterate thru the predicted errors from small to big # for each value look how much other points have equal or bigger error FPR = [] # fp/n https://en.wikipedia.org/wiki/Sensitivity_and_specificity TPR = [] # tp/p p = len(true_anomaly_predicted_errors) Thresholds = [] for predictederror in true_anomaly_predicted_errors_extended: threshold = predictederror tp = len(true_anomaly_predicted_errors[true_anomaly_predicted_errors>= threshold]) fp = len(self.errors[self.errors>=threshold])-len(true_anomaly_predicted_errors[true_anomaly_predicted_errors>=threshold]) fpr =fp/len(self.errors) FPR.append(fpr) TPR.append(tp/p) if verbose: print("Threshold: {0:25} - FP: {1:4} - TP: {2:4} - FPR: {3:21} - TPR: {4:4}".format(threshold,fp, tp, fpr, tp/p)) import matplotlib.pyplot as plt if plot: plt.figure() plt.axis([0, 1, 0, 1]) plt.plot(FPR,TPR) plt.show() # This is the AUC from sklearn.metrics import auc print('AUC: ' ,auc(FPR,TPR) ) return auc(FPR,TPR) # iforest = OneClassSVM_AnomalyDetection('Univariate/YahooServiceNetworkTraffic/A1Benchmark/real_1.csv',30,0.7,0.3) # iforest.fit() # iforest.plot() # iforest.get_roc_auc(verbose=False) ###Output _____no_output_____ ###Markdown Evaluation Results of NYCT ###Code import os import datetime startTime = datetime.datetime.now() import glob cnn = OneClassSVM_AnomalyDetection( 'drive/My Drive/MT/Experiments/Univariate/NYC_Taxi/nyc_taxi.csv', 40,0.1,0.3) cnn.fit() cnn.get_roc_auc(plot=False,verbose=False) endTime = datetime.datetime.now() diff = endTime - startTime print('Time: ',diff) cnn = OneClassSVM_AnomalyDetection( 'drive/My Drive/MT/Experiments/Univariate/NYC_Taxi/nyc_taxi.csv', 40,0.1,0.3) startTime = datetime.datetime.now() cnn.fit() endTime = datetime.datetime.now() diff = endTime - startTime print('Inference Time: ',diff) ###Output 7184:7184AUC: 0.5210562102565175 Time: 0:00:06.942960 7184:7184Inference Time: 0:00:07.040107 ###Markdown XGBoost ###Code import warnings from sklearn.cluster import KMeans from statsmodels.tsa.ar_model import AR from sklearn.metrics import mean_squared_error from math import sqrt from pandas import read_csv import pandas as pd from pandas import DataFrame from pandas import concat import numpy as np from matplotlib import pyplot from xgboost import XGBRegressor from sklearn import preprocessing import sys class XGBRegressor_AnomalyDetection: def __init__(self,path, window_width, nu, train_rate): self.df = read_csv(path, header=0, index_col=0, parse_dates=True,squeeze=True) self.df = self.df.reset_index(drop=True) self.df.rename(columns={'anomaly':'is_anomaly'}, inplace=True) self.nu = nu self.window_width = window_width series = pd.DataFrame(self.df.iloc[:,0].values) self.values = DataFrame(series.values) self.dataframe = concat([self.values.shift(1), self.values], axis=1) self.dataframe.columns = ['t', 't+1'] self.train_size = int(len(self.values) * train_rate) # train_labeled, test_labeled = self.dataframe.values[1:self.train_size], self.dataframe.values[self.train_size:] # self.train_X, self.train_y = train_labeled[:,0], train_labeled[:,1] # self.test_X, self.test_y = test_labeled[:,0], test_labeled[:,1] # self.create_persistence() # X = series.values # self.train, self.test = X[1:self.train_size], X[self.train_size:] def __build_sets(self): train_labeled, test_labeled = self.dataframe.values[1:self.train_size], self.dataframe.values[self.train_size:] self.train_X, self.train_y = train_labeled[:,0], train_labeled[:,1] self.test_X, self.test_y = test_labeled[:,0], test_labeled[:,1] X = self.dataframe.iloc[:,1].values self.train, self.test = X[1:self.train_size], X[self.train_size:] def standardize_dataframe(self): X = self.dataframe.values self.scalar = preprocessing.StandardScaler().fit(X) X = self.scalar.transform(X) self.dataframe = pd.DataFrame(X) def inverse_standardize_dataframe(self): X = self.dataframe.values X = self.scalar.inverse_transform(X) self.dataframe = pd.DataFrame(X) def model_persistence(self, x): return x def create_persistence(self): rmse = sqrt(mean_squared_error(self.dataframe['t'].iloc[self.train_size:], self.dataframe['t+1'].iloc[self.train_size::])) # print('Persistent Model RMSE: %.3f' % rmse) def fit(self): self.create_persistence() self.__build_sets() self.compute_anomalyScores() def getWindowedVectors(self, X): vectors = [] for i,_ in enumerate(X[:-self.window_width+1]): vectors.append(X[i:i+self.window_width]) return vectors def compute_anomalyScores(self): self.xgb = XGBRegressor() self.xgb.fit(self.train_X.reshape(-1,1),self.train_y.reshape(-1,1)) self.predictions = self.xgb.predict(self.test_X.reshape(-1,1)) rmse = sqrt(mean_squared_error(self.test, self.predictions)) self.errors = np.absolute(self.test - np.array(self.predictions)) # print('Prediction Test RMSE: %.3f' % rmse) def plot(self): # plot predicted error pyplot.figure(figsize=(50,5)) pyplot.plot(self.test) pyplot.plot(self.predictions, color='blue') pyplot.plot(self.errors, color = 'red', linewidth=0.5) pyplot.show() def get_roc_auc(self, plot=True, verbose=True): # get the predicted errors of the anomaly points indices = self.df[self.df['is_anomaly']==1].index >self.train_size true_anomaly_predicted_errors = self.errors[self.df[self.df['is_anomaly']==1].index[indices] - self.train_size ] if len(true_anomaly_predicted_errors) == 0: return np.nan # sort them true_anomaly_predicted_errors = np.sort(true_anomaly_predicted_errors,axis=0).reshape(-1) true_anomaly_predicted_errors_extended = np.r_[np.linspace(0,true_anomaly_predicted_errors[0],40)[:-1],true_anomaly_predicted_errors] true_anomaly_predicted_errors_extended = np.r_[true_anomaly_predicted_errors_extended, true_anomaly_predicted_errors_extended[-1] + np.mean(true_anomaly_predicted_errors_extended)] # now iterate thru the predicted errors from small to big # for each value look how much other points have equal or bigger error FPR = [] # fp/n https://en.wikipedia.org/wiki/Sensitivity_and_specificity TPR = [] # tp/p p = len(true_anomaly_predicted_errors) Thresholds = [] for predictederror in true_anomaly_predicted_errors_extended: threshold = predictederror tp = len(true_anomaly_predicted_errors[true_anomaly_predicted_errors>= threshold]) fp = len(self.errors[self.errors>=threshold])-len(true_anomaly_predicted_errors[true_anomaly_predicted_errors>=threshold]) fpr =fp/len(self.errors) FPR.append(fpr) TPR.append(tp/p) if verbose: print("Threshold: {0:25} - FP: {1:4} - TP: {2:4} - FPR: {3:21} - TPR: {4:4}".format(threshold,fp, tp, fpr, tp/p)) import matplotlib.pyplot as plt if plot: plt.figure() plt.axis([0, 1, 0, 1]) plt.plot(FPR,TPR) plt.show() # This is the AUC from sklearn.metrics import auc print('AUC: ' ,auc(FPR,TPR) ) return auc(FPR,TPR) ###Output _____no_output_____ ###Markdown Evaluation Results of NYCT ###Code import os import datetime startTime = datetime.datetime.now() import glob cnn = XGBRegressor_AnomalyDetection( 'drive/My Drive/MT/Experiments/Univariate/NYC_Taxi/nyc_taxi.csv', 1000,0.5,0.3) cnn.fit() cnn.get_roc_auc(plot=False,verbose=False) endTime = datetime.datetime.now() diff = endTime - startTime print('Time: ',diff) startTime = datetime.datetime.now() cnn.xgb.predict(cnn.test_X.reshape(-1,1)) endTime = datetime.datetime.now() diff = endTime - startTime print('Inference Time: ',diff) ###Output [03:44:49] WARNING: /workspace/src/objective/regression_obj.cu:152: reg:linear is now deprecated in favor of reg:squarederror. AUC: 0.4602617410866643 Time: 0:00:00.139282 Inference Time: 0:00:00.020563
notebooks/Tutorial2_Units+Database.ipynb
###Markdown AMPEL introAMPEL is software framework designed for processing heterogeneous streamed data. AMPEL was not developed to provide a specific scientific resource, but rather an environment where it is easy to ensure that a scientific program fulfills the strict requirement of the next generation real-time experiments: efficient and powerful analysis, where provenance and reproducibiltiy is paramount. In particular, to guarantee the last point requires algorithms (which make real-time deicsions) be separated from infrastructure (which will likely evolve with time and project phase).An AMPEL _user_ constructs a configuration file which describes every step of how an incoming alert stream should be processed. This can be broken down into selecting which _units_ should be executed, and which _parameters_ each of these should be provided. An AMPEL _live instance_ executes these units, based on the input data, as requested and stores all intermediate and final data in a databse. Provenance/reproducibility is ensured through multiple layers. First, each live instance is run from a container which can be retrieved later and together with a data archive replay the full stream. Second, AMPEL contains an extensive set of logs and a transient-specific _Journal_ which details all related actions/decisions. Finally, each unit and channel configuration file is drawn from a specific (tagged) github version. The series of notebooks provided here gradually builds toward a sample full configration. Sample science caseEach AMPEl _channel_ is designed with a science goal (or "hypothesis/test") in mind. A much discussed current topic is the origin of the extragalactic neutrino flux observed e.g. by IceCube, with one of the potential sources being supernovae interacting with circumstellar material (SNIIn). We here wish to investigate whether a particular subtype of these, SN2009ip-like SNe with recent previous outbursts, are regularly found within the uncertainty region of neutrino alerts. The steps for this science program would be: Identify transients with optical lightcurves compatible with SN2009ip AND which coincide with neutrino alerts. For such targets, obtain follow-up spectroscopy to confirm classification (i.e. an external reaction). This notebook - Tutorial 2This notebook will repeat the analysis presented in the first tutorial, but where all units are read from tagged (github) repositories. They can from there be referenced, be included in a computer center AMPEL instance and be distributed to the community.These can be found in the `ampel/contrib/sample/` subdirectories of the `Ampel-contrib-sample` repository.Furthermore, intermediate results will her be stored in a local MongoDB database and the AMPEL schedulers will be used to carry out the requested operations. This notebook thus assumes a mongod instance to be running and accessible through 27017. (The port can be changed through the `mongo` key of the `ampel_config.yml` file). ###Code import os %load_ext ampel_quick_import %qi DevAmpelContext AmpelLogger T2Processor T3Processor ChannelModel AlertProcessor TarAlertLoader ChannelModel AbsAlertFilter AMPEL_CONF = "../../ampel_config.yml" ALERT_ARCHIVE = '../sample_data/ztfpub_200917_pruned.tar.gz' # The operation context is created based on a setup configuration file. # db_prefix sets the DB name to use ctx = DevAmpelContext.load( config_file_path = AMPEL_CONF, db_prefix = "AmpelTutorial", purge_db = True, ) # A scientific program, a channel, is added ctx.add_channel( name="demo_SN09if", access=['ZTF', 'ZTF_PUB'] ) # The channel is constructed from two units, each controlled by parameters. # Lets start with the straightforward filter filter_conf = { 'min_rb':0.3, 'min_ndet':7, 'min_tspan':10, 'max_tspan' : 200, 'min_gal_lat':15, } filter_config_id = ctx.add_config_id( filter_conf ) # The template matching has now been moved into a separate unit: # T2SNcosmoComp # where we added some configurability. match_conf = { 'target_model_name':'v19-2009ip-corr', 'base_model_name':'salt2', 'chi2dof_cut':2., 'chicomp_scaling':0.5, } match_config_id = ctx.add_config_id( match_conf ) # A channel can specify which streams to read, how these should be combined and what units # should be run on each data combination. # This is provided as directives to the AlertProcessor, which besides processing the alerts # also submit tickets to the DB concerning further operations to execute for any transients # that pass the initial filter stage. ap = AlertProcessor( context = ctx, process_name = "ipyton_notebook_test", supplier = "ZiAlertSupplier", log_profile = "debug", directives = [ { "channel": "demo_SN09if", "filter": {"unit": "SimpleDecentFilterCopy","config": filter_config_id }, "stock_update": "ZiStockIngester", 't0_add': { "ingester": "ZiAlertContentIngester", "t1_combine": [ { "ingester": "PhotoCompoundIngester", "config": {"combiner": "ZiT1Combiner"}, "t2_compute": { "ingester": "PhotoT2Ingester", "config": {"tags": ["ZTF"]}, "units": [ {'unit': 'T2SNcosmoComp', 'config': match_config_id }, ] } } ], } } ] ) # Provide a link to the alert collection to use ap.set_loader(TarAlertLoader(file_path=ALERT_ARCHIVE)) ap.set_iter_max(1000) ap.run() t2p = T2Processor(context=ctx, process_name="T2Processor_test", log_profile="debug") t2p.run() ###Output _____no_output_____ ###Markdown So far an input data stream has been filtered, and some sort of calculation has been done on the accepted sample. The next step for a full channel is usually some sort of _reaction_. These can vary between sending immediate alarms (e.g. through Slack or GCN), triggering follow-up observations or propagating information (e.g. for inspection in a frontent such as SkyPortal). Such reactions take place in the T3 tier. A simple `T3HelloWorld` unit is used, but the `react` method can be configured to do most other things. Sample T3 units react with TNS, Slack, Dropbox, SkyPortal and GCN.Key steps of configuring the T3 procss comes through the `selection` directive, where we select tansients that produced the required target match, and the `execute` directive which regulates which T3 units are run. ###Code # Test base t3 = T3Processor( context=ctx, process_name = "T3Processor_test", log_profile = "default", # debug channel = "demo_SN09if", directives = [ { "select": { "unit": "T3FilteringStockSelector", "config": { 't2_filter': { 'unit': 'T2SNcosmoComp', 'match': {'target_match': True} }, } }, "load": { "unit": "T3SimpleDataLoader", "config": { "directives": ["TRANSIENT", "DATAPOINT", "COMPOUND", "T2RECORD"], } }, "run": { "unit": "T3UnitRunner", "config": { "directives": [ { "project": { "unit": "T3ChannelProjector", "config": { "channel": "demo_SN09if" } }, "execute": [ { "unit": "T3HelloWorld", "config": { 't2info_from' : ['T2SNcosmoComp'] }, }, ] } ] } } } ] ) t3.run() ###Output _____no_output_____
Notebooks/.ipynb_checkpoints/12.MachineLearningBulkModulusRandomForest-checkpoint.ipynb
###Markdown Predicting bulk moduli with matminer----------------------------------------------------------------------------- Fit data mining models to ~10,000 calculated bulk moduli from Materials Project-----------------------------------------------------------------------------**Time to complete: 30 minutes**The tutorial is based on the matminer tutorials from https://github.com/hackingmaterials/matminer_examples.This notebook is an example of using the MP data retrieval tool :code:`retrieve_MP.py` to retrieve computed bulk moduli from theMaterials Project database in the form of a pandas dataframe, using matminer's tools to populate the dataframe with descriptors/features from pymatgen, and then fitting regression models from the scikit-learn library to the dataset. OverviewIn this notebook, we will:1. Load and examine a dataset in a pandas dataframe2. Add descriptors to the dataframe using matminer3. Train, compare, and visualize two machine learning methods with scikit-learn and Plotly. Software installationTo run this and other Python notebooks, I recommend using Jupyter. Here are the steps to install JupyterLab:- If you do not have Conda already installed on your computer, please follow the instructions at https://docs.conda.io/en/latest/miniconda.html. Then you can proceed to install JupyterLab, which will also install Python if it is not already installed.- `conda install jupyterlab`- `conda install ipywidgets`For the visualization, we are using Plotly. To install it:- `conda install plotly`- Install Node.js from https://nodejs.org/en/download/- `jupyter labextension install [email protected]` 1. Load and process data setWe use matminer to load a data set of computed elastic properties of materials from MaterialsProject. 1.1 First load needed Python packages ###Code # filter warnings messages from the notebook import warnings warnings.filterwarnings('ignore') import numpy as np import pandas as pd # Set pandas view options pd.set_option('display.width', 800) pd.set_option('display.max_columns', None) pd.set_option('display.max_rows', None) # Set path for saved images import os if not os.path.exists("images"): os.mkdir("images") ###Output _____no_output_____ ###Markdown 1.1 Use matminer to obtain data from MP (automatically) in a "pandas" dataframe ###Code from matminer.data_retrieval.retrieve_MP import MPDataRetrieval api_key = None # Set your MP API key here. If set as an environment variable 'MAPI_KEY', set it to 'None' mpr = MPDataRetrieval(api_key) # Create an adapter to the MP Database. # criteria is to get all entries with elasticity (K_VRH is bulk modulus) data criteria = {'elasticity.K_VRH': {'$ne': None}} # properties are the materials attributes we want # See https://github.com/materialsproject/mapidoc for available properties you can specify properties = ['pretty_formula', 'spacegroup.symbol', 'elasticity.K_VRH', 'formation_energy_per_atom', 'band_gap', 'e_above_hull', 'density', 'volume', 'nsites'] # get the data! df_mp = mpr.get_dataframe(criteria=criteria, properties=properties) print('Number of bulk moduli extracted = ', len(df_mp)) ###Output 100%|██████████| 13172/13172 [00:10<00:00, 1231.82it/s] ###Markdown 1.2 Explore the dataset.The data set comes as a pandas DataFrame, which is a kind of "spreadsheet" object in Python. DataFrames have several useful methods you can use to explore and clean the data, some of which we'll explore below. ###Code df_mp.head() ###Output _____no_output_____ ###Markdown A pandas DataFrame includes a function called `describe()` that helps determine statistics for the various numerical / categorical columns in the data. ###Code df_mp.describe() ###Output _____no_output_____ ###Markdown Sometimes, the `describe()` function will reveal outliers that indicate mistakes in the data. For example, negative hence unphysical minimum bulk/shear moduli or maximum bulk/shear moduli that are too high.The data looks ok at first glance; meaning that there are no clear problems with the ranges of the various properties. Therefore, and we won't filter out any data.Note that the `describe()` function only describes numerical columns by default. 1.3 Filter out unstable entries and negative bulk moduliThe data set above has some entries that correspond to thermodynamically or mechanically unstable materials. We filter these materials out using the distance from the convex hull and `K_VRH` (the Voight-Reuss-Hill average of the bulk modulus). ###Code df = df_mp df = df[df['elasticity.K_VRH'] > 0] df = df[df['e_above_hull'] < 0.1] df.describe() ###Output _____no_output_____ ###Markdown 1.4 Add descriptors/featuresCreate a new desciptor for the volume per atom and add it to the pandas data frame. ###Code # add volume per atom descriptor df['vpa'] = df['volume']/df['nsites'] # explore columns df.head() ###Output _____no_output_____ ###Markdown 1.5 Add several more descriptors using MatMiner’s pymatgen descriptor getter tools ###Code from matminer.featurizers.composition import ElementProperty from matminer.utils.data import PymatgenData from pymatgen.core import Composition df["composition"] = df['pretty_formula'].map(lambda x: Composition(x)) dataset = PymatgenData() descriptors = ['row', 'group', 'atomic_mass', 'atomic_radius', 'boiling_point', 'melting_point', 'X'] stats = ["mean", "std_dev"] ep = ElementProperty(data_source=dataset, features=descriptors, stats=stats) df = ep.featurize_dataframe(df, "composition") #Remove NaN values df = df.dropna() df.head() ###Output _____no_output_____ ###Markdown 2. Fit a Linear Regression model using SciKitLearn 2.1 Define what column is the target output, and what are the relevant descriptorsThe data set above has many columns - we won't need all this data for our modeling. We'll mainly be trying to predict `K_VRH` and `G_VRH` (the Voight-Reuss-Hill average of the bulk and shear modulus, respectively) and the `elastic_anisotropy`. We can drop most of the other output data. ###Code # target output column y = df['elasticity.K_VRH'].values # possible descriptor columns excluded = ["elasticity.K_VRH", "pretty_formula", "volume", "nsites", "spacegroup.symbol", "e_above_hull", "composition"] X = df.drop(excluded, axis=1) print("There are {} possible descriptors:\n\n{}".format(X.shape[1], X.columns.values)) ###Output _____no_output_____ ###Markdown 2.2 Fit the linear regression model ###Code from sklearn.linear_model import LinearRegression from sklearn.metrics import mean_squared_error import numpy as np lr = LinearRegression() lr.fit(X, y) # get fit statistics print('Training R2 = ' + str(round(lr.score(X, y), 3))) print('Training RMSE = %.3f' % np.sqrt(mean_squared_error(y_true=y, y_pred=lr.predict(X)))) ###Output _____no_output_____ ###Markdown 2.3 Cross validate the results ###Code from sklearn.model_selection import KFold, cross_val_score, cross_val_predict # Use 10-fold cross validation (90% training, 10% test) crossvalidation = KFold(n_splits=10, shuffle=True) # compute cross validation scores for random forest model r2_scores = cross_val_score(lr, X, y, scoring='r2', cv=crossvalidation, n_jobs=1) print(r2_scores) mse_scores = cross_val_score(lr, X, y, scoring='neg_mean_squared_error', cv=crossvalidation, n_jobs=1) rmse_scores = [np.sqrt(abs(s)) for s in mse_scores] print('Cross-validation results:') print('Folds: %i, mean R2: %.3f' % (len(r2_scores), np.mean(np.abs(r2_scores)))) print('Folds: %i, mean RMSE: %.3f' % (len(rmse_scores), np.mean(np.abs(rmse_scores)))) ###Output _____no_output_____ ###Markdown 2.4 Scatter density plot the results with Plotly and kernel density estimate¶ ###Code import plotly.graph_objects as PlotlyFig from scipy import stats # a line to represent a perfect model with 1:1 prediction xy_params = {'x_col': [0, 400], 'y_col': [0, 400], 'color': 'black', 'mode': 'lines', 'legend': None, 'text': None, 'size': None} xx=y yy=lr.predict(X) # Calculate the point density kde = stats.gaussian_kde([xx,yy]) zz = kde([xx,yy]) # Sort the points by density, so that the densest points are plotted last idx = zz.argsort() xx, yy, z = xx[idx], yy[idx], zz[idx] fig = PlotlyFig.Figure(data=PlotlyFig.Scattergl( x=xx, y=yy, mode='markers', marker=dict( size=5, color=z, #set color equal to a variable colorscale='Viridis', # one of plotly colorscales ), text=df['pretty_formula'] )) fig.update_layout(xaxis_title='DFT (MP) bulk modulus (GPa)', yaxis_title='Predicted bulk modulus (GPa)', title='Linear regression', width=800, height=800, font=dict(family="Helvetica", size=18, color="black") ) fig.update_yaxes(scaleanchor="x") fig.add_trace(PlotlyFig.Scatter(x=[0,400], y=[0,400], mode='lines')) fig.update_layout(xaxis=dict(range=[0, 400]), yaxis=dict(range=[0, 400]), showlegend=False) fig.show() # Save image, can change format by simply changing the file suffix fig.write_image("images/LinearRegression.jpeg") ###Output _____no_output_____ ###Markdown Great! We just fit a linear regression model to pymatgen features using matminer and sklearn. Now let’s use a Random Forest model to examine the importance of our features. 3. Follow similar steps for a Random Forest model 3.1 Fit the Random Forest model, get R2 and RMSE ###Code from sklearn.ensemble import RandomForestRegressor rf.fit(X, y) print('R2 = ' + str(round(rf.score(X, y), 3))) print('RMSE = %.3f' % np.sqrt(mean_squared_error(y_true=y, y_pred=rf.predict(X)))) ###Output _____no_output_____ ###Markdown 3.2 Cross-validate the results ###Code # compute cross validation scores for random forest model scores = cross_val_score(rf, X, y, scoring='neg_mean_squared_error', cv=crossvalidation, n_jobs=-1) rmse_scores = [np.sqrt(abs(s)) for s in scores] print('Cross-validation results:') print('Folds: %i, mean RMSE: %.3f' % (len(scores), np.mean(np.abs(rmse_scores)))) ###Output _____no_output_____ ###Markdown 3.3 Plot the random forest model ###Code # a line to represent a perfect model with 1:1 prediction xy_params = {'x_col': [0, 400], 'y_col': [0, 400], 'color': 'black', 'mode': 'lines', 'legend': None, 'text': None, 'size': None} xx=y #yy=rf.predict(X) yy=cross_val_predict(rf, X, y, cv=crossvalidation) # Calculate the point density kde = stats.gaussian_kde([xx,yy]) zz = kde([xx,yy]) # Sort the points by density, so that the densest points are plotted last idx = zz.argsort() xx, yy, z = xx[idx], yy[idx], zz[idx] fig = PlotlyFig.Figure(data=PlotlyFig.Scattergl( x=xx, y=yy, mode='markers', marker=dict( size=5, color=z, #set color equal to a variable colorscale='Viridis', # one of plotly colorscales ), text=df['pretty_formula'] )) fig.update_layout(xaxis_title='DFT (MP) bulk modulus (GPa)', yaxis_title='Predicted bulk modulus (GPa)', title='Random Forest Model', width=800, height=800, font=dict(family="Helvetica", size=18, color="black") ) fig.update_yaxes(scaleanchor="x") fig.add_trace(PlotlyFig.Scatter(x=[0,400], y=[0,400], mode='lines')) fig.update_layout(xaxis=dict(range=[0, 400]), yaxis=dict(range=[0, 400]), showlegend=False) fig.show() # Save image, can change format by simply changing the file suffix fig.write_image("images/RandomForest.jpeg") ###Output _____no_output_____ ###Markdown This looks clearly better than the linear regression model and reflects the improved cross-validated R2 and RMSE.You could (optionally) visualize the training error by replacing the code `cross_val_predict(rf, X, y, cv=crossvalidation)` in the above cell with `rf.predict(X)`. That would look a better still, but would not be an accurate representation of your prediction error. 3.4 Visualize the distribution of training and testing errors ###Code from sklearn.model_selection import train_test_split X['pretty_formula'] = df['pretty_formula'] X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=1) train_formula = X_train['pretty_formula'] X_train = X_train.drop('pretty_formula', axis=1) test_formula = X_test['pretty_formula'] X_test = X_test.drop('pretty_formula', axis=1) rf_reg = RandomForestRegressor(n_estimators=50, random_state=1) rf_reg.fit(X_train, y_train) # get fit statistics print('training R2 = ' + str(round(rf_reg.score(X_train, y_train), 3))) print('training RMSE = %.3f' % np.sqrt(mean_squared_error(y_true=y_train, y_pred=rf_reg.predict(X_train)))) print('test R2 = ' + str(round(rf_reg.score(X_test, y_test), 3))) print('test RMSE = %.3f' % np.sqrt(mean_squared_error(y_true=y_test, y_pred=rf_reg.predict(X_test)))) import plotly.graph_objects as PlotlyFig data_train=y_train-rf_reg.predict(X_train) data_test=y_test-rf_reg.predict(X_test) fig = PlotlyFig.Figure() fig.add_trace(PlotlyFig.Histogram(x=data_train, name='Training', histnorm='probability', xbins=dict( start=-50.0, end=50, size=2 ))) fig.add_trace(PlotlyFig.Histogram(x=data_test, name='Testing', histnorm='probability', xbins=dict( start=-50.0, end=50, size=2 ))) fig.update_layout(xaxis_title='Bulk modulus prediction residual (GPa)', yaxis_title='Probability', title='Random Forest Regression Residuals', barmode='stack', width=1200, height=600, font=dict(family="Helvetica", size=18, color="black") ) fig.show() ###Output _____no_output_____ ###Markdown 3.5 Plot the importance of the features we usedLet's see what are the most important features used by the random forest model. ###Code import plotly.graph_objects as PlotlyFig importances = rf.feature_importances_ included = X.columns.values indices = np.argsort(importances)[::-1] fig = PlotlyFig.Figure(data=PlotlyFig.Bar( x=included[indices], y=importances[indices])) fig.update_layout(yaxis_title='Importance', title='Random Forest Model Features', width=1200, height=800, font=dict(family="Helvetica", size=18, color="black") ) fig.show() ###Output _____no_output_____ ###Markdown Features relating to the materials volume per atom and density and the melting and boiling points of the components are the most important in the random forest model.This concludes the tutorial! You are now familiar with some of the basic features of data retrieval and machine learning. ###Code # Import libraries from sklearn.kernel_ridge import KernelRidge krr = KernelRidge(alpha=1.0,kernel='rbf') krr.fit(X, y) print('R2 = ' + str(round(krr.score(X, y), 3))) print('RMSE = %.3f' % np.sqrt(mean_squared_error(y_true=y, y_pred=krr.predict(X)))) ###Output _____no_output_____
MelSpecVAE_v1.ipynb
###Markdown MelSpecVAE v.1.5 Author: Moisés Horta Valenzuela, 2021> Website: [moiseshorta.audio](https://moiseshorta.audio)> Twitter: [@hexorcismos](https://twitter.com/hexorcismos)```MelSpecVAE is a Variational Autoencoder which synthesizes Mel-Spectrograms thay can be inverted into raw audio waveform.Currently you can train it with any dataset of .wav audio at 44.1khz Sample Rate and 16bit bitdepth.> Features:* Interpolate through 2 different points in the latent space and synthesize the 'in between' sounds.* Generate short one-shot audio* Synthesize arbitrarily long audio samples by generating seeds and sampling from the latent space.> Credits:* VAE neural network architecture coded following 'The Sound of AI' Youtube tutorial series by Valerio Velardo* Some utility functions from Marco Passini's MelGAN-VC Jupyter Notebook.> Last update:* 12.12.2021: Added experimental feature 'Timbre Transfer'``` ###Code #@title Mount Google Drive from google.colab import drive drive.mount('/content/drive') ###Output _____no_output_____ ###Markdown Run the next cells first for training, generating or timbre transfer ###Code #@title Import Tensorflow and torchaudio !pip install tensorflow-gpu==2.0.0 !pip install h5py==2.10.0 --force-reinstall !pip install soundfile #to save wav files !pip install --no-deps torchaudio==0.5 !pip install git+https://github.com/pvigier/perlin-numpy #for generating perlin and fractal noise #@title Import libraries from glob import glob import librosa import librosa.display import matplotlib.pyplot as plt import numpy as np from numpy import asarray from numpy.random import randn from numpy.random import randint from numpy import linspace import soundfile as sf import time import IPython import tensorflow as tf from perlin_numpy import ( generate_fractal_noise_2d, generate_perlin_noise_2d, ) #@title Hyperparameters learning_rate = 0.001 #@param {type:"raw"} num_epochs_to_train = 10#@param {type:"integer"} batch_size = 32#@param {type:"integer"} vector_dimension = 64 #@param {type:"integer"} hop=256 #hop size (window size = 4*hop) sr=44100 #sampling rate min_level_db=-100 #reference values to normalize data ref_level_db=20 LEARNING_RATE = learning_rate BATCH_SIZE = batch_size EPOCHS = num_epochs_to_train VECTOR_DIM=vector_dimension shape=128 #length of time axis of split specrograms spec_split=1 #@title Waveform to Spectrogram conversion ''' Decorsière, Rémi, Peter L. Søndergaard, Ewen N. MacDonald, and Torsten Dau. "Inversion of auditory spectrograms, traditional spectrograms, and other envelope representations." IEEE/ACM Transactions on Audio, Speech, and Language Processing 23, no. 1 (2014): 46-56.''' #ORIGINAL CODE FROM https://github.com/yoyololicon/spectrogram-inversion import torch import torch.nn as nn import torch.nn.functional as F from tqdm import tqdm from functools import partial import math import heapq from torchaudio.transforms import MelScale, Spectrogram torch.set_default_tensor_type('torch.cuda.FloatTensor') specobj = Spectrogram(n_fft=4*hop, win_length=4*hop, hop_length=hop, pad=0, power=2, normalized=False) specfunc = specobj.forward melobj = MelScale(n_mels=hop, sample_rate=sr, f_min=0.) melfunc = melobj.forward def melspecfunc(waveform): specgram = specfunc(waveform) mel_specgram = melfunc(specgram) return mel_specgram def spectral_convergence(input, target): return 20 * ((input - target).norm().log10() - target.norm().log10()) def GRAD(spec, transform_fn, samples=None, init_x0=None, maxiter=1000, tol=1e-6, verbose=1, evaiter=10, lr=0.002): spec = torch.Tensor(spec) samples = (spec.shape[-1]*hop)-hop if init_x0 is None: init_x0 = spec.new_empty((1,samples)).normal_(std=1e-6) x = nn.Parameter(init_x0) T = spec criterion = nn.L1Loss() optimizer = torch.optim.Adam([x], lr=lr) bar_dict = {} metric_func = spectral_convergence bar_dict['spectral_convergence'] = 0 metric = 'spectral_convergence' init_loss = None with tqdm(total=maxiter, disable=not verbose) as pbar: for i in range(maxiter): optimizer.zero_grad() V = transform_fn(x) loss = criterion(V, T) loss.backward() optimizer.step() lr = lr*0.9999 for param_group in optimizer.param_groups: param_group['lr'] = lr if i % evaiter == evaiter - 1: with torch.no_grad(): V = transform_fn(x) bar_dict[metric] = metric_func(V, spec).item() l2_loss = criterion(V, spec).item() pbar.set_postfix(**bar_dict, loss=l2_loss) pbar.update(evaiter) return x.detach().view(-1).cpu() def normalize(S): return np.clip((((S - min_level_db) / -min_level_db)*2.)-1., -1, 1) def denormalize(S): return (((np.clip(S, -1, 1)+1.)/2.) * -min_level_db) + min_level_db def prep(wv,hop=192): S = np.array(torch.squeeze(melspecfunc(torch.Tensor(wv).view(1,-1))).detach().cpu()) S = librosa.power_to_db(S)-ref_level_db return normalize(S) def deprep(S): S = denormalize(S)+ref_level_db S = librosa.db_to_power(S) wv = GRAD(np.expand_dims(S,0), melspecfunc, maxiter=2500, evaiter=10, tol=1e-8) return np.array(np.squeeze(wv)) #@title Helper functions #Generate spectrograms from waveform array def tospec(data): specs=np.empty(data.shape[0], dtype=object) for i in range(data.shape[0]): x = data[i] S=prep(x) S = np.array(S, dtype=np.float32) specs[i]=np.expand_dims(S, -1) print(specs.shape) return specs #Generate multiple spectrograms with a determined length from single wav file def tospeclong(path, length=4*44100): x, sr = librosa.load(path,sr=44100) x,_ = librosa.effects.trim(x) loudls = librosa.effects.split(x, top_db=50) xls = np.array([]) for interv in loudls: xls = np.concatenate((xls,x[interv[0]:interv[1]])) x = xls num = x.shape[0]//length specs=np.empty(num, dtype=object) for i in range(num-1): a = x[i*length:(i+1)*length] S = prep(a) S = np.array(S, dtype=np.float32) try: sh = S.shape specs[i]=S except AttributeError: print('spectrogram failed') print(specs.shape) return specs #Waveform array from path of folder containing wav files def audio_array(path): ls = glob(f'{path}/*.wav') adata = [] for i in range(len(ls)): x, sr = tf.audio.decode_wav(tf.io.read_file(ls[i]), 1) x = np.array(x, dtype=np.float32) adata.append(x) return np.array(adata) #Waveform array from path of folder containing wav files def single_audio_array(path): ls = glob(path) adata = [] for i in range(len(ls)): x, sr = tf.audio.decode_wav(tf.io.read_file(ls[i]), 1) x = np.array(x, dtype=np.float32) adata.append(x) return np.array(adata) #Concatenate spectrograms in array along the time axis def testass(a): but=False con = np.array([]) nim = a.shape[0] for i in range(nim): im = a[i] im = np.squeeze(im) if not but: con=im but=True else: con = np.concatenate((con,im), axis=1) return np.squeeze(con) #Split spectrograms in chunks with equal size def splitcut(data): ls = [] mini = 0 minifinal = spec_split*shape #max spectrogram length for i in range(data.shape[0]-1): if data[i].shape[1]<=data[i+1].shape[1]: mini = data[i].shape[1] else: mini = data[i+1].shape[1] if mini>=3*shape and mini<minifinal: minifinal = mini for i in range(data.shape[0]): x = data[i] if x.shape[1]>=3*shape: for n in range(x.shape[1]//minifinal): ls.append(x[:,n*minifinal:n*minifinal+minifinal,:]) ls.append(x[:,-minifinal:,:]) return np.array(ls) # Generates timestamp string of "day_month_year_hourMin" def get_time_stamp(): secondsSinceEpoch = time.time() timeObj = time.localtime(secondsSinceEpoch) x = ('%d_%d_%d_%d%d' % (timeObj.tm_mday, timeObj.tm_mon, timeObj.tm_year, timeObj.tm_hour, timeObj.tm_min)) return x ###Output _____no_output_____ ###Markdown Training ###Code #@title Import folder containing .wav files for training #Generating Mel-Spectrogram dataset (Uncomment where needed) #adata: source spectrograms audio_directory = "/path/to/your/audio/dataset" #@param {type:"string"} #AUDIO TO CONVERT awv = audio_array(audio_directory) #get waveform array from folder containing wav files aspec = tospec(awv) #get spectrogram array adata = splitcut(aspec) #split spectrogams to fixed print(np.shape(adata)) #@title Build VAE Neural Network #VAE import os import pickle from tensorflow.keras import Model from tensorflow.keras.layers import Input, Conv2D, ReLU, BatchNormalization, Flatten, Dense, Reshape, Conv2DTranspose, Activation, Lambda from tensorflow.keras import backend as K from tensorflow.keras.optimizers import Adam from tensorflow.keras.losses import MeanSquaredError from tensorflow.keras.callbacks import ModelCheckpoint import numpy as np import tensorflow as tf tf.compat.v1.disable_eager_execution() class VAE: """ VAE represents a Deep Convolutional autoencoder architecture with mirrored encoder and decoder components. """ def __init__(self, input_shape, #shape of the input data conv_filters, #convolutional network filters conv_kernels, #convNet kernel size conv_strides, #convNet strides latent_space_dim): self.input_shape = input_shape # [28, 28, 1], in this case is 28 x 28 pixels on 1 channel for greyscale self.conv_filters = conv_filters # is a list for each layer, i.e. [2, 4, 8] self.conv_kernels = conv_kernels # list of kernels per layer, [1,2,3] self.conv_strides = conv_strides # stride for each filter [1, 2, 2], note: 2 means you are downsampling the data in half self.latent_space_dim = latent_space_dim # how many neurons on bottleneck self.reconstruction_loss_weight = 1000000 self.encoder = None self.decoder = None self.model = None self._num_conv_layers = len(conv_filters) self._shape_before_bottleneck = None self._model_input = None self._build() def summary(self): self.encoder.summary() print("\n") self.decoder.summary() print("\n") self.model.summary() def _build(self): self._build_encoder() self._build_decoder() self._build_autoencoder() def compile(self, learning_rate=0.0001): optimizer = Adam(learning_rate=learning_rate) self.model.compile(optimizer=optimizer, loss=self._calculate_combined_loss, metrics=[self._calculate_reconstruction_loss, self._calculate_kl_loss]) def train(self, x_train, batch_size, num_epochs): # checkpoint = ModelCheckpoint("best_model.hdf5", monitor='loss', verbose=1, # save_best_only=True, mode='auto', period=1) self.model.fit(x_train, x_train, batch_size=batch_size, epochs=num_epochs, shuffle=True) #callbacks=[checkpoint]) def save(self, save_folder="."): self._create_folder_if_it_doesnt_exist(save_folder) self._save_parameters(save_folder) self._save_weights(save_folder) def load_weights(self, weights_path): self.model.load_weights(weights_path) def reconstruct(self, spec): latent_representations = self.encoder.predict(spec) reconstructed_spec = self.decoder.predict(latent_representations) return reconstructed_spec, latent_representations def encode(self, spec): latent_representation = self.encoder.predict(spec) return latent_representation def sample_from_latent_space(self, z): z_vector = self.decoder.predict(z) return z_vector @classmethod def load(cls, save_folder="."): parameters_path = os.path.join(save_folder, "parameters.pkl") with open(parameters_path, "rb") as f: parameters = pickle.load(f) autoencoder = VAE(*parameters) weights_path = os.path.join(save_folder, "weights.h5") autoencoder.load_weights(weights_path) return autoencoder def _calculate_combined_loss(self, y_target, y_predicted): reconstruction_loss = self._calculate_reconstruction_loss(y_target, y_predicted) kl_loss = self._calculate_kl_loss(y_target, y_predicted) combined_loss = self.reconstruction_loss_weight * reconstruction_loss + kl_loss return combined_loss def _calculate_reconstruction_loss(self, y_target, y_predicted): error = y_target - y_predicted reconstruction_loss = K.mean(K.square(error), axis=[1, 2, 3]) return reconstruction_loss def _calculate_kl_loss(self, y_target, y_predicted): kl_loss = -0.5 * K.sum(1 + self.log_variance - K.square(self.mu) - K.exp(self.log_variance), axis =1) return kl_loss def _create_folder_if_it_doesnt_exist(self, folder): if not os.path.exists(folder): os.makedirs(folder) def _save_parameters(self, save_folder): parameters = [ self.input_shape, self.conv_filters, self.conv_kernels, self.conv_strides, self.latent_space_dim ] save_path = os.path.join(save_folder, "parameters.pkl") with open(save_path, "wb") as f: pickle.dump(parameters, f) def _save_weights(self, save_folder): save_path = os.path.join(save_folder, "weights.h5") self.model.save_weights(save_path) #-----------AUTOENCODER----------# def _build_autoencoder(self): model_input = self._model_input model_output = self.decoder(self.encoder(model_input)) self.model = Model(model_input, model_output, name="autoencoder") #--------------DECODER------------# def _build_decoder(self): decoder_input = self._add_decoder_input() dense_layer = self._add_dense_layer(decoder_input) reshape_layer = self._add_reshape_layer(dense_layer) conv_transpose_layers = self._add_conv_transpose_layers(reshape_layer) decoder_output = self._add_decoder_output(conv_transpose_layers) self.decoder = Model(decoder_input, decoder_output, name="decoder") def _add_decoder_input(self): return Input(shape=self.latent_space_dim, name="decoder_input") def _add_dense_layer(self, decoder_input): num_neurons = np.prod(self._shape_before_bottleneck) # [ 1, 2, 4] -> 8 dense_layer = Dense(num_neurons, name="decoder_dense")(decoder_input) return dense_layer def _add_reshape_layer(self, dense_layer): return Reshape(self._shape_before_bottleneck)(dense_layer) def _add_conv_transpose_layers(self, x): """Add conv transpose blocks.""" # Loop through all the conv layers in reverse order and # stop at the first layer for layer_index in reversed(range(1, self._num_conv_layers)): x = self._add_conv_transpose_layer(layer_index, x) return x def _add_conv_transpose_layer(self, layer_index, x): layer_num = self._num_conv_layers - layer_index conv_transpose_layer = Conv2DTranspose( filters=self.conv_filters[layer_index], kernel_size = self.conv_kernels[layer_index], strides = self.conv_strides[layer_index], padding = "same", name=f"decoder_conv_transpose_layer_{layer_num}" ) x = conv_transpose_layer(x) x = ReLU(name=f"decoder_relu_{layer_num}")(x) x = BatchNormalization(name=f"decoder_bn_{layer_num}")(x) return x def _add_decoder_output(self, x): conv_transpose_layer = Conv2DTranspose( filters = 1, kernel_size = self.conv_kernels[0], strides = self.conv_strides[0], padding = "same", name=f"decoder_conv_transpose_layer_{self._num_conv_layers}" ) x = conv_transpose_layer(x) output_layer = Activation("sigmoid", name="sigmoid_output_layer")(x) return output_layer #----------------ENCODER-----------------# def _build_encoder(self): encoder_input = self._add_encoder_input() conv_layers = self._add_conv_layers(encoder_input) bottleneck = self._add_bottleneck(conv_layers) self._model_input = encoder_input self.encoder = Model(encoder_input, bottleneck, name="encoder") def _add_encoder_input(self): return Input(shape=self.input_shape, name="encoder_input") def _add_conv_layers(self, encoder_input): """Creates all convolutional blocks in encoder""" x = encoder_input for layer_index in range(self._num_conv_layers): x = self._add_conv_layer(layer_index, x) return x def _add_conv_layer(self, layer_index, x): """Adds a convolutional block to a graph of layers, consisting of Conv 2d + ReLu activation + batch normalization. """ layer_number = layer_index + 1 conv_layer = Conv2D( filters= self.conv_filters[layer_index], kernel_size = self.conv_kernels[layer_index], strides = self.conv_strides[layer_index], padding = "same", name = f"encoder_conv_layer_{layer_number}" ) x = conv_layer(x) x = ReLU(name=f"encoder_relu_{layer_number}")(x) x = BatchNormalization(name=f"encoder_bn_{layer_number}")(x) return x #-------------Bottleneck (Latent Space)-------------# def _add_bottleneck(self, x): """Flatten data and add bottleneck with Gaussian sampling (Dense layer)""" self._shape_before_bottleneck = K.int_shape(x)[1:] x = Flatten()(x) self.mu = Dense(self.latent_space_dim,name="mu")(x) self.log_variance = Dense(self.latent_space_dim, name="log_variance")(x) def sample_point_from_normal_distribution(args): mu, log_variance = args epsilon = K.random_normal(shape=K.shape(self.mu), mean=0., stddev=1.) sampled_point = mu + K.exp(log_variance / 2) * epsilon return sampled_point x = Lambda(sample_point_from_normal_distribution, name="encoder_output")([self.mu, self.log_variance]) return x print("VAE successfully built") #@title Training functions def train(x_train, learning_rate, batch_size, epochs): vae = VAE( input_shape = (hop, shape*spec_split, 1), conv_filters=(512, 256, 128, 64, 32), conv_kernels=(3, 3, 3, 3, 3), conv_strides=(2, 2, 2, 2, (2,1)), latent_space_dim = VECTOR_DIM ) vae.summary() vae.compile(learning_rate) vae.train(x_train, batch_size, epochs) return vae def train_tfdata(x_train, learning_rate, epochs=10): vae = VAE( input_shape = (hop, 3*shape, 1), conv_filters=(512, 256, 128, 64, 32), conv_kernels=(3, 3, 3, 3, 3), conv_strides=(2, 2, 2, 2, (2,1)), latent_space_dim = VECTOR_DIM ) vae.summary() vae.compile(learning_rate) vae.train(x_train, num_epochs=epochs) return vae def continue_training(checkpoint): vae = VAE.load(checkpoint) vae.summary() vae.compile(LEARNING_RATE) vae.train(adata,BATCH_SIZE,EPOCHS) return vae def load_model(checkpoint): vae = VAE.load(checkpoint) vae.summary() vae.compile(LEARNING_RATE) return vae #@title Start training from scratch or resume training training_run_name = "my_melspecvae_model" #@param {type:"string"} checkpoint_save_directory = "/path/to/your/checkpoints/" #@param {type:"string"} resume_training = False #@param {type:"boolean"} resume_training_checkpoint_path = "/path/to/your/checkpoints/" #@param {type:"string"} current_time = get_time_stamp() if not resume_training: vae = train(adata, LEARNING_RATE, BATCH_SIZE, EPOCHS) #vae = train_tfdata(dsa, LEARNING_RATE, EPOCHS) vae.save(f"{checkpoint_save_directory}{training_run_name}_{current_time}_h{hop}_w{shape}_z{VECTOR_DIM}") else: vae = continue_training(resume_training_checkpoint_path) vae.save(f"{checkpoint_save_directory}{training_run_name}_{current_time}_h{hop}_w{shape}_z{VECTOR_DIM}") ###Output _____no_output_____ ###Markdown Generation ###Code #@title Build VAE Neural Network #VAE import os import pickle from tensorflow.keras import Model from tensorflow.keras.layers import Input, Conv2D, ReLU, BatchNormalization, Flatten, Dense, Reshape, Conv2DTranspose, Activation, Lambda from tensorflow.keras import backend as K from tensorflow.keras.optimizers import Adam from tensorflow.keras.losses import MeanSquaredError from tensorflow.keras.callbacks import ModelCheckpoint import numpy as np import tensorflow as tf tf.compat.v1.disable_eager_execution() class VAE: """ VAE represents a Deep Convolutional autoencoder architecture with mirrored encoder and decoder components. """ def __init__(self, input_shape, #shape of the input data conv_filters, #convolutional network filters conv_kernels, #convNet kernel size conv_strides, #convNet strides latent_space_dim): self.input_shape = input_shape # [28, 28, 1], in this case is 28 x 28 pixels on 1 channel for greyscale self.conv_filters = conv_filters # is a list for each layer, i.e. [2, 4, 8] self.conv_kernels = conv_kernels # list of kernels per layer, [1,2,3] self.conv_strides = conv_strides # stride for each filter [1, 2, 2], note: 2 means you are downsampling the data in half self.latent_space_dim = latent_space_dim # how many neurons on bottleneck self.reconstruction_loss_weight = 1000000 self.encoder = None self.decoder = None self.model = None self._num_conv_layers = len(conv_filters) self._shape_before_bottleneck = None self._model_input = None self._build() def summary(self): self.encoder.summary() print("\n") self.decoder.summary() print("\n") self.model.summary() def _build(self): self._build_encoder() self._build_decoder() self._build_autoencoder() def compile(self, learning_rate=0.0001): optimizer = Adam(learning_rate=learning_rate) self.model.compile(optimizer=optimizer, loss=self._calculate_combined_loss, metrics=[self._calculate_reconstruction_loss, self._calculate_kl_loss]) def train(self, x_train, batch_size, num_epochs): # checkpoint = ModelCheckpoint("best_model.hdf5", monitor='loss', verbose=1, # save_best_only=True, mode='auto', period=1) self.model.fit(x_train, x_train, batch_size=batch_size, epochs=num_epochs, shuffle=True) #callbacks=[checkpoint]) def save(self, save_folder="."): self._create_folder_if_it_doesnt_exist(save_folder) self._save_parameters(save_folder) self._save_weights(save_folder) def load_weights(self, weights_path): self.model.load_weights(weights_path) def reconstruct(self, spec): latent_representations = self.encoder.predict(spec) reconstructed_spec = self.decoder.predict(latent_representations) return reconstructed_spec, latent_representations def encode(self, spec): latent_representation = self.encoder.predict(spec) return latent_representation def sample_from_latent_space(self, z): z_vector = self.decoder.predict(z) return z_vector @classmethod def load(cls, save_folder="."): parameters_path = os.path.join(save_folder, "parameters.pkl") with open(parameters_path, "rb") as f: parameters = pickle.load(f) autoencoder = VAE(*parameters) weights_path = os.path.join(save_folder, "weights.h5") autoencoder.load_weights(weights_path) return autoencoder def _calculate_combined_loss(self, y_target, y_predicted): reconstruction_loss = self._calculate_reconstruction_loss(y_target, y_predicted) kl_loss = self._calculate_kl_loss(y_target, y_predicted) combined_loss = self.reconstruction_loss_weight * reconstruction_loss + kl_loss return combined_loss def _calculate_reconstruction_loss(self, y_target, y_predicted): error = y_target - y_predicted reconstruction_loss = K.mean(K.square(error), axis=[1, 2, 3]) return reconstruction_loss def _calculate_kl_loss(self, y_target, y_predicted): kl_loss = -0.5 * K.sum(1 + self.log_variance - K.square(self.mu) - K.exp(self.log_variance), axis =1) return kl_loss def _create_folder_if_it_doesnt_exist(self, folder): if not os.path.exists(folder): os.makedirs(folder) def _save_parameters(self, save_folder): parameters = [ self.input_shape, self.conv_filters, self.conv_kernels, self.conv_strides, self.latent_space_dim ] save_path = os.path.join(save_folder, "parameters.pkl") with open(save_path, "wb") as f: pickle.dump(parameters, f) def _save_weights(self, save_folder): save_path = os.path.join(save_folder, "weights.h5") self.model.save_weights(save_path) #-----------AUTOENCODER----------# def _build_autoencoder(self): model_input = self._model_input model_output = self.decoder(self.encoder(model_input)) self.model = Model(model_input, model_output, name="autoencoder") #--------------DECODER------------# def _build_decoder(self): decoder_input = self._add_decoder_input() dense_layer = self._add_dense_layer(decoder_input) reshape_layer = self._add_reshape_layer(dense_layer) conv_transpose_layers = self._add_conv_transpose_layers(reshape_layer) decoder_output = self._add_decoder_output(conv_transpose_layers) self.decoder = Model(decoder_input, decoder_output, name="decoder") def _add_decoder_input(self): return Input(shape=self.latent_space_dim, name="decoder_input") def _add_dense_layer(self, decoder_input): num_neurons = np.prod(self._shape_before_bottleneck) # [ 1, 2, 4] -> 8 dense_layer = Dense(num_neurons, name="decoder_dense")(decoder_input) return dense_layer def _add_reshape_layer(self, dense_layer): return Reshape(self._shape_before_bottleneck)(dense_layer) def _add_conv_transpose_layers(self, x): """Add conv transpose blocks.""" # Loop through all the conv layers in reverse order and # stop at the first layer for layer_index in reversed(range(1, self._num_conv_layers)): x = self._add_conv_transpose_layer(layer_index, x) return x def _add_conv_transpose_layer(self, layer_index, x): layer_num = self._num_conv_layers - layer_index conv_transpose_layer = Conv2DTranspose( filters=self.conv_filters[layer_index], kernel_size = self.conv_kernels[layer_index], strides = self.conv_strides[layer_index], padding = "same", name=f"decoder_conv_transpose_layer_{layer_num}" ) x = conv_transpose_layer(x) x = ReLU(name=f"decoder_relu_{layer_num}")(x) x = BatchNormalization(name=f"decoder_bn_{layer_num}")(x) return x def _add_decoder_output(self, x): conv_transpose_layer = Conv2DTranspose( filters = 1, kernel_size = self.conv_kernels[0], strides = self.conv_strides[0], padding = "same", name=f"decoder_conv_transpose_layer_{self._num_conv_layers}" ) x = conv_transpose_layer(x) output_layer = Activation("sigmoid", name="sigmoid_output_layer")(x) return output_layer #----------------ENCODER-----------------# def _build_encoder(self): encoder_input = self._add_encoder_input() conv_layers = self._add_conv_layers(encoder_input) bottleneck = self._add_bottleneck(conv_layers) self._model_input = encoder_input self.encoder = Model(encoder_input, bottleneck, name="encoder") def _add_encoder_input(self): return Input(shape=self.input_shape, name="encoder_input") def _add_conv_layers(self, encoder_input): """Creates all convolutional blocks in encoder""" x = encoder_input for layer_index in range(self._num_conv_layers): x = self._add_conv_layer(layer_index, x) return x def _add_conv_layer(self, layer_index, x): """Adds a convolutional block to a graph of layers, consisting of Conv 2d + ReLu activation + batch normalization. """ layer_number = layer_index + 1 conv_layer = Conv2D( filters= self.conv_filters[layer_index], kernel_size = self.conv_kernels[layer_index], strides = self.conv_strides[layer_index], padding = "same", name = f"encoder_conv_layer_{layer_number}" ) x = conv_layer(x) x = ReLU(name=f"encoder_relu_{layer_number}")(x) x = BatchNormalization(name=f"encoder_bn_{layer_number}")(x) return x #-------------Bottleneck (Latent Space)-------------# def _add_bottleneck(self, x): """Flatten data and add bottleneck with Gaussian sampling (Dense layer)""" self._shape_before_bottleneck = K.int_shape(x)[1:] x = Flatten()(x) self.mu = Dense(self.latent_space_dim,name="mu")(x) self.log_variance = Dense(self.latent_space_dim, name="log_variance")(x) def sample_point_from_normal_distribution(args): mu, log_variance = args epsilon = K.random_normal(shape=K.shape(self.mu), mean=0., stddev=1.) sampled_point = mu + K.exp(log_variance / 2) * epsilon return sampled_point x = Lambda(sample_point_from_normal_distribution, name="encoder_output")([self.mu, self.log_variance]) return x print("VAE successfully built") #@title Load Checkpoint for Generating checkpoint_load_directory = "/path/to/your/checkpoint" #@param {type:"string"} #-------LOAD MODEL FOR GENERATING-------------# vae = VAE.load(checkpoint_load_directory) print("Loaded checkpoint") #@title Import synthesis utility functions #-----TESTING FUNCTIONS ----------- # def select_spec(spec, labels, num_spec=10): sample_spec_index = np.random.choice(range(len(spec)), num_spec) sample_spec = spec[sample_spec_index] sample_labels = labels[sample_spec_index] return sample_spec, sample_labels def plot_reconstructed_spec(spec, reconstructed_spec): fig = plt.figure(figsize=(15, 3)) num_spec = len(spec) for i, (image, reconstructed_image) in enumerate(zip(spec, reconstructed_spec)): image = image.squeeze() ax = fig.add_subplot(2, num_spec, i + 1) ax.axis("off") ax.imshow(image, cmap="gray_r") reconstructed_image = reconstructed_image.squeeze() ax = fig.add_subplot(2, num_spec, i + num_spec + 1) ax.axis("off") ax.imshow(reconstructed_image, cmap="gray_r") plt.show() def plot_spec_encoded_in_latent_space(latent_representations, sample_labels): plt.figure(figsize=(10, 10)) plt.scatter(latent_representations[:, 0], latent_representations[:, 1], cmap="rainbow", c=sample_labels, alpha=0.5, s=2) plt.colorbar() plt.show() #---------------NOISE GENERATOR FUNCTIONS ------------# def generate_random_z_vect(seed=1001,size_z=1,scale=1.0): np.random.seed(seed) x = np.random.uniform(low=(scale * -1.0), high=scale, size=(size_z,VECTOR_DIM)) return x def generate_z_vect_from_perlin_noise(seed=1001, size_z=1, scale=1.0): np.random.seed(seed) x = generate_perlin_noise_2d((size_z, VECTOR_DIM), (1,1)) x = x*scale return x def generate_z_vect_from_fractal_noise(seed=1001, size_z=1, scale=1.0,): np.random.seed(seed) x = generate_fractal_noise_2d((size_z, VECTOR_DIM), (1,1),) x = x*scale return x #-------SPECTROGRAM AND SOUND SYNTHESIS UTILITY FUNCTIONS -------- # #Assembling generated Spectrogram chunks into final Spectrogram def specass(a,spec): but=False con = np.array([]) nim = a.shape[0] for i in range(nim-1): im = a[i] im = np.squeeze(im) if not but: con=im but=True else: con = np.concatenate((con,im), axis=1) diff = spec.shape[1]-(nim*shape) a = np.squeeze(a) con = np.concatenate((con,a[-1,:,-diff:]), axis=1) return np.squeeze(con) #Splitting input spectrogram into different chunks to feed to the generator def chopspec(spec): dsa=[] for i in range(spec.shape[1]//shape): im = spec[:,i*shape:i*shape+shape] im = np.reshape(im, (im.shape[0],im.shape[1],1)) dsa.append(im) imlast = spec[:,-shape:] imlast = np.reshape(imlast, (imlast.shape[0],imlast.shape[1],1)) dsa.append(imlast) return np.array(dsa, dtype=np.float32) #Converting from source Spectrogram to target Spectrogram def towave_reconstruct(spec, spec1, name, path='../content/', show=False, save=False): specarr = chopspec(spec) specarr1 = chopspec(spec1) print(specarr.shape) a = specarr print('Generating...') ab = specarr1 print('Assembling and Converting...') a = specass(a,spec) ab = specass(ab,spec1) awv = deprep(a) abwv = deprep(ab) if save: print('Saving...') pathfin = f'{path}/{name}' sf.write(f'{pathfin}.wav', awv, sr) print('Saved WAV!') IPython.display.display(IPython.display.Audio(np.squeeze(abwv), rate=sr)) IPython.display.display(IPython.display.Audio(np.squeeze(awv), rate=sr)) if show: fig, axs = plt.subplots(ncols=2) axs[0].imshow(np.flip(a, -2), cmap=None) axs[0].axis('off') axs[0].set_title('Reconstructed') axs[1].imshow(np.flip(ab, -2), cmap=None) axs[1].axis('off') axs[1].set_title('Input') plt.show() return abwv #Converting from Z vector generated spectrogram to waveform def towave_from_z(spec, name, path='../content/', show=False, save=False): specarr = chopspec(spec) print(specarr.shape) a = specarr print('Generating...') print('Assembling and Converting...') a = specass(a,spec) awv = deprep(a) if save: print('Saving...') pathfin = f'{path}/{name}' sf.write(f'{pathfin}.wav', awv, sr) print('Saved WAV!') IPython.display.display(IPython.display.Audio(np.squeeze(awv), rate=sr)) if show: fig, axs = plt.subplots(ncols=1) axs.imshow(np.flip(a, -2), cmap=None) axs.axis('off') axs.set_title('Decoder Synthesis') plt.show() return awv #@title Compare resynthesized MelSpec with ground truth MelSpec. Note: Only works if you imported an audio dataset num_spec_to_resynthesize = 5 #@param {type:"integer"} num_sample_spec_to_show = num_spec_to_resynthesize sample_spec, _ = select_spec(adata, adata, num_sample_spec_to_show) reconstructed_spec, _ = vae.reconstruct(sample_spec) plot_reconstructed_spec(sample_spec, reconstructed_spec) reconst = num_sample_spec_to_show for i in range(reconst): y = towave_reconstruct(reconstructed_spec[i],sample_spec[i],name='reconstructions',show=True, save=False) #@title Generate one-shot samples from latent space with random or manual seed num_samples_to_generate = 10#@param {type:"integer"} use_seed = False #@param {type:"boolean"} seed = 0 #@param {type:"slider", min:0, max:4294967295, step:1} scale_z_vectors = 1.5 #@param {type:"slider", min:-5.0, max:5.0, step:0.1} save_audio = False #@param {type:"boolean"} audio_name = "one_shot" #@param {type:"string"} audio_save_directory = "/content/" #@param {type:"string"} y = np.random.randint(0, 2**32-1) # generated random int to pass and convert into vector i=0 while i < num_samples_to_generate: if not use_seed: z = generate_random_z_vect(y, num_samples_to_generate,scale=scale_z_vectors) else: z = generate_random_z_vect(seed, num_samples_to_generate,scale=scale_z_vectors) z_sample = np.array(vae.sample_from_latent_space(z)) towave_from_z(z_sample[i], name=f'{audio_name}_{i}',path=audio_save_directory,show=True, save=save_audio) i+=1 if not use_seed: print("Generated from seed:", y) else: print("Generated from seed:", seed) #@title Generate arbitrary long audio from latent space with random or custom seed using uniform, Perlin or fractal noise num_seeds_to_generate = 32#@param {type:"integer"} noise_type = "uniform" #@param ["uniform", "perlin", "fractal"] use_seed = False #@param {type:"boolean"} seed = 0 #@param {type:"slider", min:0, max:4294967295, step:1} scale_z_vectors = 1.5 #@param {type:"slider", min:-5.0, max:5.0, step:0.1} save_audio = False #@param {type:"boolean"} audio_name = "VAE_synthesis2" #@param {type:"string"} audio_save_directory = "/content" #@param {type:"string"} y = np.random.randint(0, 2**32-1) # generated random int to pass and convert into vector if not use_seed: if noise_type == "uniform": z = generate_random_z_vect(y, num_seeds_to_generate,scale_z_vectors) # vectors to input into latent space if noise_type == "perlin": z = generate_z_vect_from_perlin_noise(y, num_seeds_to_generate,scale_z_vectors) # vectors to input into latent space if noise_type == "fractal": z = generate_z_vect_from_fractal_noise(y, num_seeds_to_generate,scale_z_vectors) # vectors to input into latent space if use_seed: if noise_type == "uniform": z = generate_random_z_vect(seed, num_seeds_to_generate,scale_z_vectors) # vectors to input into latent space if noise_type == "perlin": z = generate_z_vect_from_perlin_noise(seed, num_seeds_to_generate,scale_z_vectors) # vectors to input into latent space if noise_type == "fractal": z = generate_z_vect_from_fractal_noise(seed, num_seeds_to_generate,scale_z_vectors) # vectors to input into latent space z_sample = np.array(vae.sample_from_latent_space(z)) assembled_spec = testass(z_sample) towave_from_z(assembled_spec,audio_name,audio_save_directory,show=True,save=save_audio) if not use_seed: print("Generated from seed:", y) else: print("Generated from seed:", seed) #@title Interpolate between two seeds for n-amount of steps use_seed = False #@param {type:"boolean"} seed_a = 0 #@param {type:"slider", min:0, max:4294967295, step:1} seed_b = 4294967295 #@param {type:"slider", min:0, max:4294967295, step:1} num_interpolation_steps = 8#@param {type:"integer"} scale_z_vectors = 1.5 #@param {type:"slider", min:-5.0, max:5.0, step:0.1} scale_interpolation_ratio = 1 #@param {type:"slider", min:-5.0, max:5.0, step:0.1} save_audio = False #@param {type:"boolean"} audio_name = "random_seeds_interpolation" #@param {type:"string"} audio_save_directory = "/content/" #@param {type:"string"} # generate points in latent space as input for the generator def generate_latent_points(latent_dim, n_samples, n_classes=10): # generate points in the latent space x_input = randn(latent_dim * n_samples) # reshape into a batch of inputs for the network z_input = x_input.reshape(n_samples, latent_dim) return z_input # uniform interpolation between two points in latent space def interpolate_points(p1, p2,scale, n_steps=10): # interpolate ratios between the points ratios = linspace(-scale, scale, num=n_steps) # linear interpolate vectors vectors = list() for ratio in ratios: v = (1.0 - ratio) * p1 + ratio * p2 vectors.append(v) return asarray(vectors) y = np.random.randint(0, 2**32-1) if not use_seed: pts = generate_random_z_vect(y,10,scale_z_vectors) interpolated = interpolate_points(pts[0], pts[1], scale_interpolation_ratio, num_interpolation_steps) # interpolate points in latent space else: pts_a = generate_random_z_vect(seed_a,10,scale_z_vectors) pts_b = generate_random_z_vect(seed_b,10,scale_z_vectors) interpolated = interpolate_points(pts_a[0], pts_b[0], scale_interpolation_ratio, num_interpolation_steps) # interpolate points in latent space interp = np.array(vae.sample_from_latent_space(interpolated)) assembled_spec = testass(interp) towave_from_z(assembled_spec,audio_name,audio_save_directory,show=True, save=save_audio) if not use_seed: print("Generated from seed:", y) else: print("Generated from seed:", seed) ###Output _____no_output_____ ###Markdown EXPERIMENTAL: Timbre Transfer Note: You need to Restart Runtime (Runtime > Restart Runtime) per each Timbre Transfer ###Code #@title Import wav input_audio = "your_audio_to_timbre_transfer.wav" #@param {type:"string"} audio_in = single_audio_array(input_audio) aspec = tospec(audio_in) out_data = splitcut(aspec) #@title Build VAE Neural Network #VAE import os import pickle from tensorflow.keras import Model from tensorflow.keras.layers import Input, Conv2D, ReLU, BatchNormalization, Flatten, Dense, Reshape, Conv2DTranspose, Activation, Lambda from tensorflow.keras import backend as K from tensorflow.keras.optimizers import Adam from tensorflow.keras.losses import MeanSquaredError from tensorflow.keras.callbacks import ModelCheckpoint import numpy as np import tensorflow as tf tf.compat.v1.disable_eager_execution() class VAE: """ VAE represents a Deep Convolutional autoencoder architecture with mirrored encoder and decoder components. """ def __init__(self, input_shape, #shape of the input data conv_filters, #convolutional network filters conv_kernels, #convNet kernel size conv_strides, #convNet strides latent_space_dim): self.input_shape = input_shape # [28, 28, 1], in this case is 28 x 28 pixels on 1 channel for greyscale self.conv_filters = conv_filters # is a list for each layer, i.e. [2, 4, 8] self.conv_kernels = conv_kernels # list of kernels per layer, [1,2,3] self.conv_strides = conv_strides # stride for each filter [1, 2, 2], note: 2 means you are downsampling the data in half self.latent_space_dim = latent_space_dim # how many neurons on bottleneck self.reconstruction_loss_weight = 1000000 self.encoder = None self.decoder = None self.model = None self._num_conv_layers = len(conv_filters) self._shape_before_bottleneck = None self._model_input = None self._build() def summary(self): self.encoder.summary() print("\n") self.decoder.summary() print("\n") self.model.summary() def _build(self): self._build_encoder() self._build_decoder() self._build_autoencoder() def compile(self, learning_rate=0.0001): optimizer = Adam(learning_rate=learning_rate) self.model.compile(optimizer=optimizer, loss=self._calculate_combined_loss, metrics=[self._calculate_reconstruction_loss, self._calculate_kl_loss]) def train(self, x_train, batch_size, num_epochs): # checkpoint = ModelCheckpoint("best_model.hdf5", monitor='loss', verbose=1, # save_best_only=True, mode='auto', period=1) self.model.fit(x_train, x_train, batch_size=batch_size, epochs=num_epochs, shuffle=True) #callbacks=[checkpoint]) def save(self, save_folder="."): self._create_folder_if_it_doesnt_exist(save_folder) self._save_parameters(save_folder) self._save_weights(save_folder) def load_weights(self, weights_path): self.model.load_weights(weights_path) def reconstruct(self, spec): latent_representations = self.encoder.predict(spec) reconstructed_spec = self.decoder.predict(latent_representations) return reconstructed_spec, latent_representations def encode(self, spec): latent_representation = self.encoder.predict(spec) return latent_representation def sample_from_latent_space(self, z): z_vector = self.decoder.predict(z) return z_vector @classmethod def load(cls, save_folder="."): parameters_path = os.path.join(save_folder, "parameters.pkl") with open(parameters_path, "rb") as f: parameters = pickle.load(f) autoencoder = VAE(*parameters) weights_path = os.path.join(save_folder, "weights.h5") autoencoder.load_weights(weights_path) return autoencoder def _calculate_combined_loss(self, y_target, y_predicted): reconstruction_loss = self._calculate_reconstruction_loss(y_target, y_predicted) kl_loss = self._calculate_kl_loss(y_target, y_predicted) combined_loss = self.reconstruction_loss_weight * reconstruction_loss + kl_loss return combined_loss def _calculate_reconstruction_loss(self, y_target, y_predicted): error = y_target - y_predicted reconstruction_loss = K.mean(K.square(error), axis=[1, 2, 3]) return reconstruction_loss def _calculate_kl_loss(self, y_target, y_predicted): kl_loss = -0.5 * K.sum(1 + self.log_variance - K.square(self.mu) - K.exp(self.log_variance), axis =1) return kl_loss def _create_folder_if_it_doesnt_exist(self, folder): if not os.path.exists(folder): os.makedirs(folder) def _save_parameters(self, save_folder): parameters = [ self.input_shape, self.conv_filters, self.conv_kernels, self.conv_strides, self.latent_space_dim ] save_path = os.path.join(save_folder, "parameters.pkl") with open(save_path, "wb") as f: pickle.dump(parameters, f) def _save_weights(self, save_folder): save_path = os.path.join(save_folder, "weights.h5") self.model.save_weights(save_path) #-----------AUTOENCODER----------# def _build_autoencoder(self): model_input = self._model_input model_output = self.decoder(self.encoder(model_input)) self.model = Model(model_input, model_output, name="autoencoder") #--------------DECODER------------# def _build_decoder(self): decoder_input = self._add_decoder_input() dense_layer = self._add_dense_layer(decoder_input) reshape_layer = self._add_reshape_layer(dense_layer) conv_transpose_layers = self._add_conv_transpose_layers(reshape_layer) decoder_output = self._add_decoder_output(conv_transpose_layers) self.decoder = Model(decoder_input, decoder_output, name="decoder") def _add_decoder_input(self): return Input(shape=self.latent_space_dim, name="decoder_input") def _add_dense_layer(self, decoder_input): num_neurons = np.prod(self._shape_before_bottleneck) # [ 1, 2, 4] -> 8 dense_layer = Dense(num_neurons, name="decoder_dense")(decoder_input) return dense_layer def _add_reshape_layer(self, dense_layer): return Reshape(self._shape_before_bottleneck)(dense_layer) def _add_conv_transpose_layers(self, x): """Add conv transpose blocks.""" # Loop through all the conv layers in reverse order and # stop at the first layer for layer_index in reversed(range(1, self._num_conv_layers)): x = self._add_conv_transpose_layer(layer_index, x) return x def _add_conv_transpose_layer(self, layer_index, x): layer_num = self._num_conv_layers - layer_index conv_transpose_layer = Conv2DTranspose( filters=self.conv_filters[layer_index], kernel_size = self.conv_kernels[layer_index], strides = self.conv_strides[layer_index], padding = "same", name=f"decoder_conv_transpose_layer_{layer_num}" ) x = conv_transpose_layer(x) x = ReLU(name=f"decoder_relu_{layer_num}")(x) x = BatchNormalization(name=f"decoder_bn_{layer_num}")(x) return x def _add_decoder_output(self, x): conv_transpose_layer = Conv2DTranspose( filters = 1, kernel_size = self.conv_kernels[0], strides = self.conv_strides[0], padding = "same", name=f"decoder_conv_transpose_layer_{self._num_conv_layers}" ) x = conv_transpose_layer(x) output_layer = Activation("sigmoid", name="sigmoid_output_layer")(x) return output_layer #----------------ENCODER-----------------# def _build_encoder(self): encoder_input = self._add_encoder_input() conv_layers = self._add_conv_layers(encoder_input) bottleneck = self._add_bottleneck(conv_layers) self._model_input = encoder_input self.encoder = Model(encoder_input, bottleneck, name="encoder") def _add_encoder_input(self): return Input(shape=self.input_shape, name="encoder_input") def _add_conv_layers(self, encoder_input): """Creates all convolutional blocks in encoder""" x = encoder_input for layer_index in range(self._num_conv_layers): x = self._add_conv_layer(layer_index, x) return x def _add_conv_layer(self, layer_index, x): """Adds a convolutional block to a graph of layers, consisting of Conv 2d + ReLu activation + batch normalization. """ layer_number = layer_index + 1 conv_layer = Conv2D( filters= self.conv_filters[layer_index], kernel_size = self.conv_kernels[layer_index], strides = self.conv_strides[layer_index], padding = "same", name = f"encoder_conv_layer_{layer_number}" ) x = conv_layer(x) x = ReLU(name=f"encoder_relu_{layer_number}")(x) x = BatchNormalization(name=f"encoder_bn_{layer_number}")(x) return x #-------------Bottleneck (Latent Space)-------------# def _add_bottleneck(self, x): """Flatten data and add bottleneck with Gaussian sampling (Dense layer)""" self._shape_before_bottleneck = K.int_shape(x)[1:] x = Flatten()(x) self.mu = Dense(self.latent_space_dim,name="mu")(x) self.log_variance = Dense(self.latent_space_dim, name="log_variance")(x) def sample_point_from_normal_distribution(args): mu, log_variance = args epsilon = K.random_normal(shape=K.shape(self.mu), mean=0., stddev=1.) sampled_point = mu + K.exp(log_variance / 2) * epsilon return sampled_point x = Lambda(sample_point_from_normal_distribution, name="encoder_output")([self.mu, self.log_variance]) return x print("VAE successfully built") #@title Load Checkpoint for Generating checkpoint_load_directory = "/path/to/your/checkpoint" #@param {type:"string"} #-------LOAD MODEL FOR GENERATING-------------# vae = VAE.load(checkpoint_load_directory) print("Loaded checkpoint") #@title Import synthesis utility functions #-----TESTING FUNCTIONS ----------- # def select_spec(spec, labels, num_spec=10): sample_spec_index = np.random.choice(range(len(spec)), num_spec) sample_spec = spec[sample_spec_index] sample_labels = labels[sample_spec_index] return sample_spec, sample_labels def plot_reconstructed_spec(spec, reconstructed_spec): fig = plt.figure(figsize=(15, 3)) num_spec = len(spec) for i, (image, reconstructed_image) in enumerate(zip(spec, reconstructed_spec)): image = image.squeeze() ax = fig.add_subplot(2, num_spec, i + 1) ax.axis("off") ax.imshow(image, cmap="gray_r") reconstructed_image = reconstructed_image.squeeze() ax = fig.add_subplot(2, num_spec, i + num_spec + 1) ax.axis("off") ax.imshow(reconstructed_image, cmap="gray_r") plt.show() def plot_spec_encoded_in_latent_space(latent_representations, sample_labels): plt.figure(figsize=(10, 10)) plt.scatter(latent_representations[:, 0], latent_representations[:, 1], cmap="rainbow", c=sample_labels, alpha=0.5, s=2) plt.colorbar() plt.show() #---------------NOISE GENERATOR FUNCTIONS ------------# def generate_random_z_vect(seed=1001,size_z=1,scale=1.0): np.random.seed(seed) x = np.random.uniform(low=(scale * -1.0), high=scale, size=(size_z,VECTOR_DIM)) return x def generate_z_vect_from_perlin_noise(seed=1001, size_z=1, scale=1.0): np.random.seed(seed) x = generate_perlin_noise_2d((size_z, VECTOR_DIM), (1,1)) x = x*scale return x def generate_z_vect_from_fractal_noise(seed=1001, size_z=1, scale=1.0,): np.random.seed(seed) x = generate_fractal_noise_2d((size_z, VECTOR_DIM), (1,1),) x = x*scale return x #-------SPECTROGRAM AND SOUND SYNTHESIS UTILITY FUNCTIONS -------- # #Assembling generated Spectrogram chunks into final Spectrogram def specass(a,spec): but=False con = np.array([]) nim = a.shape[0] for i in range(nim-1): im = a[i] im = np.squeeze(im) if not but: con=im but=True else: con = np.concatenate((con,im), axis=1) diff = spec.shape[1]-(nim*shape) a = np.squeeze(a) con = np.concatenate((con,a[-1,:,-diff:]), axis=1) return np.squeeze(con) #Splitting input spectrogram into different chunks to feed to the generator def chopspec(spec): dsa=[] for i in range(spec.shape[1]//shape): im = spec[:,i*shape:i*shape+shape] im = np.reshape(im, (im.shape[0],im.shape[1],1)) dsa.append(im) imlast = spec[:,-shape:] imlast = np.reshape(imlast, (imlast.shape[0],imlast.shape[1],1)) dsa.append(imlast) return np.array(dsa, dtype=np.float32) #Converting from source Spectrogram to target Spectrogram def towave_reconstruct(spec, spec1, name, path='../content/', show=False, save=False): specarr = chopspec(spec) specarr1 = chopspec(spec1) print(specarr.shape) a = specarr print('Generating...') ab = specarr1 print('Assembling and Converting...') a = specass(a,spec) ab = specass(ab,spec1) awv = deprep(a) abwv = deprep(ab) if save: print('Saving...') pathfin = f'{path}/{name}' sf.write(f'{pathfin}.wav', awv, sr) print('Saved WAV!') IPython.display.display(IPython.display.Audio(np.squeeze(abwv), rate=sr)) IPython.display.display(IPython.display.Audio(np.squeeze(awv), rate=sr)) if show: fig, axs = plt.subplots(ncols=2) axs[0].imshow(np.flip(a, -2), cmap=None) axs[0].axis('off') axs[0].set_title('Reconstructed') axs[1].imshow(np.flip(ab, -2), cmap=None) axs[1].axis('off') axs[1].set_title('Input') plt.show() return abwv #Converting from Z vector generated spectrogram to waveform def towave_from_z(spec, name, path='../content/', show=False, save=False): specarr = chopspec(spec) print(specarr.shape) a = specarr print('Generating...') print('Assembling and Converting...') a = specass(a,spec) awv = deprep(a) if save: print('Saving...') pathfin = f'{path}/{name}' sf.write(f'{pathfin}.wav', awv, sr) print('Saved WAV!') IPython.display.display(IPython.display.Audio(np.squeeze(awv), rate=sr)) if show: fig, axs = plt.subplots(ncols=1) axs.imshow(np.flip(a, -2), cmap=None) axs.axis('off') axs.set_title('Decoder Synthesis') plt.show() return awv #@title Perform Timbre Transfer save_audio = True #@param {type:"boolean"} audio_name = "timbre_transfered_audio" #@param {type:"string"} audio_save_directory = "/content/" #@param {type:"string"} encoded_spec = vae.encode(out_data) print(np.shape(encoded_spec)) z_sample = np.array(vae.sample_from_latent_space(encoded_spec)) assembled_spec = testass(z_sample) towave_from_z(assembled_spec, name=audio_name,path=audio_save_directory,show=True, save=save_audio) ###Output _____no_output_____
notebook/core/module/Def.ipynb
###Markdown 함수 ###Code def movie1(name): print('영화명:',name) #리턴타입은 명시하지 않는다. movie1('킹덤') movie1('청춘 기록') movie1('테넷') def movie2(name, genre): # 인수 여러개 사용 가능 print('영화명: ' + name) print('장 르: ' + genre) #인수 부족하면 에러 movie2('1987','역사') movie2('종이의 집','범죄 드라마') def movie3(name, genre, score=5.0): # 기본값 사용,파이썬의 기능 print('영화명: ' + name) print('장 르: ' + genre) print('평 점: ' + str(score)) movie3('기생충','드라마') #기본값이 전달된 상태 movie3('검은 사제들','퇴마',9.5) movie3('컨저링2') # X 기본값 지정 안된것은 에러 def movie4(time, name, genre): print('영화명: ' + name) print('시 간: ' + str(time)) #문자열과 숫자는 +사용 불가하므로 형변환 필요 print('장 르: ' + genre) movie4(120,'안시성','역사') movie4(name='업사이드',time=100,genre='휴먼') #파라미터 순서 변경 가능 def movie5(*actors): # tuple, 가변인자 처리, 전달되는 변수의 수나,타입등 제한없다 print(type(actors)) print(actors) movie5('안시성', '조인성', '김태리') movie5('맘마미아', '피어스브로스넌', '메릴스트립', '아만다 사이프리드') movie5('인터스텔라', '메튜 맥커너희', '엔 헤서웨이', 2014) def movie6(movie, *actors): # 고정과 가변인자 병합 처리 print(type(actors)) print(movie) print(actors) # Dictionary # 첫번째 인자가 고정으로 들어간 후(movie) 나머지 인자가 가변인자(actors)로 들어감 movie6('안시성', '조인성', '김태리') movie6('맘마미아', '피어스브로스넌', '메릴스트립', '아만다 사이프리드') movie6('인터스텔라', '메튜 맥커너희', '엔 헤서웨이', 2014) def movie7(movie, **actors): # 고정과 가변인자 병합 처리, Dictionary print(movie) print(type(actors)) print(actors) # Dictionary # movie7('안시성', '조인성', '김태리') # X movie7('인터스텔라', actor1='맥커너희', actor2='앤 헤서웨이', actor3='마이클 케인') # Dictionary def season(month): #기본적으로 파이썬은 리턴타잆이 없어도 된다. season='' if month == 1: season='January' elif month == 2: season='February' elif month == 3: season="March" else: season="Only input 1 ~ 3" return season #그러나 리턴타입을 쓰는 경우도 존재한다 test=season(1) #리턴타입을 명시했으므로, 사용 가능, 없으면 None 출력 print(test) str1='전역변수' def fun1(): str1='지역 변수' print(str1) fun1() #전역 변수와 지역변수 중 지역 변수가 우선순위가 높음 print(str1) #함수 밖에서는 전역변수가 우선순위가 높음 str1='전역변수' def fun2(): global str1 # 전역변수를 사용 str1 = '지역 변수값 적용' print(str1) fun2() print(str1) #global 변수 선언시,지역 변수가 전역변수로 대체. def fun3(): print(str1) # 해당하는 지역변수가 없으면 전역 변수 자동 인식 fun3() def maxno(x,y): #x,y:지역 변수 if x > y: return x else: return y su= maxno(100,200) print(type(su)) print(su) def reverse(x,y,z): #리턴값 다수 선언 가능 return z,y,x ret=reverse(10,20,30) print(ret) r1,r2,r3=reverse(10,20,30) print(r1,r2,r3) r1,r2,r3=reverse("A","B","C") print(r1,r2,r3) # swap 알고리즘,중요한 개념★★ r1,r2=10,20 print(r1,r2) r1,r2=r2,r1 #값 교채하기 print(r1,r2) #자바 방식 r1 = 10 r2 = 20 temp=r1 r1=r2 r2=temp print(r1,r2) def data(): return(1,2),('A','B') # a,b,c,d=data() #X,사용불가 ret=data() print(type(ret)) print(ret) (a,b),(c,d)=data() #개발자가 타입을 알고 있어야 한다. print(a,b,c,d) def data(): return[1,2,3] #list타입 리턴 ret =data() print(type(ret)) print(ret) def data(): return[1,2,3],[4,5,6] ret =data() print(type(ret)) print(ret) #list가 결합된tuple print(type(ret[0])) print(ret[0]) print(type(ret[1])) print(ret[1]) ret1,ret2=data() print(ret1) print(ret2) #파이썬 파일을 모듈이라고 부름 ###Output _____no_output_____
ppkgl/doitkaggle.ipynb
###Markdown ###Code from google.colab import drive drive.mount('/content/drive') # Ignore the warnings import warnings warnings.filterwarnings('ignore') # System related and data input controls import os # Data manipulation, visualization and useful functions import pandas as pd import numpy as np from itertools import product # iterative combinations from tqdm import tqdm import matplotlib.pyplot as plt import seaborn as sns # Modeling algorithms # General(Statistics/Econometrics) from sklearn import preprocessing import statsmodels.api as sm import statsmodels.tsa.api as smt import statsmodels.formula.api as smf from statsmodels.stats.outliers_influence import variance_inflation_factor from scipy import stats # Regression from sklearn.linear_model import LinearRegression, Ridge, Lasso, ElasticNet from sklearn.kernel_ridge import KernelRidge from sklearn.neighbors import KNeighborsRegressor from sklearn.svm import SVR from sklearn.tree import DecisionTreeRegressor from sklearn.ensemble import RandomForestRegressor, BaggingRegressor, GradientBoostingRegressor, AdaBoostRegressor from xgboost import XGBRegressor from lightgbm import LGBMRegressor # Classification from sklearn.linear_model import LogisticRegression from sklearn.neighbors import KNeighborsClassifier from sklearn.svm import LinearSVC, SVC from sklearn.tree import DecisionTreeClassifier from sklearn.naive_bayes import GaussianNB from sklearn.ensemble import RandomForestClassifier, GradientBoostingClassifier # Model selection from sklearn.model_selection import train_test_split,cross_validate from sklearn.model_selection import KFold from sklearn.model_selection import GridSearchCV # Evaluation metrics # for regression from sklearn.metrics import mean_squared_log_error, mean_squared_error, r2_score, mean_absolute_error # for classification from sklearn.metrics import accuracy_score, precision_score, recall_score, f1_score import pandas as pd from keras.models import Sequential, Model, load_model from keras.layers import Input, Dense, Activation, Flatten, Dropout from keras.layers import SimpleRNN, LSTM, GRU item_categories = pd.read_csv('/content/drive/MyDrive/data/item_categories.csv') items = pd.read_csv('/content/drive/MyDrive/data/items.csv') sales_train = pd.read_csv('/content/drive/MyDrive/data/sales_train.csv') #학습데이터 sample_submission = pd.read_csv('/content/drive/MyDrive/data/sample_submission.csv') shops = pd.read_csv('/content/drive/MyDrive/data/shops.csv') test = pd.read_csv('/content/drive/MyDrive/data/test.csv')#테스트 데이터 raw_data = sales_train #1.train에 중복값 제거 (raw_data 모든 컬럼값이 동일한 value를 삭제) # drop_duplicates subset= ['date','date_block_num','shop_id','item_id','item_cnt_day'] print(raw_data.duplicated(subset=subset).value_counts()) raw_data.drop_duplicates(subset=subset, inplace=True) # 2935849 - 2935825 = 24개의 중복데이터 삭제 #2.test에 있는 세일즈수만 예측하면되기때문에 train에서 test에 없는 상품 제거 #test_shops = test.shop_id.unique() #test_items = test.item_id.unique() #raw_data = raw_data[raw_data.shop_id.isin(test_shops)] #raw_data = raw_data[raw_data.item_id.isin(test_items)] #raw_data.shape # 기존2935825에서1224429으로 절반 이상 데이터 드랍 # item(22169개의 아이템)과 item_cate(83개의 카테고리) 의 데이터 그룹 생성 # item_categories_name 안에 또 큰 범주의 카테고리가 있다. # pickup first-category name items_g = item_categories['item_category_name'].apply(lambda x: str(x).split(' ')[0]) item_categories['item_group'] = pd.Categorical(items_g).codes item_c = pd.merge(item_categories, item_categories.loc[:,['item_category_id','item_group']], on=['item_category_id'], how='left') item_c.head() item_c = item_c.drop('item_group_y',axis=1) item_c = item_c.drop('item_category_name',axis=1) item_c.columns = ['item_category_id','item_group'] item_c # 상점 이름의 따른 분류 # 상점도 이름안에 큰 범주가 있다. city로 변경 city = shops.shop_name.apply(lambda x: str.replace(x, '!', '')).apply(lambda x: x.split(' ')[0]) shops['city'] = pd.Categorical(city).codes shops.head() shops = shops.drop('shop_name', axis=1) shops.head() items.info() items.head() itemsinfo = pd.merge(items, item_c) itemsinfo len(raw_data) df = pd.merge(raw_data,itemsinfo) len(df) df = pd.merge(df,shops) df.head() df = df.drop('item_name',axis=1) #df = df.drop('item_group',axis=1) df = df.drop('item_price',axis=1) df.info() df.isna().sum() sns.heatmap(df.corr(), annot=True, cmap='YlOrRd') #df.columns #sns.jointplot(x='item_price', y='item_cnt_day', data=df) #sns.jointplot(x='item_category_id', y='item_cnt_day', data=df) #sns.jointplot(x='item_id', y='item_cnt_day', data=df) #sns.jointplot(x='shop_id', y='item_cnt_day', data=df) #df.columns #sns.boxplot(x='shop_name', y='item_cnt_day', data=df) #sns.boxplot(x='date', y='item_cnt_day', data=df) df.columns #df2 = df.drop(['item_price','date_block_num'],axis=1) df #print(df['item_price'].quantile(0.95)) #print(df['item_price'].quantile(0.005)) #df2 = df2[(df2['item_price'] < df2['item_price'].quantile(0.95)) & (df2['item_price'] > df2['item_price'].quantile(0.005))] #df2['item_price'].hist() df.hist() df2 = df df2= df2[(df2['item_cnt_day'] < df2['item_cnt_day'].quantile(0.95))] df2['item_cnt_day'].hist() df2['date'] = pd.to_datetime(df2['date']) df3 = df2.set_index(['date']) df3 = df3.sort_index() df3 = df3.reset_index() df3.index.name = 'ID' df3.head() df3 = df3.rename_axis('ID').reset_index() df3.head() df3 = df3.set_index('date') df3.head() dfformerge = df3[['item_id','item_category_id','item_group']] dfgormerge2 = df3[['shop_id','city']] dfformerge dfgormerge2 df4 = df3.groupby(by=['shop_id','item_id'])['item_cnt_day'].resample('M').sum() #df7 = df3.groupby(by=['shop_id','item_id'])['item_price'].resample('M').sum() #dfprice = pd.DataFrame(df7) #dfprice = dfprice.reset_index() df4 df5 = pd.DataFrame(df4) df5 = df5.reset_index() df5.head() df5.info() items = items.drop('item_name',axis=1) df6 = pd.merge(df5, items, how='left', left_on='item_id',right_on='item_id') df6 dform_du = dfformerge.drop_duplicates(subset = 'item_id') dform_du dform_du2 = dfgormerge2.drop_duplicates(subset = 'shop_id') dform_du2.info() df6= pd.merge(df5, dform_du, how='left', left_on='item_id', right_on='item_id') df6= pd.merge(df6, dform_du2, how='left', left_on='shop_id', right_on='shop_id') df6.info() #dfsummonth = pd.merge(df5, dfprice) #dfsummonth #itemsinfo.head() #newitemsinfo = itemsinfo.drop(['item_name','item_category_name'],axis=1) #newitemsinfo.head() #mergetest = pd.merge(df5, newitemsinfo) #mergetest2 = pd.merge(df5, newitemsinfo, how='left',left_on='item_id', right_on='item_id') #mergetest3 = pd.merge(df5, newitemsinfo, how='left',left_on='item_id', right_on='item_id' ) #mergetest3.head() df6['date'] = pd.to_datetime(df6['date']) df6=df6.set_index('date') df6 = df6.sort_index() df6= df6.reset_index() df6 = df6.rename_axis('ID').reset_index() df6 df6 = df6.drop('date', axis=1) df6 df6 = df6.drop('index', axis=1) #mergetest3.head() from sklearn.preprocessing import StandardScaler scaler = StandardScaler() scaler.fit(df6) X_scaled = scaler.transform(df6) X_scaled = pd.DataFrame(X_scaled, index=df6.index, columns =df6.columns) X_scaled.head() X_scaled.info() #X = X_scaled.drop(['item_cnt_day'],axis=1) #y = df6['item_cnt_day'] len(X_scaled) len(y) #정규화를 하지않고 돌려본다 y = df6.loc[:,'item_cnt_day'] X = df6.drop(['item_cnt_day'], axis=1) from sklearn.model_selection import train_test_split X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.3, random_state=1) from xgboost import XGBRegressor model_reg = XGBRegressor() model_reg.fit(X_train, y_train) from sklearn.metrics import mean_absolute_error, mean_squared_error from math import sqrt pred = model_reg.predict(X_test) print(mean_absolute_error(y_test, pred)) print(sqrt(mean_squared_error(y_test, pred))) pred.mean() # 주요하게 적용하는 변수를 판단 from xgboost import plot_importance import matplotlib.pyplot as plt %matplotlib inline plot_importance(model_reg, height=0.9) test.head() #testmerge = pd.merge(test, newitemsinfo, how='left', left_on='item_id', right_on='item_id') #testmerge #mergefortest = df3.drop(['shop_id', 'item_category_id','item_cnt_day','ID'],axis=1) #mergefortest #df3.head() evalcsv = pd.read_csv('/content/forsubmission.csv') eval = evalcsv eval =pd.merge(eval, itemsinfo,how='left',left_on='item_id',right_on='item_id') eval =pd.merge(eval, shops,how='left',left_on='shop_id',right_on='shop_id') eval = eval.drop('item_name',axis=1) test.head() eval = pd.merge(test, items, how='left', left_on='item_id',right_on='item_id') eval.head() eval = eval.drop('item_price',axis=1) pred = model_reg.predict(eval) pred.mean() adfadsf items.head() evalcsv.head() evalcsv = pd.merge(evalcsv, items, how='left',left_on='item_id',right_on='item_id') evalcsv =evalcsv.drop('item_name',axis=1) submission =pd.DataFrame(pred) submission_copy = submission.rename_axis('ID').reset_index() submission_copy submission_copy.columns = ['ID', 'item_cnt_month'] submission_copy submission_copy.mean() submission_copy.info() submission_copy['item_cnt_month'] = submission_copy['item_cnt_month'].astype(float) submission_copy.to_csv('XG_submission3.csv',index=False) items.head() len(test) ###Output _____no_output_____
01.14/Class_transcript_01_14_LU.ipynb
###Markdown In-class transcript from Lecture 3, January 14, 2019 ###Code # These are the standard imports for CS 111. # This list may change as the quarter goes on. import os import time import math import numpy as np import numpy.linalg as npla import scipy from scipy import sparse from scipy import linalg import scipy.sparse.linalg as spla import matplotlib.pyplot as plt from matplotlib import cm from mpl_toolkits.mplot3d import axes3d %matplotlib tk # Example of an upper triangular matrix U = np.array([[2,7,1,8],[0,2,8,1],[0,0,8,2],[0,0,0,8]]) print(U) # Example of a unit lower triangular matrix L = np.array([[1,0,0,0],[.5,1,0,0],[0,.5,1,0],[-.5,-.5,0,1]]) print(L) # Experiments with a random matrix A = np.round(20*np.random.rand(5,5)) print(A) npla.matrix_rank(A) npla.norm(A) npla.cond(A) # Creating a right-hand side for which we know the answer to Ax=b xorig = np.round(10*np.random.rand(5)) print("original x:", xorig) b = A @ xorig print("right-hand side b:", b) x = npla.solve(A,b) print("computed x:", x) error = x - xorig print("error in x:", error) print("relative norm of error:", npla.norm(x-xorig) / npla.norm(x)) residual = b - A@x print("residual:", residual) print("relative norm of residual:", npla.norm(residual) / npla.norm(b)) # Now a different matrix A and right-hand side vector b A = np.array([[ 2. , 7. , 1. , 8. ], [ 1. , 5.5, 8.5, 5. ], [ 0. , 1. , 12. , 2.5], [-1. , -4.5, -4.5, 3.5]]) print("A:", A,'\n') b = np.array([17. , 2.5, -7. , 10.5]) print("b:", b) # We used Gaussian elimination on the blackboard to triangularize A, giving U print(U) # During Gaussian elimination, we wrote down the multipliers in # a lower triangular array and then put ones on the diagonal, giving L print(L) # The theorem: Gaussian elimination factors A as the product L time U print( L @ U) print() print(A) def LUfactorNoPiv(A): """Factor a square matrix, A == L @ U (no partial pivoting) Parameters: A: the matrix. Outputs (in order): L: the lower triangular factor, same dimensions as A, with ones on the diagonal U: the upper triangular factor, same dimensions as A """ # Check the input m, n = A.shape assert m == n, 'input matrix A must be square' # Make a copy of the matrix that we will transform into L and U LU = A.astype(np.float64).copy() # Eliminate each column in turn for piv_col in range(n): # Update the rest of the matrix pivot = LU[piv_col, piv_col] assert pivot != 0., "pivot is zero, can't continue" for row in range(piv_col + 1, n): multiplier = LU[row, piv_col] / pivot LU[row, piv_col] = multiplier LU[row, (piv_col+1):] -= multiplier * LU[piv_col, (piv_col+1):] # Separate L and U in the result U = np.triu(LU) L = LU - U + np.eye(n) return (L, U) def Lsolve(L, b): """Forward solve a unit lower triangular system Ly = b for y Parameters: L: the matrix, must be square, lower triangular, with ones on the diagonal b: the right-hand side vector Output: y: the solution vector to L @ y == b """ # Check the input m, n = L.shape assert m == n, "matrix L must be square" assert np.all(np.tril(L) == L), "matrix L must be lower triangular" assert np.all(np.diag(L) == 1), "matrix L must have ones on the diagonal" # Make a copy of b that we will transform into the solution y = b.astype(np.float64).copy() # Forward solve for col in range(n): y[col+1:] -= y[col] * L[col+1:, col] return y def Usolve(U, y): """Backward solve an upper triangular system Ux = y for x Parameters: U: the matrix, must be square, upper triangular, with nonzeros on the diagonal y: the right-hand side vector Output: x: the solution vector to U @ x == y """ print("you will write Usolve in hw2") return L,U = LUfactorNoPiv(A) y = Lsolve(L,b) print("y:", y) x = Usolve(U,y) print("\nx:", x) print("\nresidual norm:", npla.norm(b - A @ x)) A @ [1,0,-1,2] # But LU factorization (without pivoting) fails if it encounters a zero pivot A = np.array([[0, 1], [1, 2]]) L,U = LUfactorNoPiv(A) def LUfactor(A, pivoting = True): """Factor a square matrix with partial pivoting, A[p,:] == L @ U Parameters: A: the matrix. pivoting: whether or not to do partial pivoting Outputs (in order): L: the lower triangular factor, same dimensions as A, with ones on the diagonal U: the upper triangular factor, same dimensions as A p: the permutation vector that permutes the rows of A by partial pivoting """ # Check the input m, n = A.shape assert m == n, 'input matrix A must be square' # Initialize p to be the identity permutation p = np.array(range(n)) # Make a copy of the matrix that we will transform into L and U LU = A.astype(np.float64).copy() # Eliminate each column in turn for piv_col in range(n): # Choose the pivot row and swap it into place if pivoting: piv_row = piv_col + np.argmax(LU[piv_col:, piv_col]) assert LU[piv_row, piv_col] != 0., "can't find nonzero pivot, matrix is singular" LU[[piv_col, piv_row], :] = LU[[piv_row, piv_col], :] p[[piv_col, piv_row]] = p[[piv_row, piv_col]] # Update the rest of the matrix pivot = LU[piv_col, piv_col] assert pivot != 0., "pivot is zero, can't continue" for row in range(piv_col + 1, n): multiplier = LU[row, piv_col] / pivot LU[row, piv_col] = multiplier LU[row, (piv_col+1):] -= multiplier * LU[piv_col, (piv_col+1):] # Separate L and U in the result U = np.triu(LU) L = LU - U + np.eye(n) return (L, U, p) # Our 2-by-2 counterexample again print('A:\n', A) L, U, p = LUfactor(A) print('\nL:\n', L) print('\nU:\n', U) print('\np: ', p) # A larger example of LU with partial pivoting A = np.round(20*np.random.rand(5,5)) print('matrix A:\n', A) xorig = np.round(10*np.random.rand(5)) print('\noriginal x:', xorig) b = A @ xorig print('\nright-hand side b:', b) # Factor the larger example L, U, p = LUfactor(A) print("norm of difference between L times U and permuted A:", npla.norm( L@U - A[p,:])) # Solve with the larger example y = Lsolve(L,b[p]) print("y:", y) x = Usolve(U,y) print("\nx:", x) print("\nresidual norm:", npla.norm(b - A @ x)) ###Output y: [340. 184. 85.90909091 284.97916667 103.91794872] x: [7. 2. 6. 2. 8.] residual norm: 6.355287432313019e-14
notebooks/test_interpolation.ipynb
###Markdown First let's make a sample curvilinear grid. The size is given by Nx, and Ny ###Code Nx = 80 Ny = 60 def make_sample_grid(Ny, Nx): 'return sample grid of dimension [Ny, Nx]' yc, xc = np.mgrid[1:10:Ny*1j, 1:20:Nx*1J] def rot2d(x, y, ang): '''rotate vectors by geometric angle''' xr = x*np.cos(ang) - y*np.sin(ang) yr = x*np.sin(ang) + y*np.cos(ang) return xr, yr x, y = rot2d(xc, (5+yc)**1.2*(3+xc)**0.3, 0.2) y /= y.ptp()/10. x /= x.ptp()/10. x -= x.mean() y -= y.mean() return x, y x, y = make_sample_grid(Ny, Nx) ###Output _____no_output_____ ###Markdown Now some trial points where the gridded values will be interpolated. These cover a number of test cases at specific grid locations, or random points. ###Code # Some sample grid locations x_nodes, y_nodes = x.flatten(), y.flatten() x_u = 0.5*(x[:, 1:] + x[:, :-1]).flatten() y_u = 0.5*(y[:, 1:] + y[:, :-1]).flatten() x_v = 0.5*(x[1:, :] + x[:-1, :]).flatten() y_v = 0.5*(y[1:, :] + y[:-1, :]).flatten() x_centers = 0.25*(x[1:, 1:] + x[1:, :-1] + x[:-1, 1:] + x[:-1, :-1]).flatten() y_centers = 0.25*(y[1:, 1:] + y[1:, :-1] + y[:-1, 1:] + y[:-1, :-1]).flatten() # centers offset halfway toward the lower right node x_nc = 0.5*(0.25*(x[1:, 1:] + x[1:, :-1] + x[:-1, 1:] + x[:-1, :-1]) + x[:-1, 1:]).flatten() y_nc = 0.5*(0.25*(y[1:, 1:] + y[1:, :-1] + y[:-1, 1:] + y[:-1, :-1]) + y[:-1, 1:]).flatten() x_rand, y_rand = 3.0*np.random.randn(2, 100000) ###Output _____no_output_____ ###Markdown Change the values of points, xi, and yi to the desired points to test. ###Code points = np.vstack((x_rand, y_rand)).T xi, yi = points.T ###Output _____no_output_____ ###Markdown Generate the cell tree, locate the points ###Code nodes, faces = interpolate.array2grid(x, y) squares = np.array([nodes[face] for face in faces]) ct2d = cell_tree2d.CellTree(nodes, faces) gridpoint_indices = ct2d.locate(points) ###Output _____no_output_____ ###Markdown Sanity check. See if cell_tree2d really finds the right cell. This can take some time.... ###Code eps = 0.01 # some buffering required if the points are along the edge of a square. # remove indices with value -1, outside the grid domain inside = gridpoint_indices >= 0 # good points, inside domain gridpoint_indices = gridpoint_indices[inside] points_i = points[inside] squares_i = squares[gridpoint_indices] mesh = MultiPolygon([Polygon(p).buffer(eps) for p in squares_i]) trial = MultiPoint([Point(p) for p in points_i]) contains = [m.contains(p) for m, p in zip(mesh, trial)] assert(np.alltrue(contains)) ###Output _____no_output_____ ###Markdown Now check interpolation. This is fast.. ###Code def zfunc(x, y): 'Sample field for interpolation' return np.sin(x/10.) + np.cos(y/10.) ct = interpolate.CellTree(x, y) loc = ct.locate(points) zgrid = zfunc(x, y) zi = loc.interpolate(zgrid) # use loc.points, as this contains only the points in the domain. zi_true = zfunc(*loc.points.T) assert(np.allclose(zi, zi_true, rtol=0.0001)) ###Output _____no_output_____ ###Markdown Plot this up some.. ###Code fig = plt.figure(figsize=(14, 10)) ax = fig.add_subplot(111) zc = zfunc( 0.25*(x[1:, 1:] + x[1:, :-1] + x[:-1, 1:] + x[:-1, :-1]), 0.25*(y[1:, 1:] + y[1:, :-1] + y[:-1, 1:] + y[:-1, :-1]) ) ax.pcolormesh(x, y, zgrid, cmap='viridis', clim=(0.5, 1.5)) # we only want to plot the points that are within the grid domain. ax.scatter(loc.points[:, 0], loc.points[:, 1], 10, zi, cmap='viridis', edgecolor='none') ax.plot(xi, yi, '.k', alpha=0.5, markersize=0.5) # all points ax.plot(loc.points[:, 0], loc.points[:, 1], '.k', markersize=1) # only points in the domain ax.set_xlim(-6, 6) ax.set_ylim(-6, 6) ax.set_aspect(1.0) ###Output _____no_output_____
Hotels rating prediction/Hotel reviews - baseline solution.ipynb
###Markdown Train set, testing set ###Code from sklearn import linear_model Z['Combined'] = Z['Title']+' | '+Z['Text'] X_train = Z.loc[train_inds,'Combined'].values Y_train = Z.loc[train_inds,'Rating'].values X_test = Z.loc[test_inds,'Combined'].values Y_test = Z.loc[test_inds,'Rating'].values text_model = Pipeline([('vect', CountVectorizer()), ('tfidf', TfidfTransformer()), ('model', linear_model.Ridge()), ]) from sklearn.model_selection import GridSearchCV parameters = {'vect__ngram_range': [(1, 2)], 'model__alpha': 10**linspace(-5,5,6), } gs_model = GridSearchCV(text_model, parameters, scoring='neg_mean_squared_error', n_jobs=-1) gs_model = gs_model.fit(X_train, Y_train) gs_model.best_score_, gs_model.best_params_ sqrt(-gs_model.best_score_) ###Output _____no_output_____ ###Markdown Submission ###Code Z.loc[train_inds].index print len(X_train), len(X_test) concatenate((X_train,[X_test[0]])) Y_hat = gs_model.predict(X_train) Y_hat = [20 if i<20 else 100 if i>100 else round(i,4) for i in Y_hat] Y_hat[6] S = pd.DataFrame(Y_hat, columns=['Rating'], index = Z.loc[train_inds].index) S.head() S.to_csv('Data/solution.csv', index=True) ###Output _____no_output_____
sample-code/notebooks/3-05.ipynb
###Markdown 第3章 pandasでデータを処理しよう 3-5: さまざまなデータの読み込み ###Code # リスト3.5.1:CSVファイルの読み込み import os import pandas as pd base_url = ( "https://raw.githubusercontent.com/practical-jupyter/sample-data/master/anime/" ) anime_csv = os.path.join(base_url, "anime.csv") df = pd.read_csv(anime_csv) df.head() # リスト3.5.2:インデックス列を番号で指定 # インデックスにする列を番号で指定 df = pd.read_csv(anime_csv, index_col=0) df.head() # リスト3.5.3:インデックス列を列名で指定 # インデックスにする列を列名で指定 df = pd.read_csv(anime_csv, index_col="anime_id") df.head() # リスト3.5.4:型を指定 df = pd.read_csv(anime_csv, dtype={"members": float}) df.head() # リスト3.5.5:datetime型の変換 anime_stock_price_csv = os.path.join(base_url, "anime_stock_price.csv") df = pd.read_csv(anime_stock_price_csv, parse_dates=["Date"]) df.dtypes # リスト3.5.7:区切り文字を指定 anime_tsv = os.path.join(base_url, "anime.tsv") df = pd.read_csv(anime_tsv, sep="\t") # リスト3.5.8:Excelファイルの読み込み anime_xlsx = os.path.join(base_url, "anime.xlsx") df = pd.read_excel(anime_xlsx) df.head() # リスト3.5.9:シート名を指定した読み込み df = pd.read_excel(anime_xlsx, sheetname="Movie") # リスト3.5.10:SQLiteの読み込み from urllib.request import urlopen import sqlite3 anime_db = os.path.join(base_url, "anime.db") res = urlopen(anime_db) with open("anime.db", "wb") as f: f.write(res.read()) with sqlite3.connect(f.name) as conn: df = pd.read_sql("SELECT * FROM anime", conn) # リスト3.5.11:HTMLファイルの読み込み url = "https://docs.python.org/3/py-modindex.html" tables = pd.read_html(url, index_col=1) tables[0].loc[:, 1:].dropna().head(10) # 1番目のDataFrameから空の列と欠損値を除外 ###Output _____no_output_____
CN DSA/CA problems.ipynb
###Markdown Find the Equillbirium Point in Array ###Code arr = [2, 3, 10, -10, 4, 2, 9] n = 7 # complexity O(n^2) def sum_arr(arr): ans = 0 for i in range(len(arr)): ans += arr[i] return ans for i in range(n): if sum_arr(arr[:i]) == sum_arr(arr[i+1:]): print(i) break # complexity O(n) left = 0 right = sum(arr) for i in range(n): right = right - arr[i] if left == right: print(i) break left += arr[i] ###Output _____no_output_____ ###Markdown Find the Unique ElementYou have been given an integer array/list(ARR) of size N. Where N is equal to [2M + 1].Now, in the given array/list, 'M' numbers are present twice and one number is present only once.You need to find and return that number which is unique in the array/list. Note:Unique element is always present in the array/list according to the given condition.Input format :The first line contains an Integer 't' which denotes the number of test cases or queries to be run. Then the test cases follow.First line of each test case or query contains an integer 'N' representing the size of the array/list.Second line contains 'N' single space separated integers representing the elements in the array/list.Output Format :For each test case, print the unique element present in the array.Output for every test case will be printed in a separate line.Constraints :1 <= t <= 10^20 <= N <= 10^6Time Limit: 1 secSample Input 1:172 3 1 6 3 6 2Sample Output 1:1Sample Input 2:252 4 7 2 791 3 1 3 6 6 7 10 7Sample Output 2:410 ###Code arr = [2, 3, 2, 4, 6, 3, 6] n = 7 ###Output _____no_output_____ ###Markdown Solution 1 Complexity = O(n^2) ###Code def find_unique(n,arr): for i in range(n): for j in range(n+1): if j == n: return arr[i] if i != j: if arr[i] == arr[j]: break find_unique(7,arr) ###Output _____no_output_____ ###Markdown Solution 2 Complexity O(nlogn) We need to sort first and sorting takes nlogn and then for loop will take logn complexity ###Code merge_sort(arr) print(arr) for i in range(0,n,2): if arr[i] != arr[i+1]: print(arr[i]) break ###Output _____no_output_____ ###Markdown Solution 3 Complexity O(n) solved by biwise XOR => ^ operator that return 0 if opearated with same number otherwise something else.i.e in arr = [2,3,4,6,3,6,2]This approach uses 2 properties of XOR:1. XOR of a number with itself is 0.2. XOR of a number with 0 is number itself.Let us understand this approach with the help of an example:arr[]= 2 3 1 6 3 6 2Taking their xor:Answer = 2 ^ 3 ^ 1 ^ 6 ^ 3 ^ 6 ^ 2Now XOR is associative and commutative, so I can write it as:Answer = (2^2) ^ (3^3) ^ 1 ^ (6 ^ 6) = 0 ^ 0 ^ 1 ^ 0 = 1 Time complexity of this solution is O(n) ###Code arr = [2, 3, 1, 6, 3, 6, 2] n = 7 ans = 0 for i in range(n): ans = ans ^ arr[i] print(ans) ###Output _____no_output_____ ###Markdown Find duplicate element in the Given ArrayDuplicate in arraySend FeedbackYou have been given an integer array/list(ARR) of size N which contains numbers from 0 to (N - 2). Each number is present at least once. That is, if N = 5, the array/list constitutes values ranging from 0 to 3, and among these, there is a single integer value that is present twice. You need to find and return that duplicate number present in the array.Note :Duplicate number is always present in the given array/list.Input format :The first line contains an Integer 't' which denotes the number of test cases or queries to be run. Then the test cases follow.First line of each test case or query contains an integer 'N' representing the size of the array/list.Second line contains 'N' single space separated integers representing the elements in the array/list.Output Format :For each test case, print the duplicate element in the array/list.Output for every test case will be printed in a separate line.Constraints :1 <= t <= 10^20 <= N <= 10^6Time Limit: 1 secSample Input 1:190 7 2 5 4 7 1 3 6Sample Output 1:7Sample Input 2:250 2 1 3 170 3 1 5 4 3 2Sample Output 2:13 Solution Find duplicateProblem Description: You are given with an array of integers of size n which contains numbersfrom 0 to n - 2. Each number is present at least once. That is, if n = 5, numbers from 0 to 3 ispresent in the given array at least once and one number is present twice. You need to find andreturn that duplicate number present in the array.How to approach?Approach 1: In this question you need to run two loops, pick an element from the first loop andthen in the inner loop check if the element appears once again or not, if yes then return thatelement, otherwise move to the next element.This method doesn’t use the other useful data provided in questions like range of numbers isbetween 0 to n-2 and hence, it is increasing the time complexity.Pseudo Code for this approach:Function Findduplicate: For i = 0 to i less than size: For j = 0 to j less than size: If i not equal to j and arr[i] is equal to arr[j]: Return arr[i] return minus infinity Time Complexity for this approach: Time complexity for this approach is O( n ), which is not 2good, hence we move to the next approach.Approach 2: A better solution for this problem can be by using XOR operator. Using XORoperator, we can solve this problem in one traversal only. The following facts about XORoperation will be useful for this question:1. If we XOR a number by itself, even number of times then it will give you 0.2. If we XOR a number with itself, odd number of times, then it will give you the numberitself.3. Also XOR of a number with 0 gives you that number again.So, if we take XOR of all the elements present in the array with every element in the range 0 ton-2, then all the elements of that array except the duplicate element are XORed 2 times andhence, their resultant is 0. But the duplicate element is XORed 3 times, hence, its resultant is thenumber itself. Hence, you will get you answer as the duplicate number present in the array.For example, if you are given with n=5 and let us say array is 0 1 3 2 2, then according to thisapproach, we have to XOR all elements present in the array with every element in the range 0 to3.Answer = (0^1^3^2^2)^(0^1^2^3)As XOR operation is associative and commutative, so, by rearrangingAnswer = (0^0) ^ (1^1) ^ (2^2^2) ^ (3^3) = 0 ^ 0 ^ 2 ^ 0 = 2Pseudo Code for this approach:Function Findduplicate:answer=0For i =0 to i less than n:answer=answer xor arr[i]For i=0 to i less than or equal to n-2:answer=answer xor iReturn answerTime Complexity for this approach: Time complexity for this approach is O(n) as you aretraversing the array only once for XORing.Approach 3: Another approach is to make use of the condition that all elements lies between 0and n-2. So first calculate the sum of all natural numbers between 0 to n-2 by using the directformula ((n - 1) * (n - 2)) / 2 and sum of all elements of the array. Now, subtract the sum of allnatural numbers between 0 to n-2 from sum of all elements of the array. This will give you theduplicate element present in the array.Pseudo Code for this approach:Function findduplicate: sum=0 For i = 0 to i less than size: sum = sum + input[i]; n = size sumOfNaturalNumbers = ((n - 1) * (n - 2)) / 2 return sum - sumOfNaturalNumbers Time Complexity for this approach: Time complexity for this approach is O(n) as you aretraversing the array only once to calculate the sum of all elements present in the array.❏ Let us dry run the code for the N= 9 arr[]= 0 7 2 5 4 7 1 3 6 Sum = 0+7+2+5+4+7+1+3+6 =35sumOfNaturalNumbers=8*7/2=28Output = 35-28 =7So 7 should get printed. ###Code arr = [0, 7, 2, 5, 4, 7, 1, 3, 6] n = len(arr) print(arr, n) def find_dup(arr,n): for i in range(n): for j in range(i+1, n): if arr[i] == arr[j]: return arr[i] find_dup(arr,n) merge_sort(arr) for i in range(n-1): if arr[i] == arr[i+1]: print(arr[i]) def find_dup_xor(arr,n): ans = 0 for i in range(n): ans = ans ^ arr[i] for i in range(n-1): ans = ans ^ i return ans find_dup_xor(arr,n) def sum_arr(arr): ans = 0 for i in range(len(arr)): ans += arr[i] return ans def find_dup_sum(arr,n): total_sum = sum_arr(arr) n2_sum = (n-1)*(n-2) // 2 return total_sum - n2_sum find_dup_sum(arr,n) ###Output _____no_output_____ ###Markdown Find the number of Pairs of Elements in the Array that sums equal to the given number ###Code ## Pair sum in array arr = [1,3,6,2,5,4,3,2,4,20] n = len(arr) num = 7 print(arr, n, num) ###Output _____no_output_____ ###Markdown Solution 1. O(n^2) ###Code ## O(n^2) c = 0 for i in range(n): for j in range(i+1,n): if arr[i] + arr[j] == num: c += 1 print(c) ###Output _____no_output_____ ###Markdown Solution 2. O(n) but looping two times ###Code ## creating map # Space taking but O(n) and taking two for loops m = [0]*1000 for i in range(n): m[arr[i]] += 1 twice_count = 0 for i in range(n): twice_count += m[num - arr[i]] if num - arr[i] == arr[i]: twice_count -= 1 print(twice_count//2) ###Output _____no_output_____ ###Markdown Solution 3. O(n) in single for loop ###Code ## Single loop m = {} count = 0 for i in range(n): if num - arr[i] in m: count += m[num - arr[i]] if arr[i] in m: m[arr[i]] += 1 else: m[arr[i]] = 1 print(count) ###Output _____no_output_____ ###Markdown Qus: Rotate array left side k times ###Code arr = [2,4,3,5,1] n = len(arr) k = 3 arr = [1,2,3,4,5,6,7,8] n= len(arr) k = 3 ###Output _____no_output_____ ###Markdown By List Slicing (Easy Solution) ###Code # Slicing method k = k%n print(arr[k:] + arr[:k]) ###Output _____no_output_____ ###Markdown By one by one rotation (Bad Solution) O(kn) ###Code # rotate array k = k%n for i in range(k): s = arr[0] for i in range(n-1): arr[i] = arr[i+1] arr[n-1] = s print(arr) ###Output _____no_output_____ ###Markdown Using O(k) extra Space ###Code k = k%n # store first k elements in some temp temp = [0]*k for i in range(k): temp[i] = arr[i] # make ith element as i+k th element for i in range(n-k): arr[i] = arr[i+k] # replace last k elements by the temp for i in range(n-k,n): arr[i] = temp[i-n+k] arr ###Output _____no_output_____ ###Markdown By Reversing the Array ###Code # Reverse the Array by Two Pointer Approach def rev_arr(arr,si,ei): while si < ei: arr[si], arr[ei] = arr[ei], arr[si] si += 1 ei -=1 k = k%n rev_arr(arr,0,n-1) print(arr) # reverse the first n-k elements and reverse last k elemets separately for desired result rev_arr(arr,0,n-k-1) rev_arr(arr,n-k,n-1) ###Output _____no_output_____ ###Markdown Find the number of triplets in the array which sum equal to num ###Code arr = [2,-5,8,-6,0,5,10,11,-3] n= len(arr) num = 10 ###Output _____no_output_____ ###Markdown Soultion 1: Three loops O(n^3) ###Code count = 0 for i in range(n-2): for j in range(i+1, n-1): for k in range(j+1, n): if arr[i]+arr[j]+arr[k] == num: count += 1 print(count) ###Output 5 ###Markdown Solution 2: Sort the Array and the use two pointer approach ###Code arr = [2,-5,8,-6,0,5,10,11,-3] n= len(arr) num = 10 count = 0 arr.sort() for i in range(n-2): l = i + 1 r = n - 1 while (l<r): if arr[i]+arr[l]+arr[r] == num: count += 1 l+=1 r-=1 elif arr[i]+arr[l]+arr[r] < num: l+=1 else: r-=1 print(count) ###Output 5 ###Markdown Solution 3: Storing the map ###Code count = 0 for i in range(n-1): s = set() for j in range(i+1, n): if num - arr[i] - arr[j] in s: count += 1 s.add(arr[j]) count ###Output _____no_output_____
class/MODULE_1-python_introduction/.ipynb_checkpoints/10.02_CondicionesBucles-checkpoint.ipynb
###Markdown 2- Condiciones y BuclesCurso Introducción a Python - Tecnun, Universidad de Navarra En este documento nos centraremos en la creación de condiciones y bucles. A diferencia de otros lenguajes de programación no se utilizan llaves o sentencias *end* para determinar lo que está incluido dentro de la condición o el bucle. En Python, todo esto se hace mediante indentación. A continuación veremos unos ejemplos. Condiciones La sintaxis general de las condiciones es la siguiente: ###Code a=1 if a==1: ###Output _____no_output_____ ###Markdown Los comandos que se emplean para comparar son los siguientes:- **==** y **!=** para comprobar igualdad o desigualdad, respectivamente.- **\>** y **\<** para comprobar si un elemento es estrictamente mayor o estrictamente menor que otro, respectivamente.- **>=** y **<=** para comprobar si un elemento es mayor o igual, o menor o igual que otro, respectivamente.En el caso de cumplir la condición, la comprobación devolverá una variable booleana *True* y ejecutará las líneas correspondientes a dicha condición. Si, por el contrario, la condición no se satisface, obtendremos una variable booleana *False* y no se ejecutaran las lineas correspondientes a la condición.En el caso de que fuera necesario comprobar si se cumplen varias condiciones a la vez se pueden utilizar los operadores booleanos **and** y **or**. Aparte de este tipo de comprobaciones, se puede mirar si una lista contiene un elemento empleando el comando **in**. En el caso de querer negar condiciones, se puede emplear el operador booleano **not**. Siempre y cuando tenga sentido, estos operadores se pueden emplear con cualquier tipo de variables. Bucles *for* La sintaxis general de los bucles *for* es la siguiente: Es importante darse cuenta de que el comando *range(0, 3)* crea una **sucesión de números entre 0 y 2**. Los comandos *break* y *continue* pueden resultar muy útiles. El primero termina el bucle en el instante de su ejecución, y el segundo termina la iteración actual del bucle y pasa a la siguiente. De la misma manera que ocurre con las condiones, las variables que empleamos como contador en los bucles no tienen por qué ser numéricas. Bucles *while* La sintaxis general de los bucles *while* es la siguiente: El operador **+=** aumenta el valor de la variable i en el valor que escribamos a su derecha en cada iteración. Por el contrario, el operador **-=** lo disminuye. Al igual que con los bucles *for*, los operadores *break* y *continue* son válidos en estos bucles. ###Code ###Output _____no_output_____
Semana-18/.ipynb_checkpoints/Tensor Flow-checkpoint.ipynb
###Markdown Paralelizacion de entrenamiento de redes neuronales con TensorFlowEn esta seccion dejaremos atras los rudimentos de las matematicas y nos centraremos en utilizar TensorFlow, la cual es una de las librerias mas populares de arpendizaje profundo y que realiza una implementacion mas eficaz de las redes neuronales que cualquier otra implementacion de Numpy.TensorFlow es una interfaz de programacion multiplataforma y escalable para implementar y ejecutar algortimos de aprendizaje automatico de una manera mas eficaz ya que permite usar tanto la CPU como la GPU, la cual suele tener muchos mas procesadores que la CPU, los cuales, combinando sus frecuencias, presentan un rendimiento mas potente. La API mas desarrollada de esta herramienta se presenta para Python, por lo cual muchos desarrolladores se ven atraidos a este lenguaje. Primeros pasos con TensorFlowhttps://jakevdp.github.io/PythonDataScienceHandbook/02.01-understanding-data-types.html ###Code # Creando tensores # ============================================= import tensorflow as tf import numpy as np np.set_printoptions(precision=3) a = np.array([1, 2, 3], dtype=np.int32) b = [4, 5, 6] t_a = tf.convert_to_tensor(a) t_b = tf.convert_to_tensor(b) print(t_a) print(t_b) # Obteniendo las dimensiones de un tensor # =============================================== t_ones = tf.ones((2, 3)) print(t_ones) t_ones.shape # Obteniendo los valores del tensor como array # =============================================== t_ones.numpy() # Creando un tensor de valores constantes # ================================================ const_tensor = tf.constant([1.2, 5, np.pi], dtype=tf.float32) print(const_tensor) matriz = np.array([[2, 3, 4, 5], [6, 7, 8, 8]], dtype = np.int32) matriz matriz_tf = tf.convert_to_tensor(matriz) print(matriz_tf, end = '\n'*2) print(matriz_tf.numpy(), end = '\n'*2) print(matriz_tf.shape) ###Output tf.Tensor( [[2 3 4 5] [6 7 8 8]], shape=(2, 4), dtype=int32) [[2 3 4 5] [6 7 8 8]] (2, 4) ###Markdown Manipulando los tipos de datos y forma de un tensor ###Code # Cambiando el tipo de datos del tensor # ============================================== print(matriz_tf.dtype) matriz_tf_n = tf.cast(matriz_tf, tf.int64) print(matriz_tf_n.dtype) # Transponiendo un tensor # ================================================= t = tf.random.uniform(shape=(3, 5)) print(t, end = '\n'*2) t_tr = tf.transpose(t) print(t_tr, end = '\n'*2) # Redimensionando un vector # ===================================== t = tf.zeros((30,)) print(t, end = '\n'*2) print(t.shape, end = '\n'*3) t_reshape = tf.reshape(t, shape=(5, 6)) print(t_reshape, end = '\n'*2) print(t_reshape.shape) # Removiendo las dimensiones innecesarias # ===================================================== t = tf.zeros((1, 2, 1, 4, 1)) print(t, end = '\n'*2) print(t.shape, end = '\n'*3) t_sqz = tf.squeeze(t, axis=(2, 4)) print(t_sqz, end = '\n'*2) print(t_sqz.shape, end = '\n'*3) print(t.shape, ' --> ', t_sqz.shape) ###Output tf.Tensor( [[[[[0.] [0.] [0.] [0.]]] [[[0.] [0.] [0.] [0.]]]]], shape=(1, 2, 1, 4, 1), dtype=float32) (1, 2, 1, 4, 1) tf.Tensor( [[[0. 0. 0. 0.] [0. 0. 0. 0.]]], shape=(1, 2, 4), dtype=float32) (1, 2, 4) (1, 2, 1, 4, 1) --> (1, 2, 4) ###Markdown Operaciones matematicas sobre tensores ###Code # Inicializando dos tensores con numeros aleatorios # ============================================================= tf.random.set_seed(1) t1 = tf.random.uniform(shape=(5, 2), minval=-1.0, maxval=1.0) t2 = tf.random.normal(shape=(5, 2), mean=0.0, stddev=1.0) print(t1, '\n'*2, t2) # Producto tipo element-wise: elemento a elemento # ================================================= t3 = tf.multiply(t1, t2).numpy() print(t3) # Promedio segun el eje # ================================================ t4 = tf.math.reduce_mean(t1, axis=None) print(t4, end = '\n'*3) t4 = tf.math.reduce_mean(t1, axis=0) print(t4, end = '\n'*3) t4 = tf.math.reduce_mean(t1, axis=1) print(t4, end = '\n'*3) # suma segun el eje # ================================================= t4 = tf.math.reduce_sum(t1, axis=None) print('Suma de todos los elementos:', t4, end = '\n'*3) t4 = tf.math.reduce_sum(t1, axis=0) print('Suma de los elementos por columnas:', t4, end = '\n'*3) t4 = tf.math.reduce_sum(t1, axis=1) print('Suma de los elementos por filas:', t4, end = '\n'*3) # Desviacion estandar segun el eje # ================================================= t4 = tf.math.reduce_std(t1, axis=None) print('Suma de todos los elementos:', t4, end = '\n'*3) t4 = tf.math.reduce_std(t1, axis=0) print('Suma de los elementos por columnas:', t4, end = '\n'*3) t4 = tf.math.reduce_std(t1, axis=1) print('Suma de los elementos por filas:', t4, end = '\n'*3) # Producto entre matrices # =========================================== t5 = tf.linalg.matmul(t1, t2, transpose_b=True) print(t5.numpy(), end = '\n'*2) # Producto entre matrices # =========================================== t6 = tf.linalg.matmul(t1, t2, transpose_a=True) print(t6.numpy()) # Calculando la norma de un vector # ========================================== norm_t1 = tf.norm(t1, ord=2, axis=None).numpy() print(norm_t1, end='\n'*2) norm_t1 = tf.norm(t1, ord=2, axis=0).numpy() print(norm_t1, end='\n'*2) norm_t1 = tf.norm(t1, ord=2, axis=1).numpy() print(norm_t1, end='\n'*2) ###Output 1.5818709 [1.303 0.897] [1.046 0.293 0.504 0.96 0.383] ###Markdown Partir, apilar y concatenar tensores ###Code # Datos a trabajar # ======================================= tf.random.set_seed(1) t = tf.random.uniform((6,)) print(t.numpy()) # Partiendo el tensor en un numero determinado de piezas # ====================================================== t_splits = tf.split(t, num_or_size_splits = 3) [item.numpy() for item in t_splits] # Partiendo el tensor segun los tamaños definidos # ====================================================== tf.random.set_seed(1) t = tf.random.uniform((6,)) print(t.numpy()) t_splits = tf.split(t, num_or_size_splits=[3, 3]) [item.numpy() for item in t_splits] print(matriz_tf.numpy()) # m_splits = tf.split(t, num_or_size_splits = 0, axis = 1) matriz_n = tf.reshape(matriz_tf, shape = (8,)) print(matriz_n.numpy()) m_splits = tf.split(matriz_n, num_or_size_splits = 2) [item.numpy() for item in m_splits] # Concatenando tensores # ========================================= A = tf.ones((3,)) print(A, end ='\n'*2) B = tf.zeros((2,)) print(B, end ='\n'*2) C = tf.concat([A, B], axis=0) print(C.numpy()) # Apilando tensores # ========================================= A = tf.ones((3,)) print(A, end ='\n'*2) B = tf.zeros((3,)) print(B, end ='\n'*2) S = tf.stack([A, B], axis=1) print(S.numpy()) ###Output tf.Tensor([1. 1. 1.], shape=(3,), dtype=float32) tf.Tensor([0. 0. 0.], shape=(3,), dtype=float32) [[1. 0.] [1. 0.] [1. 0.]] ###Markdown Mas funciones y herramientas en:https://www.tensorflow.org/versions/r2.0/api_docs/python/tf. EJERCICIOS1. Cree dos tensores de dimensiones (4, 6), de numeros aleatorios provenientes de una distribucion normal estandar con promedio 0.0 y dsv 1.0. Imprimalos.2. Multiplique los anteriores tensores de las dos formas vistas, element-wise y producto matricial, realizando las dos transposiciones vistas. 3. Calcule los promedios, desviaciones estandar y suma de sus elementos para los dos tensores.4. Redimensione los tensores para que sean ahora de rango 1.5. Calcule el coseno de los elementos de los tensores (revise la documentacion).6. Cree un tensor de rango 1 con 1001 elementos, empezando con el 0 y hasta el 30.7. Realice un for sobre los elementos del tensor e imprimalos.8. Realice el calculo de los factoriales de los numero del 1 al 30 usando el tensor del punto 6. Imprima el resultado como un DataFrame Creación de *pipelines* de entrada con tf.data: la API de conjunto de datos de TensorFlowCuando entrenamos un modelo NN profundo, generalmente entrenamos el modelo de forma incremental utilizando un algoritmo de optimización iterativo como el descenso de gradiente estocástico, como hemos visto en clases anteriores.La API de Keras es un contenedor de TensorFlow para crear modelos NN. La API de Keras proporciona un método, `.fit ()`, para entrenar los modelos. En los casos en que el conjunto de datos de entrenamiento es bastante pequeño y se puede cargar como un tensor en la memoria, los modelos de TensorFlow (que se compilan con la API de Keras) pueden usar este tensor directamente a través de su método .fit () para el entrenamiento. Sin embargo, en casos de uso típicos, cuando el conjunto de datos es demasiado grande para caber en la memoria de la computadora, necesitaremos cargar los datos del dispositivo de almacenamiento principal (por ejemplo, el disco duro o la unidad de estado sólido) en trozos, es decir, lote por lote. Además, es posible que necesitemos construir un *pipeline* de procesamiento de datos para aplicar ciertas transformaciones y pasos de preprocesamiento a nuestros datos, como el centrado medio, el escalado o la adición de ruido para aumentar el procedimiento de entrenamiento y evitar el sobreajuste.Aplicar las funciones de preprocesamiento manualmente cada vez puede resultar bastante engorroso. Afortunadamente, TensorFlow proporciona una clase especial para construir *pipelines* de preprocesamiento eficientes y convenientes. En esta parte, veremos una descripción general de los diferentes métodos para construir un conjunto de datos de TensorFlow, incluidas las transformaciones del conjunto de datos y los pasos de preprocesamiento comunes. Creando un Dataset de TensorFlow desde tensores existentesSi los datos ya existen en forma de un objeto tensor, una lista de Python o una matriz NumPy, podemos crear fácilmente un conjunto de datos usando la función `tf.data.Dataset.from_tensor_ slices()`. Esta función devuelve un objeto de la clase Dataset, que podemos usar para iterar a través de los elementos individuales en el conjunto de datos de entrada: ###Code # Ejemplo con listas # ====================================================== a = [1.2, 3.4, 7.5, 4.1, 5.0, 1.0] ds = tf.data.Dataset.from_tensor_slices(a) print(ds) for item in ds: print(item) ###Output _____no_output_____ ###Markdown Si queremos crear lotes a partir de este conjunto de datos, con un tamaño de lote deseado de 3, podemos hacerlo de la siguiente manera: ###Code # Creando lotes de 3 elementos cada uno # =================================================== ds_batch = ds.batch(3) for i, elem in enumerate(ds_batch, 1): print(f'batch {i}:', elem.numpy()) ###Output _____no_output_____ ###Markdown Esto creará dos lotes a partir de este conjunto de datos, donde los primeros tres elementos van al lote n° 1 y los elementos restantes al lote n° 2. El método `.batch()` tiene un argumento opcional, `drop_remainder`, que es útil para los casos en los que el número de elementos en el tensor no es divisible por el tamaño de lote deseado. El valor predeterminado de `drop_remainder` es `False`. Combinar dos tensores en un DatasetA menudo, podemos tener los datos en dos (o posiblemente más) tensores. Por ejemplo, podríamos tener un tensor para características y un tensor para etiquetas. En tales casos, necesitamos construir un conjunto de datos que combine estos tensores juntos, lo que nos permitirá recuperar los elementos de estos tensores en tuplas.Suponga que tenemos dos tensores, t_x y t_y. El tensor t_x contiene nuestros valores de características, cada uno de tamaño 3, y t_y almacena las etiquetas de clase. Para este ejemplo, primero creamos estos dos tensores de la siguiente manera: ###Code # Datos de ejemplo # ============================================ tf.random.set_seed(1) t_x = tf.random.uniform([4, 3], dtype=tf.float32) t_y = tf.range(4) print(t_x) print(t_y) # Uniendo los dos tensores en un Dataset # ============================================ ds_x = tf.data.Dataset.from_tensor_slices(t_x) ds_y = tf.data.Dataset.from_tensor_slices(t_y) ds_joint = tf.data.Dataset.zip((ds_x, ds_y)) for example in ds_joint: print('x:', example[0].numpy(),' y:', example[1].numpy()) ds_joint = tf.data.Dataset.from_tensor_slices((t_x, t_y)) for example in ds_joint: #print(example) print('x:', example[0].numpy(), ' y:', example[1].numpy()) # Operacion sobre el dataset generado # ==================================================== ds_trans = ds_joint.map(lambda x, y: (x*2-1.0, y)) for example in ds_trans: print(' x:', example[0].numpy(), ' y:', example[1].numpy()) ###Output x: [-0.67 0.803 0.262] y: 0 x: [-0.131 -0.416 0.285] y: 1 x: [ 0.952 -0.13 0.32 ] y: 2 x: [0.21 0.273 0.229] y: 3 ###Markdown Mezclar, agrupar y repetirPara entrenar un modelo NN usando la optimización de descenso de gradiente estocástico, es importante alimentar los datos de entrenamiento como lotes mezclados aleatoriamente. Ya hemos visto arriba como crear lotes llamando al método `.batch()` de un objeto de conjunto de datos. Ahora, además de crear lotes, vamos a mezclar y reiterar sobre los conjuntos de datos: ###Code # Mezclando los elementos de un tensor # =================================================== tf.random.set_seed(1) ds = ds_joint.shuffle(buffer_size = len(t_x)) for example in ds: print(' x:', example[0].numpy(), ' y:', example[1].numpy()) ###Output x: [0.976 0.435 0.66 ] y: 2 x: [0.435 0.292 0.643] y: 1 x: [0.165 0.901 0.631] y: 0 x: [0.605 0.637 0.614] y: 3 ###Markdown donde las filas se barajan sin perder la correspondencia uno a uno entre las entradas en x e y. El método `.shuffle()` requiere un argumento llamado `buffer_size`, que determina cuántos elementos del conjunto de datos se agrupan antes de barajar. Los elementos del búfer se recuperan aleatoriamente y su lugar en el búfer se asigna a los siguientes elementos del conjunto de datos original (sin mezclar). Por lo tanto, si elegimos un tamaño de búfer pequeño, es posible que no mezclemos perfectamente el conjunto de datos.Si el conjunto de datos es pequeño, la elección de un tamaño de búfer relativamente pequeño puede afectar negativamente el rendimiento predictivo del NN, ya que es posible que el conjunto de datos no esté completamente aleatorizado. En la práctica, sin embargo, por lo general no tiene un efecto notable cuando se trabaja con conjuntos de datos relativamente grandes, lo cual es común en el aprendizaje profundo.Alternativamente, para asegurar una aleatorización completa durante cada época, simplemente podemos elegir un tamaño de búfer que sea igual al número de ejemplos de entrenamiento, como en el código anterior (`buffer_size = len(t_x)`). Ahora, creemos lotes a partir del conjunto de datos ds_joint: ###Code ds = ds_joint.batch(batch_size = 3, drop_remainder = False) print(ds) batch_x, batch_y = next(iter(ds)) print('Batch-x:\n', batch_x.numpy()) print('Batch-y: ', batch_y.numpy()) ###Output Batch-y: [0 1 2] ###Markdown Además, al entrenar un modelo para múltiples épocas, necesitamos mezclar e iterar sobre el conjunto de datos por el número deseado de épocas. Entonces, repitamos el conjunto de datos por lotes dos veces: ###Code ds = ds_joint.batch(3).repeat(count = 2) for i,(batch_x, batch_y) in enumerate(ds): print(i, batch_x.numpy(), batch_y.numpy(), end = '\n'*2) ###Output 0 [[0.165 0.901 0.631] [0.435 0.292 0.643] [0.976 0.435 0.66 ]] [0 1 2] 1 [[0.605 0.637 0.614]] [3] 2 [[0.165 0.901 0.631] [0.435 0.292 0.643] [0.976 0.435 0.66 ]] [0 1 2] 3 [[0.605 0.637 0.614]] [3] ###Markdown Esto da como resultado dos copias de cada lote. Si cambiamos el orden de estas dos operaciones, es decir, primero lote y luego repetimos, los resultados serán diferentes: ###Code ds = ds_joint.repeat(count=2).batch(3) for i,(batch_x, batch_y) in enumerate(ds): print(i, batch_x.numpy(), batch_y.numpy(), end = '\n'*2) ###Output 0 [[0.165 0.901 0.631] [0.435 0.292 0.643] [0.976 0.435 0.66 ]] [0 1 2] 1 [[0.605 0.637 0.614] [0.165 0.901 0.631] [0.435 0.292 0.643]] [3 0 1] 2 [[0.976 0.435 0.66 ] [0.605 0.637 0.614]] [2 3] ###Markdown Finalmente, para comprender mejor cómo se comportan estas tres operaciones (batch, shuffle y repeat), experimentemos con ellas en diferentes órdenes. Primero, combinaremos las operaciones en el siguiente orden: (1) shuffle, (2) batch y (3) repeat: ###Code # Orden 1: shuffle -> batch -> repeat tf.random.set_seed(1) ds = ds_joint.shuffle(4).batch(2).repeat(3) for i,(batch_x, batch_y) in enumerate(ds): print(i, batch_x, batch_y.numpy(), end = '\n'*2) # Orden 2: batch -> shuffle -> repeat tf.random.set_seed(1) ds = ds_joint.batch(2).shuffle(4).repeat(3) for i,(batch_x, batch_y) in enumerate(ds): print(i, batch_x, batch_y.numpy(), end = '\n'*2) # Orden 2: batch -> repeat-> shuffle tf.random.set_seed(1) ds = ds_joint.batch(2).repeat(3).shuffle(4) for i,(batch_x, batch_y) in enumerate(ds): print(i, batch_x, batch_y.numpy(), end = '\n'*2) ###Output 0 tf.Tensor( [[0.165 0.901 0.631] [0.435 0.292 0.643]], shape=(2, 3), dtype=float32) [0 1] 1 tf.Tensor( [[0.165 0.901 0.631] [0.435 0.292 0.643]], shape=(2, 3), dtype=float32) [0 1] 2 tf.Tensor( [[0.976 0.435 0.66 ] [0.605 0.637 0.614]], shape=(2, 3), dtype=float32) [2 3] 3 tf.Tensor( [[0.976 0.435 0.66 ] [0.605 0.637 0.614]], shape=(2, 3), dtype=float32) [2 3] 4 tf.Tensor( [[0.165 0.901 0.631] [0.435 0.292 0.643]], shape=(2, 3), dtype=float32) [0 1] 5 tf.Tensor( [[0.976 0.435 0.66 ] [0.605 0.637 0.614]], shape=(2, 3), dtype=float32) [2 3] ###Markdown Obteniendo conjuntos de datos disponibles de la biblioteca tensorflow_datasetsLa biblioteca tensorflow_datasets proporciona una buena colección de conjuntos de datos disponibles gratuitamente para entrenar o evaluar modelos de aprendizaje profundo. Los conjuntos de datos están bien formateados y vienen con descripciones informativas, incluido el formato de características y etiquetas y su tipo y dimensionalidad, así como la cita del documento original que introdujo el conjunto de datos en formato BibTeX. Otra ventaja es que todos estos conjuntos de datos están preparados y listos para usar como objetos tf.data.Dataset, por lo que todas las funciones que cubrimos se pueden usar directamente: ###Code # pip install tensorflow-datasets import tensorflow_datasets as tfds print(len(tfds.list_builders())) print(tfds.list_builders()[:5]) # Trabajando con el archivo mnist # =============================================== mnist, mnist_info = tfds.load('mnist', with_info=True, shuffle_files=False) print(mnist_info) print(mnist.keys()) ds_train = mnist['train'] ds_train = ds_train.map(lambda item:(item['image'], item['label'])) ds_train = ds_train.batch(10) batch = next(iter(ds_train)) print(batch[0].shape, batch[1]) import matplotlib.pyplot as plt fig = plt.figure(figsize=(15, 6)) for i,(image,label) in enumerate(zip(batch[0], batch[1])): ax = fig.add_subplot(2, 5, i+1) ax.set_xticks([]); ax.set_yticks([]) ax.imshow(image[:, :, 0], cmap='gray_r') ax.set_title('{}'.format(label), size=15) plt.show() ###Output _____no_output_____ ###Markdown Construyendo un modelo NN en TensorFlow La API de TensorFlow Keras (tf.keras)Keras es una API NN de alto nivel y se desarrolló originalmente para ejecutarse sobre otras bibliotecas como TensorFlow y Theano. Keras proporciona una interfaz de programación modular y fácil de usar que permite la creación de prototipos y la construcción de modelos complejos en solo unas pocas líneas de código. Keras se puede instalar independientemente de PyPI y luego configurarse para usar TensorFlow como su motor de backend. Keras está estrechamente integrado en TensorFlow y se puede acceder a sus módulos a través de tf.keras.En TensorFlow 2.0, tf.keras se ha convertido en el enfoque principal y recomendado para implementar modelos. Esto tiene la ventaja de que admite funcionalidades específicas de TensorFlow, como las canalizaciones de conjuntos de datos que usan tf.data.La API de Keras (tf.keras) hace que la construcción de un modelo NN sea extremadamente fácil. El enfoque más utilizado para crear una NN en TensorFlow es a través de `tf.keras.Sequential()`, que permite apilar capas para formar una red. Se puede dar una pila de capas en una lista de Python a un modelo definido como tf.keras.Sequential(). Alternativamente, las capas se pueden agregar una por una usando el método .add().Además, tf.keras nos permite definir un modelo subclasificando tf.keras.Model.Esto nos da más control sobre la propagacion hacia adelante al definir el método call() para nuestra clase modelo para especificar la propagacion hacia adelante explicitamente. Finalmente, los modelos construidos usando la API tf.keras se pueden compilar y entrenar a través de los métodos .compile() y .fit(). Construyendo un modelo de regresion lineal ###Code X_train = np.arange(10).reshape((10, 1)) y_train = np.array([1.0, 1.3, 3.1, 2.0, 5.0, 6.3, 6.6, 7.4, 8.0, 9.0]) X_train, y_train import matplotlib.pyplot as plt fig, ax = plt.subplots() ax.plot(X_train, y_train, 'o', markersize=10) ax.set_xlabel('x') ax.set_ylabel('y') import tensorflow as tf X_train_norm = (X_train - np.mean(X_train))/np.std(X_train) ds_train_orig = tf.data.Dataset.from_tensor_slices((tf.cast(X_train_norm, tf.float32),tf.cast(y_train, tf.float32))) for i in ds_train_orig: print(i[0].numpy(), i[1].numpy()) ###Output [-1.5666989] 1.0 [-1.2185436] 1.3 [-0.87038827] 3.1 [-0.52223295] 2.0 [-0.17407766] 5.0 [0.17407766] 6.3 [0.52223295] 6.6 [0.87038827] 7.4 [1.2185436] 8.0 [1.5666989] 9.0 ###Markdown Ahora, podemos definir nuestro modelo de regresión lineal como $𝑧 = 𝑤x + 𝑏$. Aquí, vamos a utilizar la API de Keras. `tf.keras` proporciona capas predefinidas para construir modelos NN complejos, pero para empezar, usaremos un modelo desde cero: ###Code class MyModel(tf.keras.Model): def __init__(self): super(MyModel, self).__init__() self.w = tf.Variable(0.0, name='weight') self.b = tf.Variable(0.0, name='bias') def call(self, x): return self.w * x + self.b model = MyModel() model.build(input_shape=(None, 1)) model.summary() ###Output Model: "my_model" _________________________________________________________________ Layer (type) Output Shape Param # ================================================================= Total params: 2 Trainable params: 2 Non-trainable params: 0 _________________________________________________________________
tsp.ipynb
###Markdown Traveling Salesman Problem We are given a [complete graph](https://en.wikipedia.org/wiki/Complete_graph) which means that every pair of distinct vertices is connected by a unique edge. Each edge has traveling cost associated with it. We need to find a shortest route that goes throught all the nodes and comes back to the original node. Representation We can represent traveling costs as a simple matrix. ###Code i = float('inf') t = [ [i, 3, 1, 8], [3, i, 4, 4], [1, 4, i, 7], [8, 4, 7, i], ] ###Output _____no_output_____ ###Markdown This represents 4 nodes (0 through 3) where cost of traveling from 0 to 3 is ###Code t[0][3] t[1][1] ###Output _____no_output_____ ###Markdown Brute force solution ###Code from itertools import permutations def total_cost(matrix, route, loop=True): # we are going from node a to node b in each step a = list(route) b = a[1:] + [a[0]] # calculate the route cost cost = 0 for ii, jj in zip(a, b): cost += matrix[ii][jj] if not loop: # exclude path back to start cost -= matrix[route[0]][route[-1]] return cost def brute_force(matrix, start=0, stop=None, loop=True): if not stop: stop = len(matrix) current_min = float('inf') current_route = [] for route in permutations(range(start, stop)): route_cost = total_cost(matrix, route, loop) # check if we got new min route if route_cost < current_min: current_min = route_cost current_route = route return list(current_route), current_min ###Output _____no_output_____ ###Markdown Greedy solutionWhat if we just always pick the shortest distance from all available routes in a current node. We start with first node and pick the cheapest node that we haven't picked yet. ###Code def get_min_node(nodes, visited_set): min_cost = pow(10, 10) node_position = None for position, cost in enumerate(nodes): # make sure we haven't visited the node yet, we are not trying to go to # the same node, and current cost is the minimum that we've seen so far if position in visited_set or isinstance(cost, str) or cost >= min_cost: continue min_cost = cost node_position = position return node_position def greedy_recursion(matrix, route): current_node = route[-1] min_node = get_min_node(matrix[current_node], set(route)) if min_node is None: return route return greedy_recursion(matrix, route + [min_node]) def greedy(matrix, route=None): """ It doesn't return an optimal route but its close to brute force answer. The benifit of the algorithm is relative speed and simplicity. """ result = greedy_recursion(matrix, route or [0]) return result, total_cost(m, result) ###Output _____no_output_____ ###Markdown Branch and Bound Implementation was inspired by [this](http://galyautdinov.ru/post/zadacha-kommivoyazhera) post. ###Code from random import randint def random_matrix(n, max_distance=10): return [[ii == jj and float('inf') or randint(1, max_distance) for ii in range(n)] for jj in range(n)] def copy_matrix(m): return [[ii for ii in jj] for jj in m] def min_in(matrix, direction='row'): vector = [] d = range(len(matrix)) for ii in d: min_element = float('inf') for jj in d: # switch between row and column access item = direction == 'row' and matrix[ii][jj] or matrix[jj][ii] if min_element > item: min_element = item vector.append(min_element) return vector def subtract(matrix, vector, direction='row'): d = range(len(matrix)) for ii in d: if vector[ii] == float('inf'): continue for jj in d: if direction == 'row': matrix[ii][jj] -= vector[ii] else: matrix[jj][ii] -= vector[ii] return matrix def reduse(matrix): v1 = min_in(matrix) redused_rows = subtract(matrix, v1) v2 = min_in(redused_rows, 'column') return subtract(redused_rows, v2, 'column') def find_optimal_segment(matrix): size = range(len(matrix)) max_value = -1 max_ii = max_jj = -1 for ii in size: for jj in size: if matrix[ii][jj] != 0: continue # calculate max value for cells that have zeros min_in_row = min([e for pos, e in enumerate(matrix[ii]) if pos != jj]) min_in_column = min([e[jj] for pos, e in enumerate(matrix) if pos != ii]) current_max = min_in_row + min_in_column if max_value < current_max: max_value = current_max max_ii = ii max_jj = jj # close the route back matrix[max_jj][max_ii] = float('inf') # "remove" row and column for jj in size: matrix[max_ii][jj] = float('inf') for ii in size: matrix[ii][max_jj] = float('inf') return matrix, max_ii, max_jj def branch_and_bound(matrix): # reduse modifies matrix in place # so we copy the original in order # to not mess it up m = copy_matrix(matrix) # initialize variables path_dict = {} distance = 0 size = range(len(matrix)) # find optimal pairs for i in size: redused = reduse(m) m, a, b = find_optimal_segment(redused) path_dict[a] = b distance += matrix[a][b] # arrange segments in walking order path = [] for i in size: # start with 0 if it is a new path else reassign to second element a = path and b or i path.append(a) b = path_dict[a] return path, distance ###Output _____no_output_____ ###Markdown Comparing algorithms ###Code m = random_matrix(5) m branch_and_bound(m) brute_force(m) greedy(m) ###Output _____no_output_____ ###Markdown Look ahead ###Code test = random_matrix(20) def sub_route_cost(matrix, route, loop=True): # we are going from node a to node b in each step a = list(route) b = a[1:] + [a[0]] # calculate the route cost cost = 0 for ii, jj in zip(a, b): cost += matrix[ii][jj] if not loop: # exclude path back to start cost -= matrix[route[0]][route[-1]] return cost def look_ahead(matrix, k): size = len(matrix) if k >= size: # in this case it should produce similar # result to brute force k = size -1 path = [] nodes = list(range(len(matrix))) # start with the first node path.append(nodes.pop(0)) while nodes: current_min = float('inf') min_route = [] for choice in permutations(nodes, k): current_route = [path[-1]] + list(choice) route_cost = sub_route_cost(matrix, current_route, loop=False) if route_cost < current_min: current_min = route_cost min_route = list(choice) #print('route', min_route, 'cost', current_min) path += min_route for ii in min_route: nodes.remove(ii) if len(nodes) == 1: # Nowhere to go from here count the last step path.append(nodes.pop(0)) #print('last step') return path, sub_route_cost(matrix, path) t2 look_ahead(t2, 2) brute_force(t2) look_ahead(t2, 4) look_ahead(t2, 3) ###Output _____no_output_____ ###Markdown Vecino más Cercano para el TSP ###Code def nearest_neighbor(Edges, weights): """ Edges: Arreglo de aristas del grafo weights: distancias de cada arista """ # Encontrar la arista de menor peso para iniciar el recorrido next_edge = np.argmin(weights) T = list(Edges[next_edge, :]) ordered_el = [Edges[next_edge]] S = weights[next_edge] E = np.delete(Edges, next_edge, axis=0) d = np.delete(weights, next_edge, axis=0) while len(E) > 0: # Encontrar minima distancia conectada a un extremo del recorrido next_edge = np.argmin(np.where(np.any(np.isin(E, [T[0], T[-1]]), axis=1), d, np.inf)) # Revisar si la arista va al principio o al final del recorrido a, b = E[next_edge] if a == T[0]: T.insert(0, b) elif b == T[0]: T.insert(0, a) b, a = a, b elif a == T[-1]: T.append(b) else: T.append(a) b, a = a, b ordered_el.append(E[next_edge]) S += d[next_edge] # Descartar aristas invalidas para el problema mask = ~np.logical_or(np.any(E == a, axis=1), np.all(np.isin(E, T), axis=1)) E = E[mask] d = d[mask] # Cerrar el recorrido a, b = np.sort([T[0], T[-1]]) ordered_el.append([a, b]) for edg, dist in zip(Edges, weights): if edg[0] == a and edg[1] == b: S += dist return T, S, ordered_el def edges_from_tour(T): return [(x, y) for x, y in zip(T, T[1:])] + [(T[-1], T[0])] ###Output _____no_output_____ ###Markdown Solución para n particular ###Code n = 10 C = np.loadtxt("data/datos_unicos.txt", max_rows=n) E = np.array(list(combinations(range(n), 2))) d = np.linalg.norm(C[E[:, 0], :] - C[E[:, 1], :], axis=1) t = time.perf_counter() T, S, ordered_el = nearest_neighbor(E, d) t = time.perf_counter() - t print("Numero de ciudades:\t{}\nCosto:\t\t\t{}\nTiempo:\t\t\t{} s".format(n, S, t)) G = nx.Graph() G.add_weighted_edges_from(zip(E[:, 0], E[:, 1], d)) locs = dict(zip(range(n), C[:n, :])) edgs = edges_from_tour(T) nx.draw_networkx(G, locs, edgelist=edgs, node_size=200) plt.show() fig, ax = plt.subplots(figsize=(10, 10)) nx.draw_networkx(G, locs, edgelist=[], node_size=200, ax=ax) def anim_tour(i): nx.draw_networkx_edges(G, locs, [ordered_el[i]], ax=ax) #ax.set_aspect('equal') anim = animation.FuncAnimation(fig, anim_tour, range(len(edgs)), interval=1000) plt.show() ###Output _____no_output_____ ###Markdown Comparación contra solución exacta ###Code # N = [5, 10, 20, 30, 40, 50, 75, 100, 125, 150, 200, 250, 300, 400, 500, 600, 634] # bench_NN = np.zeros([len(N), 3]) # for n, row in zip(N, bench_NN): # C = np.loadtxt("data/datos_unicos.txt", max_rows=n) # E = np.array(list(combinations(range(n), 2))) # d = np.linalg.norm(C[E[:, 0], :] - C[E[:, 1], :], axis=1) # t = time.perf_counter() # _, S, oredered_el = nearest_neighbor(E, d) # t = time.perf_counter() - t # row[:] =[n, t, S] # np.savetxt("data/benchmark_NN.txt", bench_NN) bench_NN = np.loadtxt("data/benchmark_NN.txt") bench_dantzig = np.loadtxt("data/benchmark_dantzig.txt") plt.plot(bench_dantzig[:, 0], bench_dantzig[:, 1]) plt.axis([5, 300, 0, 400]) plt.ylabel("Tiempo [s]") plt.xlabel("Número de Ciudades") plt.title("Tiempo de Ejecución con PLE") plt.savefig("img/res_d.png", dpi=300) plt.show() plt.plot(bench_NN[:, 0], bench_NN[:, 1]) plt.axis([5, 634, 0, 10]) plt.ylabel("Tiempo [s]") plt.xlabel("Número de Ciudades") plt.title("Tiempo de Ejecución con NN") plt.savefig("img/res_nn.png", dpi=300) plt.show() plt.plot(bench_dantzig[:, 0], bench_dantzig[:, 1] - bench_NN[:13, 1]) plt.axis([5, 300, 0, 400]) plt.ylabel("Tiempo [s]") plt.xlabel("Número de Ciudades") plt.title("Diferencia en Tiempo de Ejecución") plt.savefig("img/diff_tiempo.png", dpi=300) plt.show() plt.plot(bench_dantzig[:, 0], bench_dantzig[:, 2], label="PLE") plt.plot(bench_NN[:13, 0], bench_NN[:13, 2], label="Vecino más Cercano") plt.axis([5, 300, 0, 7000]) plt.ylabel("Costo") plt.xlabel("Número de Ciudades") plt.title("Costo de Solución") plt.legend() plt.savefig("img/cto_d_nn.png", dpi=300) plt.show() ###Output _____no_output_____
notebooks/ipca.ipynb
###Markdown Iterated PCA (IPCA) ###Code import os, sys sys.path.append(os.path.abspath('../src')) import matplotlib.pyplot as plt import numpy as np import pandas as pd import sklearn.decomposition import xpca first_year = 2008; last_year = 2020 df = pd.read_parquet('../data/equity_indices.parquet') df = df[str(first_year):str(last_year)] df.head() df.tail() model = sklearn.decomposition.PCA() model.fit(df) Z = model.transform(df) plt.plot(Z[:,0]); Z_periods_classical = [] for period in [str(x) for x in range(first_year, last_year+1)]: df_period = df.loc[period] model = sklearn.decomposition.PCA() model.fit(df_period) Z_period = model.transform(df_period) Z_periods_classical.append(Z_period) Z_periods_classical = np.vstack(Z_periods_classical) fig = plt.figure(figsize=(12, 12)) for i in range(9): ax = fig.add_subplot(3, 3, i+1) ax.plot(Z[:,i], Z_periods_classical[:,i], '.') ax.set_title(f'PC{i+1}') ax.axis('equal') ax.tick_params(axis='both', which='major', labelsize=8) ax.tick_params(axis='both', which='minor', labelsize=8) ax.tick_params(axis='x', rotation=45) fig.tight_layout(); Z_periods_ipca = [] model = xpca.IPCA() for period in [str(x) for x in range(first_year, last_year+1)]: df_period = df.loc[period] model.fit(df_period) Z_period = model.transform(df_period) Z_periods_ipca.append(Z_period) Z_periods_ipca = np.vstack(Z_periods_ipca) fig = plt.figure(figsize=(12, 12)) for i in range(9): ax = fig.add_subplot(3, 3, i+1) ax.plot(Z[:,i], Z_periods_ipca[:,i], '.') ax.set_title(f'PC{i+1}') ax.axis('equal') ax.tick_params(axis='both', which='major', labelsize=8) ax.tick_params(axis='both', which='minor', labelsize=8) ax.tick_params(axis='x', rotation=45) fig.tight_layout(); ###Output _____no_output_____
Python/7_sentiment_analysis/Sentiment Analysis - Economic News Sentiment with BERT Fine-Tuning.ipynb
###Markdown Predicting Economic News Sentiment with BERT on TF Hub If you’ve been following Natural Language Processing over the past year, you’ve probably heard of BERT: Bidirectional Encoder Representations from Transformers. It’s a neural network architecture designed by Google researchers that’s totally transformed what’s state-of-the-art for NLP tasks, like text classification, translation, summarization, and question answering.Now that BERT's been added to [TF Hub](https://www.tensorflow.org/hub) as a loadable module, it's easy(ish) to add into existing Tensorflow text pipelines. In an existing pipeline, BERT can replace text embedding layers like ELMO and GloVE. Alternatively, [finetuning](http://wiki.fast.ai/index.php/Fine_tuning) BERT can provide both an accuracy boost and faster training time in many cases.Here, we'll train a model to predict whether a piece of economic news is positive or negative using BERT in Tensorflow with tf hub. Some code was adapted from [this colab notebook](https://colab.sandbox.google.com/github/tensorflow/tpu/blob/master/tools/colab/bert_finetuning_with_cloud_tpus.ipynb). Let's get started! ###Code !pip install --upgrade tensorflow !pip uninstall protobuf -y !pip install protobuf -y !pip install tensorflow_hub !pip install bert-tensorflow from sklearn.model_selection import train_test_split import pandas as pd import tensorflow as tf import tensorflow_hub as hub from datetime import datetime ###Output /usr/local/anaconda/lib/python3.6/site-packages/h5py/__init__.py:34: FutureWarning: Conversion of the second argument of issubdtype from `float` to `np.floating` is deprecated. In future, it will be treated as `np.float64 == np.dtype(float).type`. from ._conv import register_converters as _register_converters WARNING: Logging before flag parsing goes to stderr. W0305 15:30:17.590888 140018860873536 __init__.py:56] Some hub symbols are not available because TensorFlow version is less than 1.14 ###Markdown In addition to the standard libraries we imported above, we'll need to install BERT's python package. ###Code import bert from bert import run_classifier from bert import optimization from bert import tokenization ###Output _____no_output_____ ###Markdown Below, we'll set an output directory location to store our model output and checkpoints. This can be a local directory, in which case you'd set OUTPUT_DIR to the name of the directory you'd like to create. If you're running this code in Google's hosted Colab, the directory won't persist after the Colab session ends.Alternatively, if you're a GCP user, you can store output in a GCP bucket. To do that, set a directory name in OUTPUT_DIR and the name of the GCP bucket in the BUCKET field.Set DO_DELETE to rewrite the OUTPUT_DIR if it exists. Otherwise, Tensorflow will load existing model checkpoints from that directory (if they exist). ###Code # Set the output directory for saving model file # Optionally, set a GCP bucket location OUTPUT_DIR = 'temp_out'#@param {type:"string"} #@markdown Whether or not to clear/delete the directory and create a new one DO_DELETE = True #@param {type:"boolean"} #@markdown Set USE_BUCKET and BUCKET if you want to (optionally) store model output on GCP bucket. USE_BUCKET = False #@param {type:"boolean"} BUCKET = 'economic_news_sentiment' #@param {type:"string"} if USE_BUCKET: OUTPUT_DIR = 'gs://{}/{}'.format(BUCKET, OUTPUT_DIR) from google.colab import auth auth.authenticate_user() if DO_DELETE: try: tf.gfile.DeleteRecursively(OUTPUT_DIR) except: # Doesn't matter if the directory didn't exist pass tf.gfile.MakeDirs(OUTPUT_DIR) print('***** Model output directory: {} *****'.format(OUTPUT_DIR)) ###Output ***** Model output directory: temp_out ***** ###Markdown Data First, let's download the dataset, hosted by Stanford. The code below, which downloads, extracts, and imports the IMDB Large Movie Review Dataset, is borrowed from [this Tensorflow tutorial](https://www.tensorflow.org/hub/tutorials/text_classification_with_tf_hub). ###Code # from google.colab import drive # drive.mount('/content/gdrive') import os os.getcwd() from tensorflow import keras import os import re # Load all files from a directory in a DataFrame. def load_directory_data(directory): # data = load_directory_data(os.path.join(directory, "economic_sentiment_data.csv")) data = pd.read_csv(os.path.join(directory, "economic_sentiment_data.csv")) data = data[['sentence','sentiment','polarity']] print(data.shape) return data # # Merge positive and negative examples, add a polarity column and shuffle. # def load_dataset(directory): # data_df = load_directory_data(os.path.join(directory, "economic_sentiment_data.csv")) # return pd.concat([pos_df, neg_df]).sample(frac=1).reset_index(drop=True) # Download and process the dataset files. def download_and_load_datasets(force_download=False): # dataset = tf.keras.utils.get_file( # fname="Full-Economic-News-DFE-839861.csv", # origin="https://d1p17r2m4rzlbo.cloudfront.net/wp-content/uploads/2016/03/Full-Economic-News-DFE-839861.csv", # extract=False) # print(os.path.dirname(dataset)) full_data_df = load_directory_data(os.path.join('../../data/','raw')) train_df = full_data_df.iloc[0:3000] test_df = full_data_df.iloc[3000:] print(train_df.shape) print(test_df.shape) return train_df, test_df ###Output _____no_output_____ ###Markdown To keep training fast, we'll take a sample of 5000 train and test examples, respectively. ###Code train, test = download_and_load_datasets() # train = train.sample(5000) # test = test.sample(5000) train.columns ###Output _____no_output_____ ###Markdown For us, our input data is the 'sentence' column and our label is the 'polarity' column (0, 1 for negative and positive, respecitvely) ###Code DATA_COLUMN = 'sentence' LABEL_COLUMN = 'polarity' # label_list is the list of labels, i.e. True, False or 0, 1 or 'dog', 'cat' label_list = [0, 1] ###Output _____no_output_____ ###Markdown Data PreprocessingWe'll need to transform our data into a format BERT understands. This involves two steps. First, we create `InputExample`'s using the constructor provided in the BERT library.- `text_a` is the text we want to classify, which in this case, is the `Request` field in our Dataframe. - `text_b` is used if we're training a model to understand the relationship between sentences (i.e. is `text_b` a translation of `text_a`? Is `text_b` an answer to the question asked by `text_a`?). This doesn't apply to our task, so we can leave `text_b` blank.- `label` is the label for our example, i.e. True, False ###Code # Use the InputExample class from BERT's run_classifier code to create examples from the data train_InputExamples = train.apply(lambda x: bert.run_classifier.InputExample(guid=None, # Globally unique ID for bookkeeping, unused in this example text_a = x[DATA_COLUMN], text_b = None, label = x[LABEL_COLUMN]), axis = 1) test_InputExamples = test.apply(lambda x: bert.run_classifier.InputExample(guid=None, text_a = x[DATA_COLUMN], text_b = None, label = x[LABEL_COLUMN]), axis = 1) ###Output _____no_output_____ ###Markdown Next, we need to preprocess our data so that it matches the data BERT was trained on. For this, we'll need to do a couple of things (but don't worry--this is also included in the Python library):1. Lowercase our text (if we're using a BERT lowercase model)2. Tokenize it (i.e. "sally says hi" -> ["sally", "says", "hi"])3. Break words into WordPieces (i.e. "calling" -> ["call", "ing"])4. Map our words to indexes using a vocab file that BERT provides5. Add special "CLS" and "SEP" tokens (see the [readme](https://github.com/google-research/bert))6. Append "index" and "segment" tokens to each input (see the [BERT paper](https://arxiv.org/pdf/1810.04805.pdf))Happily, we don't have to worry about most of these details. To start, we'll need to load a vocabulary file and lowercasing information directly from the BERT tf hub module: ###Code # This is a path to an uncased (all lowercase) version of BERT BERT_MODEL_HUB = "https://tfhub.dev/google/bert_uncased_L-12_H-768_A-12/1" def create_tokenizer_from_hub_module(): """Get the vocab file and casing info from the Hub module.""" with tf.Graph().as_default(): bert_module = hub.Module(BERT_MODEL_HUB) tokenization_info = bert_module(signature="tokenization_info", as_dict=True) with tf.Session() as sess: vocab_file, do_lower_case = sess.run([tokenization_info["vocab_file"], tokenization_info["do_lower_case"]]) return bert.tokenization.FullTokenizer( vocab_file=vocab_file, do_lower_case=do_lower_case) tokenizer = create_tokenizer_from_hub_module() ###Output WARNING:tensorflow:From /usr/local/anaconda/lib/python3.6/site-packages/tensorflow/python/ops/control_flow_ops.py:3632: colocate_with (from tensorflow.python.framework.ops) is deprecated and will be removed in a future version. Instructions for updating: Colocations handled automatically by placer. ###Markdown Great--we just learned that the BERT model we're using expects lowercase data (that's what stored in tokenization_info["do_lower_case"]) and we also loaded BERT's vocab file. We also created a tokenizer, which breaks words into word pieces: ###Code tokenizer.tokenize("This here's an example of using the BERT tokenizer") #tokenizer.tokenize(pred_sentences[0]) ###Output _____no_output_____ ###Markdown Using our tokenizer, we'll call `run_classifier.convert_examples_to_features` on our InputExamples to convert them into features BERT understands. ###Code # We'll set sequences to be at most 128 tokens long. MAX_SEQ_LENGTH = 128 # Convert our train and test features to InputFeatures that BERT understands. train_features = bert.run_classifier.convert_examples_to_features(train_InputExamples, label_list, MAX_SEQ_LENGTH, tokenizer) test_features = bert.run_classifier.convert_examples_to_features(test_InputExamples, label_list, MAX_SEQ_LENGTH, tokenizer) ###Output INFO:tensorflow:Writing example 0 of 3000 ###Markdown Creating a modelNow that we've prepared our data, let's focus on building a model. `create_model` does just this below. First, it loads the BERT tf hub module again (this time to extract the computation graph). Next, it creates a single new layer that will be trained to adapt BERT to our sentiment task (i.e. classifying whether a movie review is positive or negative). This strategy of using a mostly trained model is called [fine-tuning](http://wiki.fast.ai/index.php/Fine_tuning). ###Code # model_fn_builder actually creates our model function # using the passed parameters for num_labels, learning_rate, etc. def model_fn_builder(num_labels, learning_rate, num_train_steps, num_warmup_steps): """Returns `model_fn` closure for TPUEstimator.""" def model_fn(features, labels, mode, params): # pylint: disable=unused-argument """The `model_fn` for TPUEstimator.""" input_ids = features["input_ids"] input_mask = features["input_mask"] segment_ids = features["segment_ids"] label_ids = features["label_ids"] is_predicting = (mode == tf.estimator.ModeKeys.PREDICT) # TRAIN and EVAL if not is_predicting: (loss, predicted_labels, log_probs) = create_model( is_predicting, input_ids, input_mask, segment_ids, label_ids, num_labels) train_op = bert.optimization.create_optimizer( loss, learning_rate, num_train_steps, num_warmup_steps, use_tpu=False) # Calculate evaluation metrics. def metric_fn(label_ids, predicted_labels): accuracy = tf.metrics.accuracy(label_ids, predicted_labels) f1_score = tf.contrib.metrics.f1_score( label_ids, predicted_labels) auc = tf.metrics.auc( label_ids, predicted_labels) recall = tf.metrics.recall( label_ids, predicted_labels) precision = tf.metrics.precision( label_ids, predicted_labels) true_pos = tf.metrics.true_positives( label_ids, predicted_labels) true_neg = tf.metrics.true_negatives( label_ids, predicted_labels) false_pos = tf.metrics.false_positives( label_ids, predicted_labels) false_neg = tf.metrics.false_negatives( label_ids, predicted_labels) return { "eval_accuracy": accuracy, "f1_score": f1_score, "auc": auc, "precision": precision, "recall": recall, "true_positives": true_pos, "true_negatives": true_neg, "false_positives": false_pos, "false_negatives": false_neg } eval_metrics = metric_fn(label_ids, predicted_labels) if mode == tf.estimator.ModeKeys.TRAIN: return tf.estimator.EstimatorSpec(mode=mode, loss=loss, train_op=train_op) else: return tf.estimator.EstimatorSpec(mode=mode, loss=loss, eval_metric_ops=eval_metrics) else: (predicted_labels, log_probs) = create_model( is_predicting, input_ids, input_mask, segment_ids, label_ids, num_labels) predictions = { 'probabilities': log_probs, 'labels': predicted_labels } return tf.estimator.EstimatorSpec(mode, predictions=predictions) # Return the actual model function in the closure return model_fn def create_model(is_predicting, input_ids, input_mask, segment_ids, labels, num_labels): """Creates a classification model.""" bert_module = hub.Module( BERT_MODEL_HUB, trainable=True) bert_inputs = dict( input_ids=input_ids, input_mask=input_mask, segment_ids=segment_ids) bert_outputs = bert_module( inputs=bert_inputs, signature="tokens", as_dict=True) # Use "pooled_output" for classification tasks on an entire sentence. # Use "sequence_outputs" for token-level output. output_layer = bert_outputs["pooled_output"] hidden_size = output_layer.shape[-1].value # Create our own layer to tune for politeness data. output_weights = tf.get_variable( "output_weights", [num_labels, hidden_size], initializer=tf.truncated_normal_initializer(stddev=0.02)) output_bias = tf.get_variable( "output_bias", [num_labels], initializer=tf.zeros_initializer()) with tf.variable_scope("loss"): # Dropout helps prevent overfitting output_layer = tf.nn.dropout(output_layer, keep_prob=0.9) logits = tf.matmul(output_layer, output_weights, transpose_b=True) logits = tf.nn.bias_add(logits, output_bias) log_probs = tf.nn.log_softmax(logits, axis=-1) # Convert labels into one-hot encoding one_hot_labels = tf.one_hot(labels, depth=num_labels, dtype=tf.float32) predicted_labels = tf.squeeze(tf.argmax(log_probs, axis=-1, output_type=tf.int32)) # If we're predicting, we want predicted labels and the probabiltiies. if is_predicting: return (predicted_labels, log_probs) # If we're train/eval, compute loss between predicted and actual label per_example_loss = -tf.reduce_sum(one_hot_labels * log_probs, axis=-1) loss = tf.reduce_mean(per_example_loss) return (loss, predicted_labels, log_probs) ###Output _____no_output_____ ###Markdown Next we'll wrap our model function in a `model_fn_builder` function that adapts our model to work for training, evaluation, and prediction. ###Code # Compute train and warmup steps from batch size # These hyperparameters are copied from this colab notebook (https://colab.sandbox.google.com/github/tensorflow/tpu/blob/master/tools/colab/bert_finetuning_with_cloud_tpus.ipynb) BATCH_SIZE = 32 LEARNING_RATE = 2e-5 NUM_TRAIN_EPOCHS = 6 # Warmup is a period of time where the learning rate # is small and gradually increases--usually helps training. WARMUP_PROPORTION = 0.1 # Model configs SAVE_CHECKPOINTS_STEPS = 500 SAVE_SUMMARY_STEPS = 100 # Compute # train and warmup steps from batch size num_train_steps = int(len(train_features) / BATCH_SIZE * NUM_TRAIN_EPOCHS) num_warmup_steps = int(num_train_steps * WARMUP_PROPORTION) # Specify output directory and number of checkpoint steps to save run_config = tf.estimator.RunConfig( model_dir=OUTPUT_DIR, save_summary_steps=SAVE_SUMMARY_STEPS, save_checkpoints_steps=SAVE_CHECKPOINTS_STEPS) model_fn = model_fn_builder( num_labels=len(label_list), learning_rate=LEARNING_RATE, num_train_steps=num_train_steps, num_warmup_steps=num_warmup_steps) estimator = tf.estimator.Estimator( model_fn=model_fn, config=run_config, params={"batch_size": BATCH_SIZE}) ###Output INFO:tensorflow:Using config: {'_model_dir': 'temp_out', '_tf_random_seed': None, '_save_summary_steps': 100, '_save_checkpoints_steps': 500, '_save_checkpoints_secs': None, '_session_config': allow_soft_placement: true graph_options { rewrite_options { meta_optimizer_iterations: ONE } } , '_keep_checkpoint_max': 5, '_keep_checkpoint_every_n_hours': 10000, '_log_step_count_steps': 100, '_train_distribute': None, '_device_fn': None, '_protocol': None, '_eval_distribute': None, '_experimental_distribute': None, '_service': None, '_cluster_spec': <tensorflow.python.training.server_lib.ClusterSpec object at 0x7f585444aac8>, '_task_type': 'worker', '_task_id': 0, '_global_id_in_cluster': 0, '_master': '', '_evaluation_master': '', '_is_chief': True, '_num_ps_replicas': 0, '_num_worker_replicas': 1} ###Markdown Next we create an input builder function that takes our training feature set (`train_features`) and produces a generator. This is a pretty standard design pattern for working with Tensorflow [Estimators](https://www.tensorflow.org/guide/estimators). ###Code # Create an input function for training. drop_remainder = True for using TPUs. train_input_fn = bert.run_classifier.input_fn_builder( features=train_features, seq_length=MAX_SEQ_LENGTH, is_training=True, drop_remainder=False) ###Output _____no_output_____ ###Markdown Now we train our model! For me, using a Colab notebook running on Google's GPUs, my training time was about 14 minutes. ###Code !pip install dask --upgrade print(f'Beginning Training!') current_time = datetime.now() estimator.train(input_fn=train_input_fn, max_steps=num_train_steps) print("Training took time ", datetime.now() - current_time) ###Output Beginning Training! INFO:tensorflow:Calling model_fn. ###Markdown Now let's use our test data to see how well our model did: ###Code test_input_fn = run_classifier.input_fn_builder( features=test_features, seq_length=MAX_SEQ_LENGTH, is_training=False, drop_remainder=True) estimator.evaluate(input_fn=test_input_fn, steps=None) ###Output INFO:tensorflow:Calling model_fn. ###Markdown Now let's write code to make predictions on new sentences: ###Code def getPrediction(in_sentences): labels = ["Negative", "Positive"] input_examples = [run_classifier.InputExample(guid="", text_a = x, text_b = None, label = 0) for x in in_sentences] # here, "" is just a dummy label input_features = run_classifier.convert_examples_to_features(input_examples, label_list, MAX_SEQ_LENGTH, tokenizer) predict_input_fn = run_classifier.input_fn_builder(features=input_features, seq_length=MAX_SEQ_LENGTH, is_training=False, drop_remainder=False) predictions = estimator.predict(predict_input_fn) return [(sentence, prediction['probabilities'], labels[prediction['labels']]) for sentence, prediction in zip(in_sentences, predictions)] pred_sentences = [ '''While the RMB in 2017 was broadly in line with economic fundamentals and desirable policies, the current account surplus was moderately stronger. This reflects structural distortions and policies that cause excessive savings, such as low social spending. Addressing these distortions and the resulting external imbalance would benefit both China and the global economy.''', '''Favorable domestic and external conditions reduced capital outflows and exchange rate pressure. The RMB was broadly stable against the basket published by the China Foreign Exchange Trade System (CFETS) in 2017, but with more fluctuation versus the dollar, and it has appreciated by about 2 percent in real effective terms in the first half of 2018. The current account surplus continued to decline but, reflecting distortions and policy gaps that encourage excessive savings, the external position for 2017 is assessed as moderately stronger than the level consistent with medium-term fundamentals and desirable policies, with the exchange rate broadly in line(Appendix I).''', '''Large outflows and pressure on the exchange rate could resume due to tighter and more volatile global financial conditions, especially a surging dollar. Investor sentiment towards emerging markets has recently weakened, and this could intensify, potentially spreading to China.''', '''. Uncoordinated financial and local government regulatory action could have unintended consequences that trigger disorderly repricing of corporate/LGFV credit risks, losses for investors, and rollover risks for financial institutions''', '''But a lack of decisive reforms in deleveraging and rebalancing would add to the Faster reform progress could pave the way for higher and more sustainable GDP growth, already-high stock of vulnerabilities and worsen resource allocation, leading to more rapidly diminishing returns over the medium term. This scenario also raises the probability of a disruptive adjustment to Chinese demand which would result in a contractionary impulse to the global economy, as well as spillovers through commodity prices and financial markets. ''' ] predictions = getPrediction(pred_sentences) ###Output INFO:tensorflow:Writing example 0 of 5 ###Markdown Voila! We have a sentiment classifier! ###Code predictions ###Output _____no_output_____
ARC-seismic demo.ipynb
###Markdown ARC-seismic | demo Anisotropic Reflection Coefficient seismic modelling software.*** 1. IntroductionARC-seismic uses the Zoeppritz equations to compute exact reflection coefficients for anisotropic elastic media across all angles of incidence (0$^{\circ}-$90$^{\circ}$), including post-critical angles. It consists of 3 individual programs:1. **_zoeppritz.py_** - Comprehensive code for modelling reflection and transmission coefficients for isotropic and anisotropic elastic media in three dimensions.2. **_montecarlo.py_** - A code that uses zoeppritz.py to conduct Monte Carlo simulations of reflection coefficients.3. **_kirchhoff.py_** - Kirchhoff program that uses zoeppritz.py to generate synthetic seismograms of the reflected finite frequency wave-field. The zoeppritz.py code has been validated using exact plane-wave reflection coefficient solutions, and the only condition it requires is that both media possess at least **_monoclinic symmetry_**, defined by an elastic tensor of the form:$$\mathbf{C}_{ij} = \begin{bmatrix}C_{11} & C_{12} & C_{13} & & & C_{16}\\ C_{12} & C_{22} & C_{23} & & & C_{26}\\ C_{13} & C_{23} & C_{33} & & & C_{36}\\ & & & C_{44} & C_{45} & \\ & & & C_{45} & C_{55} & \\ C_{16} & C_{26} & C_{36} & & & C_{66}\end{bmatrix}$$*** 2. Background & MotivationARC-seismic was initially developed to use the anisotropic Zoeppritz equations to investigate a seismic data set, the Buzzard Field, where acoustic Full-Waveform Inversion (FWI) fails (see Figure 1). I was initially surprised when I couldn't find a program online that accurately solved the anisotropic Zoeppritz equations for pre- and post-critical incidence angles. It turns out that developing an accurate and robust anisotropic Zoeppritz modelling program is a challenging task$^{\textbf{(2)}}$, requiring the solution to multiple eigenvalue problems and the correct handling of complex numbers. Using ARC-seismic, it was discovered that the failure of acoustic FWI on the Buzzard Field was a consequence of post-critical elastic effects originating from an unconformable top-chalk interface, which resulted in almost zero post-critical energy being reflected. It was also discovered that post-critical reflection coefficients are highly sensitive to weak/moderate anisotropy within the reflecting medium of an interface.For a detailed case study example, please refer to my thesis: *'The Limitations of Acoustic Full-Waveform Inversion in an Elastic World'*. Figure 1. A) Starting reverse time migrated p-wave velocity model. The high chalk velocities have been verified using common-image gathers. B) Acoustic FWI recovered p-wave velocity model. Acoustic FWI clearly fails to recover the high p-wave velocities of the chalk in the starting model.*** 3. TheoryThe Zoeppritz equations$^{\textbf{(1)}}$ govern the proportion of seismic energy that is reflected and transmitted from an interface between two homogenous elastic media (see Figure 2). They are formulated by assuming the solution to the wave equation: $$\left ( \rho\ \delta_{ij}\ \frac{\partial^2 }{\partial t^2}\ -\ \mathbf{C}_{ikjm}\ \frac{\partial^2 }{\partial x_{k}\partial x_{m}} \right )u_{j}=0,\ \ \ i=1,2,3,$$takes the form of an elastic plane wave:$$u_{i} = q_{i}\ exp\left [ i\omega\left ( t\ -\ s_{k}\ x_{k} \right ) \right ],\ \ \ i=1,2,3,$$where $\rho$ is the density, $\mathbf{\delta}_{ij}$ is the Kronecker delta function, $\mathbf{C}_{ikjm}$ is the fourth order elastic tensor, $u_{ij}$ is the displacement, $q_{ij}$ is the polarization vector, $s_{km}$ is the slowness vector and $\omega$ is the angular frequency. This yields the following equation: $$\rho\ q_{i} = \mathbf{C}_{ikjm}\ s_{k}\ s_{m}\ q_{j},\ \ \ i=1,2,3,$$$\mathbf{C}_{ikjm}$ governs the three-dimensional relationship between stress and strain in a medium. It was condensed into a more convenient second order 6$\times$6 matrix, $\mathbf{C}_{ij}$, by exploiting its symmetrical properties. The previous equation can then be rearranged and expressed as:$$\left ( \mathbf{\Gamma}-\rho \mathbf{I}\right )\mathbf{q} = 0$$where $\mathbf{\Gamma}$ is the 3$\times$3 Christoffel matrix, defined for media with monoclinic symmetry as follows:$$\begin{matrix} \Gamma_{11}=C_{11}s_{1}^{2}+C_{66}s_{2}^{2}+C_{55}s_{3}^{2}+2C_{16}s_{1}s_{2}\\ \Gamma_{22}=C_{66}s_{1}^{2}+C_{22}s_{2}^{2}+C_{44}s_{3}^{2}+2C_{26}s_{1}s_{2}\\ \Gamma_{33}=C_{55}s_{1}^{2}+C_{44}s_{2}^{2}+C_{33}s_{3}^{2}+2C_{45}s_{1}s_{2}\\ \Gamma_{12}=\Gamma_{21}=C_{16}s_{1}^{2}+C_{26}s_{2}^{2}+C_{45}s_{3}^{2}+(C_{12}+C_{66})s_{1}s_{2}\\ \Gamma_{13}=\Gamma_{31}=(C_{13}+C_{55})s_{1}s_{3}+(C_{36}+C_{45})s_{2}s_{3}\\ \Gamma_{23}=\Gamma_{32}=(C_{36}+C_{45})s_{1}s_{3}+(C_{23}+C_{44})s_{2}s_{3}.\end{matrix}$$Explicit solutions for the reflection and transmission coefficients of incident P and S-waves were then found by imposing continuity of displacement and traction across the interface$^{\textbf{(1)}}$. Figure 2. Diagram of the system defined by the anisotropic Zoeppritz equations. $Pi$ is the incident p-wave, $Pr$ and $Pt$ are the reflected and transmitted p-waves, $qSHr$, $qSHt$, $qSVr$, $qSVt$ are the reflected and transmitted s-waves polarized in the quasi-horizontal/vertical planes, $\theta$ is the incidence angle and $\phi$ is the azimuth angle.For a comprehensive discussion of the theory and methods used, please refer to my masters thesis.*** References1. Zoeppritz, K. (1919), ‘Erdbebenwellen VIII B, Uber Reflexion and Durchgang seismischer wellen durch Unstetigkeisflachen’, *Gottinger Nachr* **1**, 66–84.2. Schoenberg, M. & Protazio, J. (1990), ‘Zoeppritz rationalized, and generalized to anisotropic media’, *The Journal of the Acoustical Society of America* **88**(S1), S46–S46.3. Aki, K. & Richards, P. (1980), *Quantitative seismology; theory and methods*, Freeman Co., San Francisco.4. Thomsen, L. (1986), ‘Weak elastic anisotropy’, *Geophysics* **51**(10), 1954–1966.5. Bond, W. L. (1943), ‘The mathematics of the physical properties of crystals’, *Bell Labs Technical Journal* **22**(1), 1–72.6. Shearer, P. M. (2009), *Introduction to seismology*, Cambridge University Press. *** Demo 1. Isotropic Reflection Coefficients ###Code from zoeppritz import * # import zoeppritz.py ###Output _____no_output_____ ###Markdown First, we need to define the isotropic physical parameters of the upper (1) and lower (2) media. E.g. $vp1$ is the p-wave velocity of the upper medium and $p2$ is the density of the lower medium. Note that all values should be given in **SI units** (e.g. $ms^{-1}$ and $kgm^{-3}$). These are the estimated parameters for the top-chalk interface from Figure 1. ###Code vp1 = 2000; vp2 = 4000 # p-wave velocities vs1 = 400; vs2 = 2150 # s-wave velocities p1 = 2000; p2 = 2600 # densities ###Output _____no_output_____ ###Markdown To generate a scattering matrix$^{\textbf{(4)}}$, call the **isotropic_zoeppritz()** function, passing in the angle of incidence (i_angle) in degrees. The first element of the first row of the scattering matrix is the complex valued p-wave reflection coefficient ($Rpp$). As shown, the magnitude and phase can be easily extracted. ###Code Rpp = isotropic_zoeppritz(vp1, vp2, vs1, vs2, p1, p2, i_angle=35)[0][0] # p-wave reflection coefficient m = abs(Rpp) # magnitude p = np.degrees(cm.phase(Rpp)) # phase print(f'magnitude = {m}, phase = {p}') ###Output magnitude = 0.21887645438940628, phase = -87.20829659064667 ###Markdown In order to make a plot of magnitude and phase vs angle of incidence, call the **isotropic_plot()** function. The solid line represents the magnitude and the dotted line represents the phase shift. ###Code isotropic_plot(vp1, vp2, vs1, vs2, p1, p2) # angle plot ###Output _____no_output_____ ###Markdown Assuming we know the depth to the interface, we can easily convert incidence angle to offset by specifying the depth (d) in metres. The maximum offset can be set by using the maxoff parameter. ###Code isotropic_plot(vp1, vp2, vs1, vs2, p1, p2, d=1000, maxoff=5000) # offset plot ###Output _____no_output_____ ###Markdown The above plot is an elastic model of reflection coefficients vs offset. If we want to invetigate whether acoustic FWI will be effective, as discussed in section 2, we can easily generate an acoustic model by setting the upper and lower s-wave velocities to zero (note that we have to use very small numbers instead of zero to ensure that matrix P is non-singular and invertible). ###Code isotropic_plot(vp1, vp2, 1e-10, 1e-10, p1, p2, d=1000, maxoff=5000) # acoustic offset plot ###Output _____no_output_____ ###Markdown As demonstrated above, there is a huge disparity between the elastic and acoustic post-critical reflection coefficients. This means that acoustic FWI will likely fail. Elastic FWI, or methods for mitigating elastic effects, are required in order to proceed with FWI. *** 2. Anisotropic Reflection Coefficients To model media with vertical transverse isotropy (VTI), we can specify Thomsen's anisotropy parameters ($\epsilon$, $\delta$, $\gamma$)$^{\textbf{(4)}}$ and generate elastic tensors for the upper and lower media using the **thomsen_c()** function. ###Code e1 = 0.1; e2 = 0.2; d1 = 0.05; d2 = 0.1; g1 = 1e-10; g2 = 1e-10 # define some random but realistic Thomsen parameters C1 = thomsen_c(vp1, vs1, p1, e1, d1, g1) # generate elastic tensors C2 = thomsen_c(vp2, vs2, p2, e2, d2, g2) ###Output _____no_output_____ ###Markdown Similarly to before, we can calculate an individual reflection coefficient using **anisotropic_zoeppritz()**, with the addition of the azimuth angle parameter (a_angle) in degrees. The first element of the first row of the first matrix that is returned is the complex valued p-wave reflection coefficient. To generate a plot of magnitude and phase vs incidence angle, call the **anisotropic_plot()** function. Just like before, to convert incidence angle to offset, simply specify the depth (d) in metres. If the results are unstable, try increasing the pre-whitening factor (p_white) to stabilise the solution. ###Code Rpp = anisotropic_zoeppritz(C1, C2, p1, p2, i_angle=10, a_angle=0)[0][0][0] # p-wave reflection coefficient m = abs(Rpp) # magnitude p = np.degrees(cm.phase(Rpp)) # phase print(f'magnitude = {m}, phase = {p}') anisotropic_plot(C1, C2, p1, p2, a_angle=0, p_white=1e-7) # angle plot ###Output _____no_output_____ ###Markdown To model horizontal transverse isotropy (HTI), without losing the convenience of Thomsen's anisotropy parameter, we can apply a 90$^{\circ}$ bond transformation to the elastic tensors$^{\textbf{(5)}}$. This rotates the tensors from VTI to HTI, whilst retaining the same magnitude of anisotropy. Note that for HTI media, the reflection coefficients depend on the azimuth angle (a_angle). ###Code anisotropic_plot(bond_transformation(C1,90), bond_transformation(C2,90), p1, p2, a_angle=70, p_white=1e-7) # hti plot ###Output _____no_output_____ ###Markdown *** 3. Monte Carlo Simulations Monte Carlo simulations can be used to investigate the full range of possible reflection coefficients from a geological interface where the precise lithology and physical parameters are unknown or poorly constrained. Additionally, if you have high quality seismic data available, Monte Carlo simulations can provide constraints on the range of physical parameters that could produce reflections with the amplitude behaviour that is observed in the data. They are also a very effective tool for sensitivity analyses. For instance, to determine the sensitivity of reflection coefficients to small changes in the density, velocity, or anisotropy of the system. ###Code from montecarlo import * # import montecarlo.py ###Output _____no_output_____ ###Markdown 3.1 Isotropic Simulation To demonstrate an isotropic Monte Carlo simulation, I am going to use the same example as before; investigating the top-chalk interface from Figure 1. The estimated parameters that I used to generate the isotropic reflection coefficient profiles were very poorly constrained. In this case, it makes more sense to estimate the minimum and maximum possible values for each parameter, and then fully explore how the reflection coefficients behave in this possible parameter space using Monte Carlo simulations. In order to conduct an isotropic Monte Carlo simulation, we must specify the **minimum** and **maximum** values for each parameter, and the **number of samples**. ###Code vp1_min = 2000; vp1_max = 2000 # upper and lower p-wave velocity range vp2_min = 4000; vp2_max = 4000 vs1_min = 200; vs1_max = 600 # upper and lower s-wave velocity range vs2_min = 1800; vs2_max = 2500 p1_min = 1800; p1_max = 2200 # upper and lower density range p2_min = 2400; p2_max = 2800 num_samples = 300 # number of samples used for the simulation ###Output _____no_output_____ ###Markdown Then, call the **isotropic_monte_carlo()** function. This creates a random uniform distribution of samples for each parameter, which are then used to generate a reflection coefficient profile for each sample. The **isotropic_monte_carlo()** function returns an array of p-wave reflection coefficient profiles of length num_samples and the mean profile (generated using the mean value of each parameter). ###Code SIM_isotropic, mean = isotropic_monte_carlo(vp1_min, vp1_max, vp2_min, vp2_max, \ vs1_min, vs1_max, vs2_min, vs2_max, \ p1_min, p1_max, p2_min, p2_max, num_samples) ###Output 100%|██████████| 300/300 [00:05<00:00, 53.51it/s] ###Markdown To visualise the result, use the **monte_carlo_plot()** function. The blue lines are the simulated reflection coefficient profiles and the red line is the mean profile. Probability density functions are calculated and plotted at incidence angles of 15$^{\circ}$, 45$^{\circ}$ and 75$^{\circ}$. ###Code monte_carlo_plot(SIM_isotropic, mean, pdf=True) ###Output _____no_output_____ ###Markdown 3.2 Anisotropic Simulation I am going to demonstrate how to implement an anisotropic Monte Carlo simulation by showing you how to conduct a sensitivity analysis of p-wave reflection coefficients to weak/moderate amounts of VTI anisotropy. A VTI Monte Carlo simulation is conducted using the same general approach. The only difference is that we need to define the azimuth angle, the minimum and maximum Thomsen parameters, and use the **anisotropic_monte_carlo()** function. Note that this will take longer than the isotropic simulation due to the added complexities and computational cost of solving the anisotropic Zoeppritz equations! ###Code vp1_min = 2000; vp1_max = 2000 # hold isotropic parameters constant vp2_min = 4000; vp2_max = 4000 vs1_min = 600; vs1_max = 600 vs2_min = 1800; vs2_max = 1800 p1_min = 2200; p1_max = 2200 p2_min = 2400; p2_max = 2400 e1_min = -0.02; e1_max = 0.2 # upper and lower epsilon ranges e2_min = -0.02; e2_max = 0.2 d1_min = -0.1; d1_max = 0.2 # upper and lower delta ranges d2_min = -0.1; d2_max = 0.2 g1_min = 1e-10; g1_max = 1e-10 # upper and lower gamma ranges g2_min = 1e-10; g2_max = 1e-10 a_angle = 0 # azimuth angle - note that this does not affect the results for VTI anisotropy num_samples = 300 # number of samples used for the simulation SIM_vti, mean = anisotropic_monte_carlo(vp1_min, vp1_max, vp2_min, vp2_max, \ vs1_min, vs1_max, vs2_min, vs2_max, \ p1_min, p1_max, p2_min, p2_max, \ e1_min, e1_max, e2_min, e2_max, \ d1_min, d1_max, d2_min, d2_max, \ g1_min, g1_max, g2_min, g2_max, \ a_angle, num_samples) monte_carlo_plot(SIM_vti, mean, pdf=True) ###Output _____no_output_____ ###Markdown *** 4. Kirchhoff Synthetic Seismograms Zoeppritz equations assume seismic energy travels in rays with infinite frequency$^{\textbf{(1)}}$. However, seismic waves have finite frequencies and accounting for this will effectively average out the reflection coefficient profile over the Fresnel zone$^{\textbf{(3)}}$. The kirchhoff.py program was developed in order to account for finite frequencies and geometrical spreading by modelling the reflected wave-field. The reflected wave-field for a specific receiver, $\phi_{R}$, is calculated using a Kirchhoff integral of the form: $$\phi_{R}=\frac{1}{4\pi c}\int_{S}\delta\ (t-\frac{r+r_{0}}{c})\ \frac{R_{pp}(\theta_{0})}{r r_{0}}\ (cos\theta_{0}+cos\theta)\ dS\ *\ \frac{\partial w(t,f)}{\partial t}$$where c is the p-wave velocity of the incident medium, $\delta$ is the Dirac delta function, $R_{pp}(\theta_{0})$ is the plane-wave p-wave reflection coefficient at incidence angle $\theta_{0}$, $r_{0}$ and $r$ are the source to surface grid-point and surface grid-point to receiver distances respectively, $\theta$ is the angle between the scattered ray and the surface normal, $w$ is the source time function (Ricker wavelet) and the integral is conducted over surface $S\ ^{\textbf{(6)}}$. For post-critical incidence angles, $\phi_{R}$ becomes complex, requiring an additional phase shift. Therefore, the final Kirchhoff seismogram is given by:$$\phi_{R}^{final}=Re(\phi_{R})\ +\ \textbf{H}\{Im(\phi_{R})\}$$where $\textbf{H}$ is the Hilbert transform, and the real and imaginary component are denoted by $Re$ and $Im$ respectively$^{\textbf{(6)}}$. ###Code from kirchhoff import * # import kirchhoff.py ###Output _____no_output_____ ###Markdown In order to model the reflected wave-field, we first need to specify the geometry of the system we wish to model. ###Code rec_min = 50 # minimum reciever distance from source rec_max = 5000 # maximum reciever distance from source drec = 100 # reciever spacing (every 100 meters in this case) d = 1000 # interface depth ###Output _____no_output_____ ###Markdown Next, we choose some modelling parameters. Note that if the time interval spacing (dt) and grid-point spacing (ds) are too large, the synthetic seismograms will be incorrect. To run this a bit faster you can use a larger grid-point spacing (ds) value. It is recommended that you start small (using the default value or lower), and increase until the resultant seismograms are no longer the same. At this point you have reached the maximum spacing allowed for the system you have defined. ###Code w = 500 # width of 3D Kirchhoff integral ds = 50 # grid-point spacing over the interface dt = 5e-4 # time interval spacing f = 6 # frequency in hz ###Output _____no_output_____ ###Markdown 4.1 Isotropic Synthetic Seismograms To generate an isotropic synthetic seismic trace for each reciever in the specified geometry, call **isotropic_synthetic()**, which will return an array of traces (one for each reciever) and a time array. I am using the same isotropic physical parameters (defined in section 3) for modelling the top-chalk interface in Figure 1. ###Code traces_elastic, time = isotropic_synthetic(rec_min, rec_max, drec, vp1, vp2, vs1, vs2, p1, p2, d, w, f, ds=ds, dt=dt) # isotropic elastic traces ###Output 100%|██████████| 50/50 [00:07<00:00, 6.95it/s] ###Markdown We can plot these traces by passing them into the **plot_synthetic()** function along with the time array. For better visualisation, you can amplify the result by setting the scale_fac parameter to a value greater than 1. The ymin and ymax parameters control the time axis limits. ###Code scale_fac = 1 plot_synthetic(traces_elastic, time, scale_fac, ymin=0.4, ymax=1.4) # plot isotropic elastic traces ###Output _____no_output_____ ###Markdown Just like in section 3, we can also generate an acoustic version of the reflected wave-field by setting the s-wave velocities to (practically) zero. ###Code traces_acoustic, time = isotropic_synthetic(rec_min, rec_max, drec, vp1, vp2, 1e-10, 1e-10, p1, p2, d, w, f, ds=ds, dt=dt) # isotropic acoustic traces plot_synthetic(traces_acoustic, time, scale_fac, ymin=0.4, ymax=1.4) # plot isotropic acoustic traces ###Output _____no_output_____ ###Markdown We can also plot the difference between the acoustic and elastic traces to visualise the disparity between the elastic and acoustic reflected wave-fields. ###Code plot_synthetic(traces_acoustic-traces_elastic, time, scale_fac, ymin=0.4, ymax=1.4) # acoustic - elastic plot ###Output _____no_output_____ ###Markdown 4.2 Anisotropic Synthetic Seismograms To generate anisotropic synthetic traces, we can use the same system geometry and modelling parameters. The only difference is that we need to call the **anisotropic_synthetic()** function which requres elastic tensors, only the upper p-wave velocity, and the azimuth angle. I am using the same anisotropic physical parameters that were defined in section 4. ###Code traces_anisotropic, time = anisotropic_synthetic(rec_min, rec_max, drec, C1, C2, p1, p2, d, w, f, vp1, a_angle=0, ds=ds, dt=dt, p_white=1e-7) # anisotropic traces plot_synthetic(traces_anisotropic, time, scale_fac, ymin=0.4, ymax=1.4) # plot anisotropic traces ###Output _____no_output_____ ###Markdown In order to visualise the minor differences between the anisotropic and elastic traces, we can increase the scale factor. ###Code scale_fac = 4 # increase scale factor to better visualise minor differences plot_synthetic(traces_anisotropic-traces_elastic, time, scale_fac, ymin=0.4, ymax=1.4) # anisotropic - isotropic elastic ###Output _____no_output_____
testing/sahel_cropmask/1_Extract_training_data.ipynb
###Markdown Extracting training data from the ODC* **Products used:** [gm_s2_semiannual](https://explorer.digitalearth.africa/gm_s2_semiannual) DescriptionThis notebook will extract training data over northern Africa using geometries within a shapefile (or geojson). To do this, we rely on a custom `deafrica-sandbox-notebooks` function called `collect_training_data`, contained within the [deafrica_tools.classification](https://github.com/digitalearthafrica/deafrica-sandbox-notebooks/blob/minty-fresh-sandbox/Tools/deafrica_tools/classification.py) script.1. Import, and preview our training data contained in the file: `'data/Northern_training_data_YYYYMMDD.geojson'`2. Extract training data from the datacube using a custom defined feature layer function that we can pass to `collect_training_data`. The training data function is stored in the python file `feature_layer_functions.py` - the functions are stored in a seperate file simply to keep this notebook tidy. - **The features used to create the cropland mask are as follows:** - For two seasons, January to June, and July to Decemeber: - A geomedian composite of nine Sentinel-2 spectral bands - Three measures of median absolute deviation - NDVI, MNDWI, and LAI - Cumulative Rainfall from CHIRPS - Slope from SRTM (not seasonal, obviously) 3. Separate the coordinate values in the returned training data from step 2, and export the coordinates as a text file.4. Export the remaining training data (features other than coordinates) to disk as a text file for use in subsequent scripts*** Getting startedTo run this analysis, run all the cells in the notebook, starting with the "Load packages" cell. Load packages ###Code %matplotlib inline import os import warnings warnings.filterwarnings("ignore") import datacube import numpy as np import xarray as xr import geopandas as gpd from odc.io.cgroups import get_cpu_quota from datacube.utils.geometry import assign_crs from datacube.utils.rio import configure_s3_access configure_s3_access(aws_unsigned=True, cloud_defaults=True) from deafrica_tools.plotting import map_shapefile from deafrica_tools.classification import collect_training_data #import the custom feature layer functions from feature_layer_functions import gm_mads_two_seasons_training ###Output _____no_output_____ ###Markdown Analysis parameters* `path`: The path to the input shapefile from which we will extract training data.* `field`: This is the name of column in your shapefile attribute table that contains the class labels. **The class labels must be integers** ###Code path = 'data/sahel_training_data_20211110.geojson' output_suffix = '20211110' field = 'Class' ###Output _____no_output_____ ###Markdown Automatically find the number of cpus> **Note**: With supervised classification, its common to have many, many labelled geometries in the training data. `collect_training_data` can parallelize across the geometries in order to speed up the extracting of training data. Setting `ncpus>1` will automatically trigger the parallelization, however, its best to set `ncpus=1` to begin with to assist with debugging before triggering the parallelization. ###Code ncpus=round(get_cpu_quota()) print('ncpus = '+str(ncpus)) ###Output ncpus = 31 ###Markdown Load & preview polygon dataWe can load and preview our input data shapefile using `geopandas`. The shapefile should contain a column with class labels (e.g. 'class'). These labels will be used to train our model. > Remember, the class labels **must** be represented by `integers`. ###Code # Load input data shapefile input_data = gpd.read_file(path) # Plot first five rows input_data.head() # Plot training data in an interactive map # map_shapefile(input_data, attribute=field) ###Output _____no_output_____ ###Markdown Now, we can pass this shapefile to `collect_training_data`. For each of the geometries in our shapefile we will extract features in accordance with the function `feature_layer_functions.gm_mads_two_seasons_training`. First, we need to set up a few extra inputs for `collect_training_data` and the datacube. See the function docs [here](https://github.com/digitalearthafrica/deafrica-sandbox-notebooks/blob/03b7b41d5f6526ff3f33618f7a0b48c0d10a155f/Scripts/deafrica_classificationtools.pyL650) for more information on these parameters. ###Code #set up our inputs to collect_training_data zonal_stats = 'median' # Set up the inputs for the ODC query time = ('2019') measurements = [ "blue", "green", "red", "nir", "swir_1", "swir_2", "red_edge_1", "red_edge_2", "red_edge_3", "bcdev", "edev", "sdev" ] resolution = (-10, 10) output_crs = 'epsg:6933' #generate a new datacube query object query = { 'time': time, 'measurements': measurements, 'resolution': resolution, 'output_crs': output_crs, 'group_by' : 'solar_day', 'resampling': 'bilinear' } ###Output _____no_output_____ ###Markdown Extract training data> Remember, if running this function for the first time, its advisable to set `ncpus=1` to assist with debugging before triggering the parallelization (which won't return errors if something is not working correctly). You can also limit the number of polygons to run for the first time by passing in `gdf=input_data[0:5]`, for example. ###Code %%time warnings.filterwarnings("ignore") column_names, model_input = collect_training_data( gdf=input_data, dc_query=query, ncpus=25, field=field, zonal_stats=zonal_stats, fail_threshold=0.0075, feature_func=gm_mads_two_seasons_training ) print(column_names) print('') print(np.array_str(model_input, precision=2, suppress_small=True)) ###Output ['Class', 'blue_S1', 'green_S1', 'red_S1', 'nir_S1', 'swir_1_S1', 'swir_2_S1', 'red_edge_1_S1', 'red_edge_2_S1', 'red_edge_3_S1', 'bcdev_S1', 'edev_S1', 'sdev_S1', 'NDVI_S1', 'LAI_S1', 'MNDWI_S1', 'rain_S1', 'blue_S2', 'green_S2', 'red_S2', 'nir_S2', 'swir_1_S2', 'swir_2_S2', 'red_edge_1_S2', 'red_edge_2_S2', 'red_edge_3_S2', 'bcdev_S2', 'edev_S2', 'sdev_S2', 'NDVI_S2', 'LAI_S2', 'MNDWI_S2', 'rain_S2', 'slope'] [[ 0. 0.15 0.22 ... -0.42 334.4 4.25] [ 0. 0.13 0.22 ... -0.49 258.16 3.17] [ 0. 0.13 0.19 ... -0.46 309.86 2.43] ... [ 1. 0.12 0.15 ... -0.54 632.3 2.43] [ 1. 0.13 0.18 ... -0.5 530.79 1.67] [ 1. 0.13 0.17 ... -0.46 530.7 4.25]] ###Markdown Export training dataOnce we've collected all the training data we require, we can write the data to disk. This will allow us to import the data in the next step(s) of the workflow. ###Code #set the name and location of the output file output_file = "results/training_data/sahel_training_data_"+output_suffix+".txt" #grab all columns except the x-y coords model_col_indices = [column_names.index(var_name) for var_name in column_names] #Export files to disk np.savetxt(output_file, model_input[:, model_col_indices], header=" ".join(column_names), fmt="%4f") ###Output _____no_output_____
_posts/CS20SI-1.ipynb
###Markdown Welcome to TensorFlow- Welcome- Overview of TensorFlow- Graphs and Sessions Why TensorFlow1. Python API;2. Portability - 可迁移性。 能够适用于单个或多个CPU、GPU、服务器,或者移动设备;3. Flexibility - 弹性。 从树莓派、安卓、Windows、iOS、Linux到服务器集群;4. Visualization - 可视化。 其中重要的TensorBoard功能。5. Checkpoints - 断点。 可用于管理诸多实验。6. Auto-differentiation - 自动微分。不再需要手动求微分。7. Large community。 丰富的社区讨论等。8. 大量已经使用TensorFlow搭建的项目。 Goals - 目标1. 理解TensorFlow的计算图方法;2. 探索TensorFlow内置函数;3. 学习构建并结构化深度学习的项目。> Off-the-shelf model are not the main purpose of TensorFlow.> TensorFlow provides an extensive suite of functions and classes that allow users to define models from scratch.> And this is what we are going to learn. 我们不是要使用TensorFlow中现成的模型,而是要学会从零搭建自己的模型。 Books- TensorFlow for Machine Intelligence- Hands-on Machine Learning with Scikit-learn and TensorFlow- Fundamentals of Deep Learning相对而言,TensorFlow发展迅速,书很快就会过时,所以要通过 [官网](https://www.tensorflow.org/)跟踪最新。 Two phases with TensorFlow- Phase 1: assemble a graph 构建一个计算图- Phase 2: use a session to execute operations in the graph. 利用一个session执行该计算图(自动进行微分更新参数) What is a tensor?- 0维:标量;- 1维:向量;- 2维:矩阵;- 等等。 What is a Session?A __Session__ object encapsulates the environment in which Operations objects are executed, and Tensor objects are evaluated. __Session__ 封装一个计算环境,其中完成了 _操作_ 的执行,并计算相应的 _张量_ 。 Why graphs1. 节省计算量。 仅运行只跟所求的变量相关的部分计算图。2. 将计算分割成小块,每一部分便于进行自动微分。3. 方便进行分布式计算,跨多个CPU、GPU或设备等。4. 许多常用的机器学习模型已经训练并通过计算图进行可视化。 小结这一章在自我摸索的情况下,竟然能够正确输出TensorBoard的展示,也算是一个小意外了。 还是从整体性介绍TensorFlow,有谁在用,为什么要用,该如何使用。__该如何使用__ 最为重要, 我们希望能够充分利用TensorFlow已经定义好的模块,如各种函数和类,但是最核心的模型部分,还是得从头搭建。并且特别强调 Computational Graph 计算图这个核心问题。 ###Code import tensorflow as tf x = tf.constant(3, name='x') y = tf.constant(5, name='y') add_op = tf.add(x, y) mul_op = tf.multiply(x, y) useless = tf.multiply(x, add_op) pow_op = tf.pow(add_op, mul_op) with tf.Session() as sess: print(sess.run(pow_op)) summary_writer = tf.summary.FileWriter('tmp/testtb', sess.graph) ###Output 0
icpw_annual_time_series.ipynb
###Markdown ICPW annual time seriesAt the Task Force meeting in May 2017, it was decided that the TOC trends analysis should include rolling regressions based on the annually aggregated data (rather than calculating a single set of statistics for a small number of pre-specified time periods). Some code outlining an approach for this in Python can be found here [here](http://nbviewer.jupyter.org/github/JamesSample/icpw/blob/master/rolling_sens_slope.ipynb), and John has also created a version using R.Heleen has asked me to provided John with annual time series for each site for the period from 1990 to 2012 (see e-mail received from Heleen on 12/05/2017 at 10.01 and also the e-mail chain involving John, Don and Heleen between 19th and 20th May 2017).As a first step, I've modified my previous trends code so that, duirng the processing, the annual time series are saved as a series of CSVs. The additional lines of code can be found on lines 567 to 574 of [toc_trend_analysis.py](https://github.com/JamesSample/icpw/blob/master/toc_trends_analysis.py).The output CSVs require a small amount of manual cleaning: * Delete the TOC series for station ID 23467 and * Delete the SO4 series for station ID 36561. (See sections 1.2 and 1.3 of [this notebook](https://github.com/JamesSample/icpw/blob/master/toc_trends_oct_2016_part3.ipynb) for justification).Annual time series for climate based on the updated climate data have already been created and are saved here:C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015\CRU_Climate_Data\cru_climate_summaries.xlsxThe temperature series also need correcting based on site elevation, as was done in [this notebook](https://github.com/JamesSample/icpw/blob/master/icpw_climate_trends.ipynb) and the two (climate and water chemistry) datasets then need restructuring and joining into a single output file for John to work with. ###Code # Data paths clim_xls = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\CRU_Climate_Data\cru_climate_summaries.xlsx') stn_xls = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\CRU_Climate_Data\cru_stn_elevs.csv') chem_fold = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\Results\annual_chemistry_series') # For performance, pre-load the climate output file # (this saves having to read it within the inner loop below, # which is very slow) clim_dict = {} for var in ['pre', 'tmp']: for tm in ['ann', 'jja', 'jas']: # Open the climate data clim_df = pd.read_excel(clim_xls, sheetname='%s_%s' % (var, tm)) clim_dict[(var, tm)] = clim_df # Read stn elev data stn_df = pd.read_csv(stn_xls) # Get list of sites stn_list = stn_df['stn_id'].unique() # List to store output data_list = [] # Loop over stations for stn in stn_list: # Read chem data chem_path = os.path.join(chem_fold, 'stn_%s.csv' % stn) # Only process the 431 files with chem data if os.path.exists(chem_path): # Allow for manually edited stations (see above) # which now have ';' as the delimiter if stn in [23467, 36561]: chem_df = pd.read_csv(chem_path, sep=';') else: chem_df = pd.read_csv(chem_path) chem_df.index = chem_df['YEAR'] # Process climate data # Dict to store output data_dict = {} # Loop over data for var in ['pre', 'tmp']: for tm in ['ann', 'jja', 'jas']: # Get the climate data clim_df = clim_dict[(var, tm)] # Filter the climate data for this station stn_clim_df = clim_df.query('stn_id == @stn') # Set index stn_clim_df.index = stn_clim_df['year'] stn_clim_df = stn_clim_df.sort_index() # Correct temperatures according to lapse rate if var == 'tmp': # Get elevations stn_elev = stn_df.query('stn_id == @stn')['elev_m'].values[0] px_elev = stn_df.query('stn_id == @stn')['px_elev_m'].values[0] # If pixel elev is negative (i.e. in sea), correct back to s.l. if px_elev < 0: px_elev = 0 # Calculate temperature difference based on 0.6C/100m t_diff = 0.6 * (px_elev - stn_elev) / 100. # Apply correction stn_clim_df['tmp'] = stn_clim_df['tmp'] + t_diff # Truncate stn_clim_df = stn_clim_df.query('(year>=1990) & (year<=2012)') # Add to dict key = '%s_%s' % (var, tm) val = stn_clim_df[var] data_dict[key] = val # Build output df stn_clim_df = pd.DataFrame(data_dict) # Join chem and clim data df = pd.merge(stn_clim_df, chem_df, how='outer', left_index=True, right_index=True) # Get desired columns # Modified 06/06/2017 to include all pars for Leah df = df[['pre_ann', 'pre_jas', 'pre_jja', 'tmp_ann', 'tmp_jas', 'tmp_jja', 'Al', 'TOC', 'EH', 'ESO4', 'ECl', 'ESO4_ECl', 'ENO3', 'ESO4X', 'ESO4_ECl', 'ECa_EMg', 'ECaX_EMgX', 'ANC']] # Transpose df = df.T # Add station ID df.reset_index(inplace=True) df['station_id'] = stn # Rename cols df.columns.name = '' cols = list(df.columns) cols[0] = 'var' df.columns = cols data_list.append(df) # Combine results for each site ann_df = pd.concat(data_list, axis=0) ann_df.head() ###Output _____no_output_____ ###Markdown The final step is to join in the site metadata used in the previous analysis. ###Code # Read site data from previous output in_xls = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\Results\toc_trends_long_format_update1.xlsx') props_df = pd.read_excel(in_xls, sheetname='toc_trends_long_format_update1', keep_default_na=False) # Otherwise 'NA' for North America becomes NaN # Get just cols of interest props_df = props_df[['project_id', 'project_name', 'station_id', 'station_code', 'station_name', 'nfc_code', 'type', 'continent', 'country', 'region', 'subregion', 'lat', 'lon']] # Drop duplicates props_df.drop_duplicates(inplace=True) # Join ann_df = pd.merge(ann_df, props_df, how='left', on='station_id') # Reorder cols ann_df = ann_df[['project_id', 'project_name', 'station_id', 'station_code', 'station_name', 'nfc_code', 'type', 'continent', 'country', 'region', 'subregion', 'lat', 'lon', 'var']+range(1990, 2013)] ann_df.head() ###Output _____no_output_____ ###Markdown Heleen previously defined various criteria for whether a series should be included in the analysis or not. In order to keep things consistent, it's probably a good idea to include this information here. ###Code # Read site data from previous output in_xls = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\Results\toc_trends_long_format_update1.xlsx') inc_df = pd.read_excel(in_xls, sheetname='toc_trends_long_format_update1') # Get just cols of interest inc_df = inc_df[['station_id', 'par_id', 'analysis_period', 'include']] # Filter to just results for 1990-2012 inc_df = inc_df.query('analysis_period == "1990-2012"') # Join ann_df = pd.merge(ann_df, inc_df, how='left', left_on=['station_id', 'var'], right_on=['station_id', 'par_id']) # Reorder cols ann_df = ann_df[['project_id', 'project_name', 'station_id', 'station_code', 'station_name', 'nfc_code', 'type', 'continent', 'country', 'region', 'subregion', 'lat', 'lon', 'var', 'include']+range(1990, 2013)] # The climate vars all have data and can be included ann_df['include'].fillna(value='yes', inplace=True) # Write output out_path = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\Results\annual_clim_chem_series_leah.csv') ann_df.to_csv(out_path, encoding='utf-8') ann_df.head() ###Output _____no_output_____ ###Markdown Finally, it's worth checking that this output matches the time series available on the ICPW website and in the plots of the climate data. ###Code in_xls = (r'C:\Data\James_Work\Staff\Heleen_d_W\ICP_Waters\TOC_Trends_Analysis_2015' r'\Results\annual_clim_chem_series.xlsx') df = pd.read_excel(in_xls, sheetname='annual_clim_chem_series') # Select series stn_id = 23455 var = 'tmp_jja' df2 = df.query("(station_id==@stn_id) & (var==@var)") df2 = df2[range(1990, 2013)].T df2.columns = [var] df2.plot(ls='-', marker='o') ###Output _____no_output_____
scripts/Auto Mun/Auto Mun.ipynb
###Markdown **Set Up Planet Objects** ###Code for name, obj in conn.space_center.bodies.items(): if name == 'Kerbin': kerbin = obj elif name == 'Mun': mun = obj mass_kerbin = 5.2915158e22 # (kg) mass_mun = 9.7599066e20 # (kg) G = 6.67408e-11 # (m**3/kg/s**2) gravitational constant ###Output _____no_output_____ ###Markdown **Reference Frames** ###Code kerbin_nrrf = kerbin.non_rotating_reference_frame ###Output _____no_output_____ ###Markdown **Helper Functions** ###Code def warp_to_node(conn, vessel, node, burn_duration): # set attitude vessel.control.sas_mode = conn.space_center.SASMode.maneuver ang_vel = vessel.angular_velocity(vessel.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) while not oriented: ang_vel = vessel.angular_velocity(vessel.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) # warp end_warp = node.ut - burn_duration conn.space_center.rails_warp_factor = 4 while conn.space_center.ut < (end_warp - 1000): time.sleep(0.1) conn.space_center.rails_warp_factor = 3 while conn.space_center.ut < (end_warp - 300): time.sleep(0.1) conn.space_center.rails_warp_factor = 2 while conn.space_center.ut < (end_warp - 100): time.sleep(0.1) conn.space_center.rails_warp_factor = 1 while conn.space_center.ut < (end_warp - 15): time.sleep(0.1) conn.space_center.rails_warp_factor = 0 def execute_node(conn, vessel, node, burn_duration): while conn.space_center.ut < (node.ut-(burn_duration/2)): time.sleep(0.1) while node.remaining_delta_v > 15: vessel.control.throttle = 1 while node.remaining_delta_v > 0.1: vessel.control.throttle = 0.05 node.remove() vessel.control.throttle = 0 ###Output _____no_output_____ ###Markdown **Orbital Mechanics** ###Code def calc_burn_duration(vessel, dV): ''' Calculates the burn time for a given vessel and a given delta-V. ''' m = vessel.mass isp = vessel.specific_impulse thrust = vessel.available_thrust term1 = (m*isp*9.81)/(thrust) term2 = (1-e**(dV/(isp*9.81))) return abs(term1*term2) def calc_circ_orb_speed(r, M): ''' Calculates the orbital speed of a vessel in a circular orbit. ''' return np.sqrt(G*M/r) def vis_viva(r, a, M): ''' Calculates the orbital speed of a vessel at some point in an elliptical orbit given the semi-major axis and mass of orbiting body. ''' return np.sqrt(G*M*((2/r)-(1/a))) def orbital_period(a, M): ''' Calculates the orbital period of a satellite given the semi-major axis of the orbit and the mass of the orbiting body. ''' mu = G*M return (2*np.pi)*np.sqrt(a**3/mu) def v_pe_hyperbolic(M, a, e): ''' Find periapsis speed of hyperbolic orbit. ''' k = G*M # gravitational parameter return ((-k/a)*(1+e)/(e-1))**0.5 ###Output _____no_output_____ ###Markdown **Kerbin Ascent**This ascent profile is custom for this vehicle. It would be nice to have a generalized script (at least attempt) to launch any vehicle to orbit.`TODO`: Convert times to use `MET` rather than `time.time()` so that physical time warp is stable. ###Code # vehicle telemetry telem = vessel.flight(vessel.orbit.body.reference_frame) # go straight up at first vessel.auto_pilot.target_pitch_and_heading(90, 90) vessel.auto_pilot.engage() time.sleep(1) # throttle 100% and ignition vessel.control.throttle = 1 vessel.control.activate_next_stage() # wait until going 100 m/s straight up v_speed = telem.vertical_speed while v_speed < 100: v_speed = telem.vertical_speed time.sleep(0.1) # pitch over at a rate of 1.125 deg/s for 40 s # this gets vehicle to 45 deg by 10 km pitch_start = time.time() pitch_end = pitch_start + 40 # 40 seconds of pitching time_to_deg = 1.125 while time.time() < pitch_end: tgt_deg = 90 - ((time.time() - pitch_start) * time_to_deg) vessel.auto_pilot.target_pitch_and_heading(tgt_deg, 90) print('Lower atmosphere pitch complete.') # pitch over at a lower rate until first stage depleted pitch_start = time.time() pitch_end = pitch_start + 25 # ~25 seconds until depleted while time.time() < pitch_end: tgt_deg = 45 - ((time.time() - pitch_start) * 0.4) vessel.auto_pilot.target_pitch_and_heading(tgt_deg, 90) # MECO stg_1_resrcs = vessel.resources_in_decouple_stage(stage=9, cumulative=False) stg_1_lqd_fu = stg_1_resrcs.amount('LiquidFuel') while stg_1_lqd_fu > 0.01: stg_1_lqd_fu = stg_1_resrcs.amount('LiquidFuel') vessel.auto_pilot.target_pitch_and_heading(35, 90) time.sleep(1) print('MECO.') # S1 Sep vessel.control.throttle = 0 vessel.control.activate_next_stage() time.sleep(1) print('Stage 1 sep.') # S2 Ignition vessel.control.activate_next_stage() vessel.control.throttle = 1 # pitch over slowly on S2 until apoapsis ~100 km pitch_start = time.time() pitch_end = pitch_start + 25 while time.time() < pitch_end: tgt_deg = 35 - ((time.time() - pitch_start) * 1.0) vessel.auto_pilot.target_pitch_and_heading(tgt_deg, 90) apo_alt = vessel.orbit.apoapsis_altitude while apo_alt < 100000: apo_alt = vessel.orbit.apoapsis_altitude vessel.auto_pilot.target_pitch_and_heading(10, 90) # stabilize for coast to apoapsis vessel.control.throttle = 0 print('SECO.') vessel.auto_pilot.disengage() time.sleep(1) vessel.control.sas = True time.sleep(1) vessel.control.sas_mode = conn.space_center.SASMode.prograde time.sleep(1) ###Output Lower atmosphere pitch complete. MECO. Stage 1 sep. SECO. ###Markdown **Circularize Kerbin Orbit** ###Code # wait until out of atmosphere conn.space_center.physics_warp_factor = 3 while vessel.flight().mean_altitude < 70000: time.sleep(1) conn.space_center.physics_warp_factor = 0 # fairing & LES sep time.sleep(3) vessel.control.activate_next_stage() print('Fairing sep.') print('LAS sep.') print('Out of atmosphere.') # go to map screen time.sleep(3) conn.space_center.camera.mode = conn.space_center.CameraMode.map # relevant times time.sleep(1) ut_set = conn.space_center.ut t_to_ap = vessel.orbit.time_to_apoapsis ut_ap = ut_set + t_to_ap # determine delta-V required to circularize V_ap_reqd = calc_circ_orb_speed(r=vessel.orbit.apoapsis, M=mass_kerbin) V_ap_curr = vis_viva(r=vessel.orbit.apoapsis, a=vessel.orbit.semi_major_axis, \ M=mass_kerbin) dV_node = abs(V_ap_reqd - V_ap_curr) # create circularization node circ = vessel.control.add_node(ut=ut_ap) circ.prograde = dV_node print('Created circularization maneuver node.') # set SAS to maneuver and wait until oriented print('Orienting for maneuver.') vessel.control.sas_mode = conn.space_center.SASMode.maneuver time.sleep(3) ang_vel = vessel.angular_velocity(vessel.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) while not oriented: ang_vel = vessel.angular_velocity(vessel.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) # time warp to maneuver print('Warping to maneuver.') burn_time = calc_burn_duration(vessel, dV_node) conn.space_center.physics_warp_factor = 3 while conn.space_center.ut < (ut_ap - (burn_time / 2) - 15): time.sleep(1) conn.space_center.physics_warp_factor = 0 conn.space_center.camera.mode = conn.space_center.CameraMode.automatic time.sleep(1) # execute maneuver while conn.space_center.ut < (ut_ap - (burn_time / 2)): time.sleep(0.1) while circ.remaining_delta_v > 15: vessel.control.throttle = 1 while circ.remaining_delta_v > 0.1: vessel.control.throttle = 0.05 circ.remove() vessel.control.throttle = 0 print('Circularization complete.') ###Output _____no_output_____ ###Markdown **CSM Detach, Flip, Dock Maneuver**This docking method depends entirely on this craft, and doesn't work generally. It'd be nice to have a general docking script that works for any two crafts. ###Code # set vessel to SAS normal and wait until oriented print('Orienting to normal... ', end='') vessel.control.sas_mode = conn.space_center.SASMode.normal time.sleep(3) ang_vel = vessel.angular_velocity(vessel.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) while not oriented: ang_vel = vessel.angular_velocity(vessel.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) print('Done.') # turn off RCS and undock CSM vessel.control.rcs = False time.sleep(1) vessel.control.activate_next_stage() # turn on RCS, thrust normal briefly, then stop vessel.control.rcs = True time.sleep(1) vessel.control.forward = 1 time.sleep(0.5) vessel.control.forward = 0 time.sleep(2) vessel.control.forward = -1 time.sleep(0.5) vessel.control.forward = 0 # turn off RCS vessel.control.rcs = False print('CSM sep.') # control CSM from docking port print('Docking maneuver... ', end='') CSM_docking_port = vessel.parts.with_title('Clamp-O-Tron Docking Port')[0] vessel.parts.controlling = CSM_docking_port # target LM docking port vessels = conn.space_center.vessels for v in vessels: if v.name == 'Auto Mun Lander': LM = v elif v.name == 'Auto Mun': CSM = v LM_docking_port = LM.parts.docking_ports[0] # DockingPort object conn.space_center.target_docking_port = LM_docking_port # set CSM SAS to target vessel.control.sas_mode = conn.space_center.SASMode.target time.sleep(1) # switch to LM conn.space_center.active_vessel = LM time.sleep(1) # control LM from docking port LM_parts = LM.parts.all for part in LM_parts: if part.name == 'dockingPort2': LM_dp = part # Part object LM.parts.controlling = LM_dp # Requires Part, not DockingPort time.sleep(1) # target CSM docking port CSM_dp = CSM.parts.docking_ports[0] conn.space_center.target_docking_port = CSM_dp time.sleep(1) # set LM SAS to target LM.control.sas_mode = conn.space_center.SASMode.target time.sleep(1) # switch to CSM conn.space_center.active_vessel = CSM time.sleep(1) CSM.control.sas_mode = conn.space_center.SASMode.target time.sleep(1) vessel.control.rcs = True time.sleep(1) # thrust slightly in direction of target docking port, turn off RCS, wait until docked # this requires some Patience & Luck (TM) n_vessels = len(conn.space_center.vessels) vessel.control.forward = 1 time.sleep(1.5) vessel.control.forward = 0 # wait until docked while len(conn.space_center.vessels) == n_vessels: time.sleep(1) # turn off RCS vessel.control.rcs = False print('Done.') ###Output _____no_output_____ ###Markdown **Trans-Munar Injection** ###Code print('Setting up Trans-Munar Injection.') # redefine newly docked vessel CSM_LM_S2 = conn.space_center.active_vessel # control from LM (for orientation purposes) for part in CSM_LM_S2.parts.all: if part.name == 'mk2LanderCabin.v2': CSM_LM_S2.parts.controlling = part # set Mun as target conn.space_center.target_body = mun time.sleep(1) # create TMI maneuver node (Hohmann Transfer) # calculate dV required for maneuver v_s_hi = calc_circ_orb_speed(r=CSM_LM_S2.orbit.apoapsis, M=mass_kerbin) v_s_lo = calc_circ_orb_speed(r=CSM_LM_S2.orbit.periapsis, M=mass_kerbin) v_s = (v_s_hi + v_s_lo) / 2 r_b = CSM_LM_S2.orbit.semi_major_axis a_tmi = (100000 + 600000 + mun.orbit.apoapsis) / 2 v_pe = vis_viva(r=r_b, a=a_tmi, M=mass_kerbin) dv = v_pe - v_s # calculate time until maneuver node # half-period of transfer orbit p = orbital_period(a=a_tmi, M=mass_kerbin) # period of one TMI orbit p_2 = 0.5 * p # time of one half TMI orbit # arc distance Mun travels in time p_2 v_m = mun.orbit.speed # Mun orbital velocity s = v_m * p_2 # arc-length distance Mun travels # angle between Mun current pos and Mun intercept pos r_m_i = mun.position(kerbin_nrrf) theta = s / np.linalg.norm(r_m_i) # radians # Mun position at intercept x1, y1, z1 = r_m_i[0], r_m_i[1], r_m_i[2] x2 = x1 * np.cos(theta) - z1 * np.sin(theta) y2 = 0 z2 = x1 * np.sin(theta) + z1 * np.cos(theta) r_m_f = (x2, y2, z2) # find time to burn r_s_b = [-1 * r_m_f[i] for i in range(len(r_m_f))] r_s_i = CSM_LM_S2.position(kerbin_nrrf) r_s_i = [r_s_i[0], 0, r_s_i[2]] # nullify inclination x1, y1 = r_s_i[0], r_s_i[2] x2, y2 = r_s_b[0], r_s_b[2] theta2 = np.arctan2(x1*y2-y1*x2, x1*x2+y1*y2) if theta2 < 0: theta2 += 2*np.pi s_to_b = CSM_LM_S2.orbit.semi_major_axis * theta2 t_to_b = s_to_b / v_s # create node ut_b = conn.space_center.ut + t_to_b tmi = CSM_LM_S2.control.add_node(ut=ut_b) tmi.prograde = dv # set SAS to maneuver and wait until oriented print('Orienting for maneuver.') CSM_LM_S2.control.sas_mode = conn.space_center.SASMode.maneuver CSM_LM_S2.control.rcs = True time.sleep(3) ang_vel = CSM_LM_S2.angular_velocity(CSM_LM_S2.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.1 for i in range(3)) while not oriented: ang_vel = CSM_LM_S2.angular_velocity(CSM_LM_S2.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.1 for i in range(3)) # time warp to maneuver print('Warping to maneuver.') burn_time = calc_burn_duration(CSM_LM_S2, dv) conn.space_center.rails_warp_factor = 2 while conn.space_center.ut < (ut_b - (burn_time / 2) - 60): time.sleep(1) conn.space_center.rails_warp_factor = 0 conn.space_center.camera.mode = conn.space_center.CameraMode.automatic time.sleep(1) print('Burn baby, burn.') # execute maneuver while conn.space_center.ut < (ut_b - (burn_time / 2)): time.sleep(0.1) while tmi.remaining_delta_v > 15: CSM_LM_S2.control.throttle = 1 while tmi.remaining_delta_v > 0.1: CSM_LM_S2.control.throttle = 0.05 tmi.remove() CSM_LM_S2.control.throttle = 0 CSM_LM_S2.control.rcs = False print('TMI complete.') ###Output _____no_output_____ ###Markdown **Outbound Trajectory Correction** ###Code time.sleep(3) # separate from Stage 2 for part in CSM_LM_S2.parts.all: if part.tag == 'Decoupler.S2': part.decoupler.decouple() # rename current vessel CSM_LM = conn.space_center.active_vessel print('S2 sep.') for part in CSM_LM.parts.all: if part.name == 'engineLargeSkipper': # activate SM engine part.engine.active = True elif part.name == 'liquidEngine2-2.v2': # ensure LM engine deactivated part.engine.active = False elif part.name == 'mk1-3pod': # control from CM CSM_LM.parts.controlling = part # create outbound trajectory correction node print('Set up OTC node.') ut_otc = conn.space_center.ut + 3000 otc = CSM_LM.control.add_node(ut=ut_otc) otc.prograde = 12.5 # maneuver to node CSM_LM.control.sas_mode = conn.space_center.SASMode.maneuver time.sleep(3) ang_vel = CSM_LM.angular_velocity(CSM_LM.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) while not oriented: ang_vel = CSM_LM.angular_velocity(CSM_LM.orbit.body.reference_frame) oriented = all(ang_vel[i] < 0.01 for i in range(3)) # warp carefully to node print('Warping to node... ', end='') conn.space_center.rails_warp_factor = 4 while conn.space_center.ut < (ut_otc - 1000): time.sleep(0.1) conn.space_center.rails_warp_factor = 3 while conn.space_center.ut < (ut_otc - 300): time.sleep(0.1) conn.space_center.rails_warp_factor = 2 while conn.space_center.ut < (ut_otc - 100): time.sleep(0.1) conn.space_center.rails_warp_factor = 1 while conn.space_center.ut < (ut_otc - 15): time.sleep(0.1) conn.space_center.rails_warp_factor = 0 print('Done.') # execute node print('Cleaning up Mun intercept... ', end='') while conn.space_center.ut < (ut_otc - 5): time.sleep(0.1) while otc.remaining_delta_v > 0.1: CSM_LM.control.throttle = 0.05 CSM_LM.control.throttle = 0 otc.remove() vessel = CSM_LM print('Done.') ###Output _____no_output_____ ###Markdown **Circularize Lunar Orbit** ###Code # wait until in Mun SOI print('Warping to Mun SOI... ', end='') time.sleep(1.0) # wait for clunkiness to clear conn.space_center.rails_warp_factor = 5 while vessel.orbit.body.name != 'Mun': time.sleep(0.1) time.sleep(1.0) # wait for SOI change clunkiness to clear conn.space_center.rails_warp_factor = 0 print('Done.') # create circularization maneuver node t_to_pe = vessel.orbit.time_to_periapsis ut_pe = conn.space_center.ut + t_to_pe speed_pe = v_pe_hyperbolic(M=mass_mun, a=vessel.orbit.semi_major_axis, \ e=vessel.orbit.eccentricity) speed_circ = calc_circ_orb_speed(r=vessel.orbit.periapsis, M=mass_mun) dV = -abs(speed_pe - speed_circ) # minus to slow down circ = vessel.control.add_node(ut=ut_pe, prograde=dV) # warp to node burn_duration = calc_burn_duration(vessel, dV) warp_to_node(conn, vessel, circ, burn_duration) # execute maneuver print('Executing node... ', end='') execute_node(conn, vessel, circ, burn_duration) print('Done.') ###Output _____no_output_____ ###Markdown **Crew Transfer**TODO: can't figure out how to transfer crew automatically. For now, need to launch with two pilots, one in each command pod. **Mun Landing** ###Code # fill lander fuel tanks from SM fuel tanks # lander detach # wait until passing over night->day meridian # deorbit burn # landing sequence ###Output _____no_output_____
Kaggle Datasets/DL/Mushroom Classification/mushroom-ann.ipynb
###Markdown Data : Attribute Information: (classes: edible=e, poisonous=p) --> cap-shape: bell=b,conical=c,convex=x,flat=f, knobbed=k,sunken=s --> cap-surface: fibrous=f,grooves=g,scaly=y,smooth=s --> cap-color: brown=n,buff=b,cinnamon=c,gray=g,green=r,pink=p,purple=u,red=e,white=w,yellow=y --> bruises: bruises=t,no=f --> odor: almond=a,anise=l,creosote=c,fishy=y,foul=f,musty=m,none=n,pungent=p,spicy=s --> gill-attachment: attached=a,descending=d,free=f,notched=n --> gill-spacing: close=c,crowded=w,distant=d --> gill-size: broad=b,narrow=n --> gill-color: black=k,brown=n,buff=b,chocolate=h,gray=g, green=r,orange=o,pink=p,purple=u,red=e,white=w,yellow=y --> stalk-shape: enlarging=e,tapering=t --> stalk-root: bulbous=b,club=c,cup=u,equal=e,rhizomorphs=z,rooted=r,missing=? --> stalk-surface-above-ring: fibrous=f,scaly=y,silky=k,smooth=s --> stalk-surface-below-ring: fibrous=f,scaly=y,silky=k,smooth=s --> stalk-color-above-ring: brown=n,buff=b,cinnamon=c,gray=g,orange=o,pink=p,red=e,white=w,yellow=y --> stalk-color-below-ring: brown=n,buff=b,cinnamon=c,gray=g,orange=o,pink=p,red=e,white=w,yellow=y --> veil-type: partial=p,universal=u --> veil-color: brown=n,orange=o,white=w,yellow=y --> ring-number: none=n,one=o,two=t --> ring-type: cobwebby=c,evanescent=e,flaring=f,large=l,none=n,pendant=p,sheathing=s,zone=z --> spore-print-color: black=k,brown=n,buff=b,chocolate=h,green=r,orange=o,purple=u,white=w,yellow=y --> population: abundant=a,clustered=c,numerous=n,scattered=s,several=v,solitary=y --> habitat: grasses=g,leaves=l,meadows=m,paths=p,urban=u,waste=w,woods=d Importing Libraries & getting Data ###Code import numpy as np import pandas as pd import matplotlib.pyplot as plt import seaborn as sns plt.style.use('dark_background') import warnings warnings.filterwarnings('ignore') data = pd.read_csv('dataset/mushrooms.csv') data.head() data.info() for i in data.columns: print('{} consists of : {} unique-values , which are --> {}\n'.format(i ,len(pd.value_counts(data[i])) ,data[i].unique())) ###Output class consists of : 2 unique-values , which are --> ['p' 'e'] cap-shape consists of : 6 unique-values , which are --> ['x' 'b' 's' 'f' 'k' 'c'] cap-surface consists of : 4 unique-values , which are --> ['s' 'y' 'f' 'g'] cap-color consists of : 10 unique-values , which are --> ['n' 'y' 'w' 'g' 'e' 'p' 'b' 'u' 'c' 'r'] bruises consists of : 2 unique-values , which are --> ['t' 'f'] odor consists of : 9 unique-values , which are --> ['p' 'a' 'l' 'n' 'f' 'c' 'y' 's' 'm'] gill-attachment consists of : 2 unique-values , which are --> ['f' 'a'] gill-spacing consists of : 2 unique-values , which are --> ['c' 'w'] gill-size consists of : 2 unique-values , which are --> ['n' 'b'] gill-color consists of : 12 unique-values , which are --> ['k' 'n' 'g' 'p' 'w' 'h' 'u' 'e' 'b' 'r' 'y' 'o'] stalk-shape consists of : 2 unique-values , which are --> ['e' 't'] stalk-root consists of : 5 unique-values , which are --> ['e' 'c' 'b' 'r' '?'] stalk-surface-above-ring consists of : 4 unique-values , which are --> ['s' 'f' 'k' 'y'] stalk-surface-below-ring consists of : 4 unique-values , which are --> ['s' 'f' 'y' 'k'] stalk-color-above-ring consists of : 9 unique-values , which are --> ['w' 'g' 'p' 'n' 'b' 'e' 'o' 'c' 'y'] stalk-color-below-ring consists of : 9 unique-values , which are --> ['w' 'p' 'g' 'b' 'n' 'e' 'y' 'o' 'c'] veil-type consists of : 1 unique-values , which are --> ['p'] veil-color consists of : 4 unique-values , which are --> ['w' 'n' 'o' 'y'] ring-number consists of : 3 unique-values , which are --> ['o' 't' 'n'] ring-type consists of : 5 unique-values , which are --> ['p' 'e' 'l' 'f' 'n'] spore-print-color consists of : 9 unique-values , which are --> ['k' 'n' 'u' 'h' 'w' 'r' 'o' 'y' 'b'] population consists of : 6 unique-values , which are --> ['s' 'n' 'a' 'v' 'y' 'c'] habitat consists of : 7 unique-values , which are --> ['u' 'g' 'm' 'd' 'p' 'w' 'l'] ###Markdown Handling Missing Values ###Code data.isnull().sum() ###Output _____no_output_____ ###Markdown Analysing Target Variable i.e class ###Code data['class'].value_counts() sns.countplot(x='class' ,data=data ) ###Output _____no_output_____ ###Markdown Encoding ###Code from sklearn.preprocessing import LabelEncoder encoder = LabelEncoder() for i in data.columns: data[i] = encoder.fit_transform(data[i]) data.head() data.shape ###Output _____no_output_____ ###Markdown Correlation ###Code data.corr()['class'].sort_values(ascending=False) plt.figure(figsize=(20,10)) sns.heatmap(data.corr() ,annot=True ,cmap='RdYlGn') plt.show() ###Output _____no_output_____ ###Markdown Model Train-Test Split ###Code X = data.drop(['class'] ,axis=1) y = data['class'].values X.shape ,y.shape from sklearn.model_selection import train_test_split X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42) X_train.shape, X_test.shape, y_train.shape, y_test.shape ###Output _____no_output_____ ###Markdown Model Building ###Code import tensorflow as tf from keras.wrappers.scikit_learn import KerasClassifier from sklearn.model_selection import cross_val_score def model_building(): model = tf.keras.models.Sequential() model.add(tf.keras.layers.Dense(units=8, kernel_initializer='uniform', activation='relu', input_dim=X_train.shape[1])) model.add(tf.keras.layers.Dense(units=8 ,kernel_initializer='uniform' ,activation='relu')) model.add(tf.keras.layers.Dense(units=1 ,kernel_initializer='uniform' ,activation='sigmoid')) model.compile(optimizer='adam', loss='binary_crossentropy' ,metrics=['accuracy']) return model classifier_model = KerasClassifier(build_fn=model_building ,epochs=70 ,batch_size=10) accuracies = cross_val_score(estimator=classifier_model ,X=X_train ,y=y_train ,cv=2) mean = accuracies.mean() variance = accuracies.std() print('Mean Accuracy :' ,str(mean)) print('Mean Variance :', str(variance)) model_history = classifier_model.fit(X_train ,y_train ,validation_split=0.20 ,epochs=70 ,batch_size=10) print(model_history.history.keys()) plt.figure(figsize=(10,5)) plt.plot(model_history.history['accuracy']) plt.plot(model_history.history['val_accuracy']) plt.title('model accuracy') plt.ylabel('accuracy') plt.xlabel('epoch') plt.legend(['train', 'test'], loc='best') plt.show() plt.figure(figsize=(10,5)) plt.plot(model_history.history['loss']) plt.plot(model_history.history['val_loss']) plt.title('model loss') plt.ylabel('loss') plt.xlabel('epoch') plt.legend(['train', 'test'], loc='best') plt.show() ###Output dict_keys(['loss', 'accuracy', 'val_loss', 'val_accuracy'])
matematica-para-datascience/Lectures/Lecture-21-Principal-Component-Analysis.ipynb
###Markdown Lecture 21: Principal Component Analysis ###Code import numpy as np import matplotlib.pyplot as plt import numpy.linalg as LA %matplotlib inline ###Output _____no_output_____ ###Markdown PCA and data preprocessingPrincipal Components Analysis (PCA) is a dimensionality reduction algorithm that can be used to significantly speed up our feature(s) learning algorithm. Mathematically speaking, PCA uses a method called Singular Value Decomposition (SVD), in which the singular value resembles the eigenvalue.Suppose we are training a model to classify our MNIST handwritten digits (28x28 grayscale images). A given training sample in `X_train` is of shape `(784,)`, which has 784 features (dimensions). However, many of these features are somewhat redundant, because the values of adjacent pixels in an image are highly correlated. Concretely, for a training vector $\mathbf{x} = (x_1,\dots, x_{784}) \in \mathbb{R}^{784}$ which is 784 dimensional vectors, with each feature $x_j$ corresponding to the intensity of $j$-th pixel (in the flattened image). Because of the correlation between adjacent pixels, PCA will allow us to approximate the input with a much lower dimensional one, while incurring very little error. (Reading) What does PCA exactly do?PCA will find the most significant "direction" in a dataset, then the second most significant direction, then the third most significant direction,.... so on and so forth.Let $A\in \mathbb{R}^{n\times d}$ matrix, usually we assume $n>d$>Singular Value Decomposition (SVD): any real matrix can be decomposed into the following form:>$$A = U S V^{\top}$$Where $S\in \mathbb{R}^{n\times d}$ is a diagonal matrix whose diagonal entries are non-negative and in decreasing order. $U\in \mathbb{R}^{n\times n}$ and $V\in \mathbb{R}^{d\times d}$ are orthogonal matrices (i.e. columns of $V$ are orthonormal, same for $U$, $U^{\top}U = UU^{\top} = I$). The columns of $V= [\mathbf{v}_1 \mathbf{v}_2 \cdots \mathbf{v}_d]$ are the significant directions we were looking for, which are known as the right singular vectors. Moreover, the columns of $V$ form a $d$-dimensional orthogonal basis.The columns of $U = [\mathbf{u}_1 \mathbf{u}_2 \cdots \mathbf{u}_n]$ are known as the left singular vectors. Sometimes we just say "eigenvectors" instead of "singular vectors".If some singular values are 0 or very small, we can essentially "discard" those singular values and the corresponding eigenvectors, and still get a reasonably good approximation of our data, hence reducing the dimensions.Finally, $U$ (or more precisely $U S$) stores how you write each coordinate vector in terms of the significant directions in $V$. ---- Geometric meaningConsider what happens to the unit sphere in our vector space as it isbeing transformed by the matrix $X$. First, we apply some transformation $V^{\top}$, which is essentially a rotation, since $V^{\top}$ is a matrix with orthonormal rows. A matrix with orthonormal rows just changes the coordinate axes via some rotation or reflection but does no scaling.Next, we apply a scaling defined by $S$, which just scales the dimensions since it is a diagonal matrix. Finally, we rotate again with $U$. In other words, any transformation can be expressed as a rotation followed by a scaling followed by another rotation. ---- Rank-$k$ approximationThe vectors $V= [\mathbf{v}_1 \mathbf{v}_2 \cdots \mathbf{v}_d]$ are such that they describe the most important axis of the data in the following sense. The first eigenvector $\mathbf{v}_1$ describes which direction has the most variance. Then since $\mathbf{v}_2, \cdots, \mathbf{v}_d$ are each orthogonal to $\mathbf{v}_1$, this implies that $\mathbf{v}_2$is the direction (after $\mathbf{v}_1$ has been factored out) that has the most variance.The rank $k$ approximation $A_k \in \mathbb{R}^{n\times d}$ of $A$ is as follows: $$A_k = U_k S_k V_k^{\top},$$where $S_k$ is still an $\mathbb{R}^{n\times d}$ matrix, but only with the first $k$ entries being nonzero. In this way, only first $k$ columns of $V$ ($k$ rows of $V^{\top}$) contribute to $A_k$.---- Example of GaussLet us recall in the beginning of this quarter, and load `Gauss.jpg`, then flatten its color dimension to make a grayscale image.Reference: [https://docs.scipy.org/doc/numpy-1.15.1/reference/generated/numpy.linalg.svd.html](https://docs.scipy.org/doc/numpy-1.15.1/reference/generated/numpy.linalg.svd.html) ###Code G = plt.imread("gauss.jpg") G_bw = np.mean(G, axis=2) plt.imshow(G_bw, cmap="gray") U, S, VT = LA.svd(G_bw) S_mat = np.zeros_like(G_bw) S_mat[:600] = np.diag(S) np.allclose(G_bw, U.dot(S_mat.dot(VT))) ###Output _____no_output_____ ###Markdown Remark: Or you can use `@`, where the `@` (at) operator is intended to be used for matrix multiplication. This is New in version 3.5. ###Code # suppose we only use first k singular values k = 30 S_k = np.zeros_like(G_bw) S_k[:k, :k] = np.diag(S[:k]) G_k = (U @ S_k) @ VT plt.imshow(G_k, cmap="gray") ###Output _____no_output_____ ###Markdown Why can we discard the smaller singular values?Let us use the following artificially generated data as an example. ###Code num_samples = 500 mean = np.array([0,0]) covariance = np.array([[12, 9], [ 9, 10]]) X = np.random.multivariate_normal(mean, covariance, num_samples) plt.scatter(X[:,0], X[:,1], alpha=0.2) U, S, VT = LA.svd(X) VT[0,:] # first eigenvectors VT[1,:] # second eigenvectors v1 = VT[0,:] * S[0]/20 v2 = VT[1,:] * S[1]/20 plt.figure(figsize=(6,6)) plt.axis([-8,8,-8,8]) plt.scatter(X[:,0], X[:,1], alpha=0.2) plt.arrow(0, 0, v1[0], v1[1], width=0.1, head_width=0.34, head_length=0.8) plt.arrow(0, 0, v2[0], v2[1], width=0.1, head_width=0.34, head_length=0.8) ###Output _____no_output_____ ###Markdown What does the "significance" of a direction mean in data science?Consider using SVD on the dataset matrix $X\in \mathbb{R}^{N\times d}$, where $X$'s $i$-th row corresponds to the $i$-th data sample. There are $N$ data samples and each data point has $n$ dimensions (features). If the data set is normalized row-wise, i.e., each row (data sample) has mean zero. The significant directions (columns of $V$) are the eigenvectors of the covariance matrix ${\frac{1}{N-1}X^T X}$. The singular values (entries of $S$) are the square roots of the eigenvalues of this covariance matrix.In other words, $S$'s entries are the amount of deviations of the data in that direction. The covariance matrix tells you how correlated each direction is with other directions in the space where the dataset lives. In-class Exercise: MNISTPCA can be used to speed up a machine learning algorithm (logistic regression) on the MNIST dataset.Download `mnist_binary_train.npz` and `mnist_binary_test.npz` on Canvas files tab, and load them using the following cell. Then try the following:* Use `scikit-learn`'s `LogisticRgression` class on the original dataset.* Import `PCA` from `scikit-learn`'s `decomposition` submodule, apply it on the dataset, re-run the `LogisticRegression` on the reduced dataset. ###Code data_train = np.load('mnist_binary_train.npz') data_test = np.load('mnist_binary_test.npz') X_train, y_train = data_train['X'], data_train['y'] X_test, y_test = data_test['X'], data_test['y'] # use scikit-learn's built-in logistic regression from sklearn.linear_model import LogisticRegression import time # measure the time mnist_reg = LogisticRegression(solver='lbfgs', max_iter=500) starting_time = time.process_time() mnist_reg.fit(X_train,y_train) print("Data fitting takes", time.process_time() - starting_time, "seconds") mnist_reg.score(X_test,y_test) ###Output _____no_output_____ ###Markdown Use PCA to reduce the dimensionFrom the [reference of the `PCA`](https://scikit-learn.org/stable/modules/generated/sklearn.decomposition.PCA.html):> If `0 < n_components < 1` and `svd_solver == 'full'`, select the number of components such that the amount of variance that needs to be explained is greater than the percentage specified by n_components.The following cell means that the `PCA` class in `scikit-learn` will choose the minimum number of principal components such that 95% of the variance is retained for our data. ###Code from sklearn.decomposition import PCA mnist_pca = PCA(n_components=0.95) mnist_pca.fit(X_train) X_train_reduced = mnist_pca.transform(X_train) X_test_reduced = mnist_pca.transform(X_test) mnist_reg_pca = LogisticRegression(solver='lbfgs', max_iter=500) starting_time = time.process_time() mnist_reg_pca.fit(X_train_reduced,y_train) print("Data fitting takes", time.process_time() - starting_time, "seconds") mnist_reg_pca.score(X_test_reduced,y_test) ###Output _____no_output_____
solutions/Practical_4.ipynb
###Markdown Practical 4: Modules and Functions - Building Conway's Game of LifeObjectives: In this practical we continue to use functions, modules and conditional statements. We also continue practicing how we access entries from 2D arrays. At the end of this notebook you will have a complete version of Conway's Game of Life which will produce an animation. This will be done through 3 different sections, each of which has an exercise for you to complete: - 1) [Creating different shapes through 2D Numpy array modifications](Part1) * [Exercise 1: Draw still 'life' from Conway's Universe](Exercise1) * [Exercise 2: Draw oscillators and space-ship 'life' from Conway's Universe](Exercise2) - 2) [Creating a function that searches a local neighbourhood for values of '1' and '0'](Part2) * [Exercise 3: Implement the 4 rules of life](Exercise3) * [Exercise 4: Loop through 20 oscillations of the 'Beacon' lifeform](Exercise4) - 3) [Populating Conway's Universe with multiple species](Part3) As with our other notebooks, we will provide you with a template for plotting the results. Also please note that you should not feel pressured to complete every exercise in class. These practicals are designed for you to take outside of class and continue working on them. Proposed solutions to all exercises can be found in the 'Solutions' folder. Please note: After reading the instructions and aims of any exercise, search the code snippets for a note that reads -------'INSERT CODE HERE'------- to identify where you need to write your code Introduction: The gameBefore we get our teeth into the exercises included in this notebook, let's remind ourselves about the basis for Conway's game of life. In Conway's game of life, the Universe is represented as a 2D space [a 2D Numpy array in our case!] on which each cell can either be alive or dead. If we refer to each cell as having one of two states, we can represent this numerically as each cell having either a value of 1 or 0. If we then assume we can draw 2D shapes that represent a 'specie', as a collection of live cells, we might find patterns changing over time.Every cell interacts with its neighbours, whether they are horizontally, vertically of diagonally adjacent. There are 4 laws that define these interactions: - Any live cell with fewer than two live neighbours dies, as if by underpopulation. - Any live cell with two or three live neighbours lives on to the next generation. - Any live cell with more than three live neighbours dies, as if by overpopulation. - Any dead cell with exactly three live neighbours becomes a live cell, as if by reproduction.So, imagine we are at the beginning of time in our 2D Universe. We need to check the status of every cell and make changes according to these laws. After one sweep through our 2D space, or time step, the status of individual cells will change. Numerically, the distribution of '1's and '0's change across our 2D space. In fact, by defining species as dinstinct groups of cells of a certain shape, as we move through multiple time steps we find 3 types of patterns emerging: - Still life: These patterns remain fixed on the board - Oscillators: These patterns change shape on every iteration, but return to their initial state after a number of generations. - Space-ships: These patterns end up moving across the board according to the rules that define life and death.From a programming perspective, implementing these rules through any number of time steps requires a number of procedures to be implemented via code: - 1) Defining 2D arrays that represent species in Conway's Universe. - 2) Creating a function that searches the immediate neighbouring space of each cell for 1's and 0's. - 3) Counting the number of 1's and 0's according to the previous point. - 4) Changing the values of each cell according to the 4 laws stated above. - 5) Looping through points 2-4 for any number of time steps.By sequentially following the proceeding exercises, we will eventually build a variant of Conway's game of life. Creating different shapes through 2D Numpy array modifications Before we can run a simulation, let's create distinct species as groups of cells, and thus patterns. This will help us practice creating 2D arrays and populating each cell with either a '0' or '1' depending on what pattern we want to draw. To generate and thus draw each specie you will be asked to initialise a 2D Numpy array that repeats the pattern seen in the picture. The code to plot, thus visualise, each pattern is given for you. Still life The pictures in Figure 1 and 2 illustrate common types of still life in Conway's Universe. Ive given you some code that reproduces the pattern for 'Block', in the code box below. Read through the code and comments and see if this makes sense. ![](images/Practical_3_figure1.png "Title") Figure 1![](images/Practical_3_figure2.png "Title") Figure 2 ###Code #%matplotlib inline #this is to help us retrieve those love animations! import numpy as np #import the numerical python library, numpy. Changing the referenced library to 'np' is solely for convenience import matplotlib.pyplot as plt #as per the above, much easier to write over and over again from matplotlib import animation, rc # Lets first create our 'Block'. Dont forget, we can call our arrays and matrices anything we want. In this case Im going to use the name of the pattern we are interested in Block = np.zeros((4,4),dtype=int) #Im telling the Python interpreter I want a numpy array that is 4 rows by 4 columns, contains '0' for now and is expecting my data to be of integer type # What does this look like? print("An empty array",Block) # Can you see a matrix of 0s? # Ok cool. Now lets add some black cells by position some values of 1. For the Block pattern, this is done as follows: Block[1,1]=1 Block[1,2]=1 Block[2,1]=1 Block[2,2]=1 # Remeber how we refer to elements in an array in Python? Everything starts at 0, so here im filling in the central 2x2 matrix with 1s. Lets check this out numerically: print(print("A finished array",Block)) #Now lets plot this to recreate the patterns given in figure x. plt.imshow(Block, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Block') plt.show() ###Output An empty array [[0 0 0 0] [0 0 0 0] [0 0 0 0] [0 0 0 0]] A finished array [[0 0 0 0] [0 1 1 0] [0 1 1 0] [0 0 0 0]] None ###Markdown Exercise 1: Draw still 'life' from Conway's Universe. In this exercise you will need to create a 2D Numpy array that essentially draws both the *Tub* and *Boat* specie from figure 2. ###Code # We have already imported both Numpy and Matplotlib so no need to import those again. # Initialise our matrices Tub = np.zeros((5,5),dtype=int) Boat = np.zeros((5,5),dtype=int) #-------'INSERT CODE HERE'------- # Now add '1's to the currently empty 2D array Tub Tub [1,2]=1 Tub [2,1]=1 Tub [3,2]=1 Tub [2,3]=1 #-------------------------------- plt.subplot(1, 2, 1).imshow(Tub, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Tub') #plt.show() #-------'INSERT CODE HERE'------- # Now add '1's to the currently empty 2D array Boat Boat [1,1]=1 Boat [1,2]=1 Boat [2,1]=1 Boat [2,3]=1 Boat [3,2]=1 #-------------------------------- plt.subplot(1, 2, 2).imshow(Boat, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Boat') plt.show() ###Output _____no_output_____ ###Markdown Exercise 2: Draw oscillators and space-ship 'life' from Conway's Universe Following exercise 1,now do the same for 2 types of both *oscillators* and *space ships*: Toad, Beacon, Glider and Light-weight spaceship (LWSS). Can you replicate the patterns shown in figures 1 and 3? Check the size of each array you need, accounting for white space around the outside. Use the space below and copy-paste the code we have already used.![](images/Practical_3_figure3.png "Title") Figure 3 ###Code #Enter the Python code here to create and then visualise a Toad, Beacon and Glider #Initialise each matrix Beacon = np.zeros((6,6),dtype=int) Toad = np.zeros((6,6),dtype=int) Glider = np.zeros((5,5),dtype=int) LWSS = np.zeros((6,7),dtype=int) #Enter values for '1' where you would like a black square #-------'INSERT CODE HERE'------- Beacon [1,1]=1 Beacon [1,2]=1 Beacon [2,1]=1 Beacon [3,4]=1 Beacon [4,3]=1 Beacon [4,4]=1 Toad [2,2]=1 Toad [2,3]=1 Toad [2,4]=1 Toad [3,1]=1 Toad [3,2]=1 Toad [3,3]=1 Glider [1,2]=1 Glider [2,3]=1 Glider [3,1]=1 Glider [3,2]=1 Glider [3,3]=1 LWSS [1,2]=1 LWSS [1,5]=1 LWSS [2,1]=1 LWSS [3,1]=1 LWSS [4,1]=1 LWSS [4,2]=1 LWSS [4,3]=1 LWSS [4,4]=1 LWSS [3,5]=1 #-------------------------------- #Now visualise your results. plt.subplot(1, 2, 1).imshow(Beacon, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Beacon') plt.subplot(1, 2, 2).imshow(Toad, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Toad') plt.show() plt.subplot(1, 2, 1).imshow(Glider, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Glider') plt.subplot(1, 2, 2).imshow(LWSS, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('LWSS') plt.show() ###Output _____no_output_____ ###Markdown Creating a function that searches a local neighbourhood for values of '1' and '0' Now we know how to define a specie by modifying values in a 2D array, we also need to now create a function that can search the neighbouring space of any cell for the occurance of '1's or '0's. We are going to perform this operation many times so creating a function to do this seems a sensible approach.As an example, let's re-create the 2D array that represents the specie 'Beacon' and then pass this array into a new function that will search the neighbouring space of every cell to detect a '1' or '0. In this example I have given you all of the code to perform this operation. Try to understand the syntax used. Does this make sense? First look at the code and then let's formulate the steps in the function as a narrative. ###Code #Initialise the Beacon matrix Beacon = np.zeros((6,6),dtype=int) #Enter values for '1' where you would like a black square Beacon [1,1]=1 Beacon [1,2]=1 Beacon [2,1]=1 Beacon [3,4]=1 Beacon [4,3]=1 Beacon [4,4]=1 # Now define a function that moves through through each cell in our 2D array and searches the neighbouring space # We pass three variables: # rows - Number of rows in our space to be searched # cols - Number of columns in our space to be searched # space - The 2D array space to be searched def search_each_cell(total_rows,total_cols,space): # 1) First, we need to start moving through each cell of our 'space'. # To do this, we will use two nested 'for' loops for row in range(total_rows): for col in range(total_cols): # So 'row' and 'col' define our current cell. # We now need to search a neighbourhood defined as 1 cell distance around this position # We thus need two more nested for loops. When searching this neighbouring space, we want # to count the number of 1's. Thus we also need a variable that we can increment by 1 # everytime we find a value of 1. Lets call this integer variable count count = 0 for row2 in range(row-1,row+2): # See here that we can define a start and end to our 'range' for col2 in range(col-1,col+2): # We need to check if our new position, defined by [row2,col2] is off the board if (row2<0) or (row2>=total_rows) or (col<0) or (col2>=total_cols): # Do nothing pass elif row2 == row and col2 == col: # Do nothing, its the cell we already have! pass # If we are not off the board or in the same cell as our starting point... # We can check if this new space has a value of 1. It it does, lets count it else: if space[row2,col2]>0: count=count+1 return # At the moment we are not returning anything. Seem odd? We will get back to this. # call the above function search_each_cell(6,6,Beacon) print("Finished function call, nothing to report!") ###Output Finished function call, nothing to report! ###Markdown Now let's try to understand what this function is actually doing. As an algorithm, we have the following steps - 1) Pass the 2D Numpy array to the new function along with variables that define the total number of rows and columns - 2) We need to move through every cell and search its local neighbourhood. Moving through each cell is defined by the first two loops that cycle through both the row and column index of our 2D space. The limits are defined by the variables total_rows and total_cols - 3) For each cell, we will want to have an integer variable that counts how many 1's there are in the local neighborhood. We need to initialise this to 0 for each cell we move through. We call this variable count - 4) Now we need to look at the local space surrounding our cell. For this we need two more nested loops that look 1 row above, 1 row below, 1 column to the left and one to the right. - 5) As we move through this neighborhood we need to check if we are either off the board OR in the same location as the cell we are interested in! - 6) If none of the above is true, then check if a cell has a value greater then 0. If it does, increment variable count by 1. - 7) For each cell on the board, repeat steps 3-6. - 8) When the entire space has been searched, stop the function and return nothing. Exercise 3 - Implement the 4 rules of life Now we have the function that can search the local neighbourhood of any cell and count how many 1s and 0's there are, we can now add on more code that can implement the 4 rules of life and thus keep the value of our current cell or change it. Let's remind ourselves what those rules are: - Any live cell with fewer than two live neighbours dies, as if by underpopulation. - Any live cell with two or three live neighbours lives on to the next generation. - Any live cell with more than three live neighbours dies, as if by overpopulation. - Any dead cell with exactly three live neighbours becomes a live cell, as if by reproduction.So in this exercise we have a shape that has been passed into our function, and then create a new shape according to the rules of life. In the exercise you will need to add a series of conditional statements that populate the value of cells in our new shape according to these rules. In other words, we can re-write the above rules as: - If our current cell is alive [=1]: a) If count < 2, current cell = 0 [it dies]. b) If 2<=count<=3, current cell = 1 [it stays alive]. c) If count>3, current cell = 0 [it dies] - If our current cell is dead [=0] a) If count == 3, current cell = 1 [born] Notice the syntax I have used for the last conditional: If count == 3 ? When checking a value we use two equals signs == as we are not *assigning* a value as we would in, e.g. x = 4 . In the code snippet below, I have identified where you need to implement these rules. Notice that we plot the 'Beacon' pattern before we call the function and then the new 2D space which should change the pattern. With this in mind, also note that our function new returns a new version of our 2D space which I have called 'new_space'. If correct, when you run your completed code you should see figure 4.![](images/Practical_3_figure4.png "Title") Figure 4Please note that where I have added 'INSERT CODE HERE' we are using the correct indentation. ###Code #Initialise the Beacon matrix Beacon = np.zeros((6,6),dtype=int) #Enter values for '1' where you would like a black square Beacon [1,1]=1 Beacon [1,2]=1 Beacon [2,1]=1 Beacon [3,4]=1 Beacon [4,3]=1 Beacon [4,4]=1 # Now define a function that moves through through each cell in our 2D array and searches the neighbouring space # We pass three variables: # rows - Number of rows in our space to be searched # cols - Number of columns in our space to be searched # space - The 2D array space to be searched def search_each_cell(total_rows,total_cols,space): new_space = np.zeros((total_rows,total_cols),dtype=int) # 1) First, we need to start moving through each cell of our 'space'. # To do this, we will use two nested 'for' loops for row in range(total_rows): for col in range(total_cols): # So 'row' and 'col' define our current cell index. # We now need to search a neighbourhood defined as 1 cell distance around this position # We thus need two more nested for loops. When searching this neighbouring space, we want # to count the number of 1's. Thus we also need a variable that we can increment by 1 # everytime we find a value of 1. Lets call this integer variable count. count = 0 for row2 in range(row-1,row+2): # See here that we can define a start and end to our 'range' for col2 in range(col-1,col+2): # We need to check if our new position, defined by [row2,col2] is off the board if (row2<0) or (row2>=total_rows) or (col<0) or (col2>=total_cols): # Do nothing pass elif row2 == row and col2 == col: # Do nothing, its the cell we already have! pass # If we are not off the board or in the same cell as our starting point... # We can check if this new space has a value of 1. It it does, lets count it else: if space[row2,col2]>0: count=count+1 #-------'INSERT CODE HERE'------- # Here you need to introduce conditional statements that act on the value of 'count' # Read through the narrative provided above and remember to obey the spacing rules # You will need to check the value of space[row,col] and then, depending on whether # this is greater than 0 OR equals to 0, implement the rules of life. I have provided # the first example. Please do try to complete this. if space[row,col] > 0: if count < 2: new_space[row,col] = 0; elif 2<=count<=3: new_space[row,col] = 1; elif count > 3: new_space[row,col] = 0; elif space[row,col] == 0: if count == 3: new_space[row,col] = 1; #-------------------------------- return new_space # call the above function Beacon_new = search_each_cell(6,6,Beacon) print("Finished function call, now lets compare our pattern before and after...") #Now visualise your results. plt.subplot(1, 2, 1).imshow(Beacon, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Beacon - before') plt.subplot(1, 2, 2).imshow(Beacon_new, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Beacon - after') plt.show() ###Output Finished function call, now lets compare our pattern before and after... ###Markdown Exercise 4 - Loop through 20 oscillations of the 'Beacon' lifeform Now that we have build the function that can implement the 4 rules of life, all that is left for us to do is to call this function a set number of times to simulate evolution across our Universe. In the code box below, drop your conditional statements from above in the relevant place and click 'Run'. Do you see the Beacon shape oscillating? As before, I have provided the code for plotting but see if the syntax makes sense. ###Code import numpy as np #import the numerical python library, numpy. Changing the referenced library to 'np' is solely for convenience import matplotlib.pyplot as plt #as per the above, much easier to write over and over again from matplotlib import animation, rc from IPython.display import HTML from IPython.display import clear_output import time #Initialise the Beacon matrix Beacon = np.zeros((6,6),dtype=int) #Enter values for '1' where you would like a black square Beacon [1,1]=1 Beacon [1,2]=1 Beacon [2,1]=1 Beacon [3,4]=1 Beacon [4,3]=1 Beacon [4,4]=1 # Now define a function that moves through through each cell in our 2D array and searches the neighbouring space # We pass three variables: # rows - Number of rows in our space to be searched # cols - Number of columns in our space to be searched # space - The 2D array space to be searched def search_each_cell(total_rows,total_cols,space): new_space = np.zeros((total_rows,total_cols),dtype=int) # 1) First, we need to start moving through each cell of our 'space'. # To do this, we will use two nested 'for' loops for row in range(total_rows): for col in range(total_cols): # So 'row' and 'col' define our current cell index. # We now need to search a neighbourhood defined as 1 cell distance around this position # We thus need two more nested for loops. When searching this neighbouring space, we want # to count the number of 1's. Thus we also need a variable that we can increment by 1 # everytime we find a value of 1. Lets call this integer variable count count = 0 for row2 in range(row-1,row+2): # See here that we can define a start and end to our 'range' for col2 in range(col-1,col+2): # We need to check if our new position, defined by [row2,col2] is off the board if (row2<0) or (row2>=total_rows) or (col<0) or (col2>=total_cols): # Do nothing pass elif row2 == row and col2 == col: # Do nothing, its the cell we already have! pass # If we are not off the board or in the same cell as our starting point... # We can check if this new space has a value of 1. It it does, lets count it else: if space[row2,col2]>0: count=count+1 #-------'INSERT CODE HERE'------- # Here you need to introduce conditional statements that act on the value of 'count' # Read through the narrative provided above and remember to obey the spacing rules if space[row,col] > 0: if count < 2: new_space[row,col] = 0; elif 2<=count<=3: new_space[row,col] = 1; elif count > 3: new_space[row,col] = 0; elif space[row,col] == 0: if count == 3: new_space[row,col] = 1; #-------------------------------- return new_space fig, ax2 = plt.subplots() plt.imshow(Beacon, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Beacon oscillating') plt.show() # Let us call the function 20 times # Each time we are given a new shape to plot on our figure. # Wait 0.2 seconds before moving on to thje next iteration # We shpuld see oscillating behaviour. for x in range(20): clear_output(wait=True) Beacon_new = search_each_cell(6,6,Beacon) Beacon = Beacon_new plt.imshow(Beacon_new, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Beacon oscillating') plt.show() time.sleep(0.2) ###Output _____no_output_____ ###Markdown Populating Conway's Universe with multiple species Now we are going to use the defintion of our shapes to populate a miniature Universe in 2D space! Once we have this, following the same procedure as above, we should see some interesting movement! So let's create a space that is big enough for all of our cell types. To do this, we need to create another matrix: ###Code Universe=np.zeros((50,50),dtype=int) print(Universe) ###Output [[0 0 0 ... 0 0 0] [0 0 0 ... 0 0 0] [0 0 0 ... 0 0 0] ... [0 0 0 ... 0 0 0] [0 0 0 ... 0 0 0] [0 0 0 ... 0 0 0]] ###Markdown You should now see a snapshot of the Universe matrix that is empty. How do we populate our Universe with individual species? We could enter a value for each cell but this is laborious. Rather, we are going to use our existing matrices that define our species and place them on the Universe grid. We do that by definining the exact space in the Universe we want our cells to go. This is practice in recognising the correct shape of an array/matrix and matching one to another. For example, look at the code below which places the top left corner of an LWSS on the cell in the 12th row and 13th column of my Universe and then visualises the results. Dont forget, indexing in Python starts at 0 so for the 12th row and 13th column, I need to refer to element [11,12]. Im also using the operator : which allows us to straddle cells bound by a start and a finish. Why have I chosen the range given below? Feel free to change the values, but if you get the size of space needed to fit in an LWSS, Python will complain it cannot broadcast a given shape: ###Code #Define the space in the Universe you would like your LWSS to appear Universe[11:17,12:19] = LWSS #Now visualise our Universe plt.imshow(Universe, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Universe [with 1 LWSS]') plt.show() ###Output _____no_output_____ ###Markdown To finish this notebook, in the following code box we fill the Universe with a range of species and then run a simulation. Can you see how we have mapped species shapes into our Universe? It is left for you to copy the working function 'search_each_cell' from above to complete the simulation.Have a play with this! What happens if you increase the number of iterations to 300? Please note, we might want to clear our Universe from the above exercise, in which case we could write: Universe[:,:]=0, but let's keep it in for now. ###Code #Define the space in the Universe you would like your different species to appear Universe[30:36,32:39] = LWSS Universe[11:17,12:19] = LWSS Universe[22:28,12:18] = Beacon Universe[33:39,2:8] = Beacon Universe[19:25,32:38] = Toad Universe[1:6,1:6] = Glider Universe[6:11,25:30] = Boat plt.imshow(Universe, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Universe [with multiple cell types]') plt.show() #-------'INSERT CODE HERE'------- def search_each_cell(total_rows,total_cols,space): new_space = np.zeros((total_rows,total_cols),dtype=int) # 1) First, we need to start moving through each cell of our 'space'. # To do this, we will use two nested 'for' loops for row in range(total_rows): for col in range(total_cols): # So 'row' and 'col' define our current cell index. # We now need to search a neighbourhood defined as 1 cell distance around this position # We thus need two more nested for loops. When searching this neighbouring space, we want # to count the number of 1's. Thus we also need a variable that we can increment by 1 # everytime we find a value of 1. Lets call this integer variable count count = 0 for row2 in range(row-1,row+2): # See here that we can define a start and end to our 'range' for col2 in range(col-1,col+2): # We need to check if our new position, defined by [row2,col2] is off the board if (row2<0) or (row2>=total_rows) or (col<0) or (col2>=total_cols): # Do nothing pass elif row2 == row and col2 == col: # Do nothing, its the cell we already have! pass # If we are not off the board or in the same cell as our starting point... # We can check if this new space has a value of 1. It it does, lets count it else: if space[row2,col2]>0: count=count+1 # Here you need to introduce conditional statements that act on the value of 'count' # Read through the narrative provided above and remember to obey the spacing rules if space[row,col] > 0: if count < 2: new_space[row,col] = 0; elif 2<=count<=3: new_space[row,col] = 1; elif count > 3: new_space[row,col] = 0; elif space[row,col] == 0: if count == 3: new_space[row,col] = 1; return new_space #-------------------------------- fig, ax2 = plt.subplots(figsize=(12, 12)) plt.imshow(Universe, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Universe simulation') plt.show() for x in range(100): clear_output(wait=True) Universe_new = search_each_cell(50,50,Universe) Universe = Universe_new plt.imshow(Universe_new, cmap='binary') #The cmap, or colour map, gives us a black and white board. plt.title('Universe simulation') plt.show() time.sleep(0.2) ###Output _____no_output_____
Assignments/Assignment_2_final.ipynb
###Markdown Second Assignment 1) Create a function called **"even_squared"** that receives an integer value **N**, and returns a list containing, in ascending order, the square of each of the even values, from 1 to N, including N if applicable. ###Code def even_squared(N): '''returns a list of squared even values from 1 to N''' my_range = range(1, N) squares = [value**2 for value in my_range if value%2 ==0] return (squares) even_squared(5) ###Output _____no_output_____ ###Markdown 2) Using a while loop and the **input()** function, read an indefinite amount of **integers** until the number read is **-1**. After this process, print two lists on the screen: The first containing the even integers, and the second containing the odd integers. Both must be in ascending order. ###Code even_list = [] odd_list = [] while True: i = int(input("Type in an integer: ")) if i == -1: break if i%2 ==0: even_list.append(i) else: odd_list.append(i) print(sorted(even_list)) print(sorted(odd_list)) ###Output Type in an integer: 2 Type in an integer: 4 Type in an integer: 5 Type in an integer: -1 ###Markdown 3) Create a function called **"even_account"** that receives a list of integers, counts the number of existing even elements, and returns this count. ###Code def even_account(l): '''count the number of given even numbers''' list1 = len([num for num in l if num%2 ==0]) return list1 my_list = [4, 5, 2, 12] even_account(my_list) ###Output _____no_output_____ ###Markdown 4) Create a function called **"squared_list"** that receives a list of integers and returns another list whose elements are the squares of the elements of the first. ###Code def square_list(lis): '''return list of swaures of all the elements in the given list''' squares = [value **2 for value in lis] return squares my_test_list = [1, 3] square_list(my_test_list) ###Output _____no_output_____ ###Markdown 5) Create a function called **"descending"** that receives two lists of integers and returns a single list, which contains all the elements in descending order, and may include repeated elements. ###Code def descending(list1, list2): '''merge the elements of two lists in decending order''' list3 = (list1 +list2) list3.sort(reverse=True) return (list3) my_first = [2, 4, 3] my_second = [5, 7, 3] descending(my_first, my_second) ###Output _____no_output_____ ###Markdown 6) Create a function called **"adding"** that receives a list **A**, and an arbitrary number of integers as input. Return a new list containing the elements of **A** plus the integers passed as input, in the order in which they were given. Here is an example: >```python>>>> A = [10,20,30]>>>> adding(A, 4, 10, 50, 1)> [10, 20, 30, 4, 10, 50, 1]``` ###Code def adding(A, *num): '''combine the received list and the arbitrary number of integers to a new list in the order that was given''' numb = [value for value in num] my_list = A + numb print(my_list) A = [1, 2, 3] adding(A, 4,5) ###Output [1, 2, 3, 4, 5] ###Markdown 7) Create a function called **"intersection"** that receives two input lists and returns another list with the values that belong to the two lists simultaneously (intersection) without repetition of values and in ascending order. Use only lists (do not use sets); loops and conditionals. See the example: >```python>>>> A = [-2, 0, 1, 2, 3]>>>> B = [-1, 2, 3, 6, 8]>>>> intersection(A,B)> [2, 3]``` ###Code def intersection(list1, list2): '''receive 2 lists and return one list with the intersection, without duplicates in ascending order''' my_intersection = [element for element in list1 if element in list2] return (my_intersection) A = [1, 2, 3] B = [2, 3, 4] intersection(A, B) ###Output _____no_output_____ ###Markdown 8) Create a function called **"union"** that receives two input lists and returns another list with the union of the elements of the two received, without repetition of elements and in ascending order. Use only lists (do not use sets); loops and conditionals. See the example: >```python>>>> A = [-2, 0, 1, 2]>>>> B = [-1, 1, 2, 10]>>>> union(A,B)> [-2, ,-1, 0, 1, 2, 10]``` ###Code def union(list1,list2): '''receive two input lists, return new list with union of elements without duplicates in ascending order''' for e in list2: if e not in list1: list1.append(e) list1.sort() return list1 A = [1, 5, 3] B = [2, 3, 1] union(A, B) ###Output _____no_output_____ ###Markdown 9) Generalize the **"intersection"** function so that it receives an indefinite number of lists and returns the intersection of all of them. Call the new function **intersection2**. ###Code A = [1, 2, 3] B = [2, 3, 4] C = [6, 7, 4] def intersection2(*lists): '''generalised intersection function: intersection of provided list elements''' my_intersection = [] my_intersection2 = [] lists_together = [part for part in lists] for myList in lists_together: for item in myList: my_intersection.append(item) for i in my_intersection: if i not in my_intersection2: my_intersection2.append(i) return (my_intersection2) intersection2(A, B, C) ###Output _____no_output_____
src/ipynbfiles/run.ipynb
###Markdown ###Code # # # TODO: # - read file data TIK # - delete enters TIK # - read subtitle folders list # - split by " " and search in TIK # - shomaresh TIK # - save dar csv # - sort list by counting TIK # - remove duplicate TIK # # # - Optional : 1 july 2021 # - write it to the picture for reading better or printing # # # import csv # 4/1AX4XfWjgyO89bswaF-OiIEHWet7wTPljmthD6YmmhelJG6x_frolThpIxzY from google.colab import drive drive.mount('/content/drive') # this block remove blank lines and replace enters with space def clean_file(filename): lines = "" with open(filename) as f: for line in f: if not line.strip(): continue # skip the empty line lines += line return lines.replace("\n", " ") #test and track lines = clean_file("/content/drive/MyDrive/dataset.txt") # word filters common_prepositions = "a the about above across after against among around at before behind below beside between by down during for from in inside into near of off on out over through to toward under up with" common_prepositions = common_prepositions.split() less_comman_prepositions = "aboard along amid as beneath beyond but concerning considering despite except following like minus next onto opposite outside past per plus regarding round save since than till underneath unlike until upon versus via within without" less_comman_prepositions = less_comman_prepositions.split() full_perps = common_prepositions+less_comman_prepositions # define counter number of repeat every string def count_repeats(input_string): countdict = [] words = input_string.split() # list of words in a input string for i in range(len(words)): # print(i) # print(lines.split()[i]) countdict.append([words[i],input_string.count(words[i])]) return countdict # duplicator remover get a list for input words = count_repeats(lines) def remove_duplicates(wordslist): res = [] [res.append(x) for x in words if x not in res] return res words = remove_duplicates(words) len(words) indexwords = [] [indexwords.append(x[0]) for x in words] len(indexwords) # remove useless words junk_list = [] [junk_list.append(x) for x in indexwords if "[" in x and "]" in x or len(x) < 3 or x in full_perps] print(len(words)) [words.remove(x) for x in words if x[0] in junk_list] print(len(words)) # sort by repeats words.sort(key=lambda x:x[1], reverse=True) words #save results in a csv file def save_to_csv(): fields = ['words', 'repeats'] with open('result.csv', 'w') as f: # using csv.writer method from CSV package write = csv.writer(f) write.writerow(fields) write.writerows(words) save_to_csv() ###Output _____no_output_____
src/Topics/Notes/modules.ipynb
###Markdown ###Code !pip install ExifRead import exifread # Open image file for reading (binary mode) f = open("foto_exif.jpeg", 'rb') # Return Exif tags tags = exifread.process_file(f) for tag in tags.keys(): if tag not in ('JPEGThumbnail', 'TIFFThumbnail', 'Filename', 'EXIF MakerNote'): print(f"Key: {tag}, value {tags[tag]}") !pip install requests3 from random import randint import requests # url = f'http://numbersapi.com/{randint(1,2019)}/year' # r = requests.get(url) # print(r.text, url) class Number: url = 'http://numbersapi.com' types = ['trivia', 'math', 'year', 'date'] def __init__(self, number): self.number = number self.facts = {} for type in self.types: self.facts[type] = self.get_fact(type) def get_fact(self, type): url = f'{self.url}/{self.number}/{type}' r = requests.get(url) return r.text def trivia(self): if not hasattr(self, "_trivia"): self._trivia = self.get_fact('trivia') return self._trivia def year(self): if not hasattr(self, "_year"): self._year = self.get_fact('year') return self._year n1945 = Number(randint(1,2019)) print(n1945.year()) print(n1945.trivia()) from requests import get class Weather: url_base = 'http://api.openweathermap.org/data' version = 2.5 forecast_endpoint = 'forecast/daily' api_key = '51b03e8d15d7df69f88921b3c1ba2f69' @staticmethod def convert(k): c = k - 273.15 return 9*c/5 + 32 def __init__(self, city, country, num_days=16): if num_days < 1 or num_days > 16: raise Exception('num_days must be between 1 and 16!') self.city = city self.country = country self.days = num_days def api_call(self): args = f'q={self.city},{self.country}&APPID={self.api_key}&cnt={self.days}' url = f'{self.url_base}/{self.version}/{self.forecast_endpoint}?{args}' r = get(url) return r.json() def get_weather(self, day): if day < 1 or day > self.days: raise Exception('days must be between 1 and 16!') data = self.api_call() return self.convert(data['list'][day-1]['temp']['day']) nyc = Weather('new york', 'usa', 2) print(nyc.get_weather(2)) ###Output 83.03000000000004 ###Markdown Modules & PackagesIn Python, a `module` is Python source file that contains pre-defined objects like variables, functions, classes, and other items we'll talk about soon. A Python `package`, sometimes used synonymously with the term `library`, is simply a collection of Python modules. The diagram below can show you this hierarchy visually.![package_def](https://365datascience.com/wp-content/uploads/2018/07/image2-min-6-768x419.png)Essentially, packages and modules are a means of `modularizing` code by grouping functions and objects into specific areas of focus. For instance, the `statsmodels` module ([here](https://www.statsmodels.org/)) contains code useful to a data scientist. The `Pyglet` library ([here](http://www.pyglet.org/)) contains code useful to game developers needing shortcuts for 3D game animation. But vice versa?`Modular programming` allows us to break out modules and packages dealing with specific topics in order make the standard library more efficient for the general public. It's sort of like "a la carte" code. This becomes especially valuable once you scale your programs. Who needs that extra baggage? Global vs. Local ScopeOne of the reasons Python leverages modular programming is because it helps avoid conflicts between `local` and `global` variables by creating separate `namespaces`. `Namespaces` are the place where variables are stored, and they exist on several independent levels, including **local, global, built-in, and nested namespaces**. For instance, the functions `builtin.open()` and `os.open()` are distinguished by their namespaces. Namespaces also aid readability and maintainability by making it clear which module implements a function. At a high level, a variable declared outside a function has `global scope`, meaning you can access a it inside or outside functions. A variable declared within a function has `local scope`, which means you can only access it within the object you created it. If you try to access it outside that, you will get a `NameError` telling you that variable is not defined.We'll get more into how to use and interpret local and global scope as we dive into modules and functions... Importing Modules & PackagesImporting modules and packages is very easy and saves you a lot of time you'd otherwise spend reinventing the wheel. Modules can even import other modules! The best practice is to place all import statements at the of your script file so you can easily see everything you've imported right at the top. Importing Modules Let's look at a few different way to import modules and their contents. The simplest way to import a module is to simply write `import module_name`. This will allow you to access all the contents within that module. If you want to easily find out exactly what is in your newly imported module, you can call the built-in function `dir()` on it. This will list all types of names: variables, modules, functions, etc. ###Code import math dir(math) # prints ['__doc__', '__file__', '__loader__', '__name__', '__package__', '__spec__', 'acos', 'acosh', 'asin', ... etc.] ###Output _____no_output_____ ###Markdown You can also import one specific object from a module like this: ###Code from math import sqrt sqrt(25) # 5 ###Output _____no_output_____ ###Markdown Notice how we included `math.` when we called the `sqrt` function. Because of *variable scope*, you need to reference the `namespace` where `sqrt` is defined. Simply importing `sqrt` does not give it `global scope`. It still has `local scope` within the math module.However, you can help avoid verbose code by importing modules and their items like this: ###Code from math import sqrt as s s(25) # 5 ###Output _____no_output_____ ###Markdown By importing the `sqrt as s`, you can call the function as `s()` instead of `math.sqrt`. The same works for modules. Note the difference in how we reference the square root function though... ###Code import math as m m.sqrt(25) # 5.0 ...we only renamed the module in this import and not the function. So we have to go back to the `module_name.function()` syntax. *However*, because we renamed the module on import, we can reference it in function calls by its shortened name, i.e. `m.sqrt`. ## Managing Dependencies In addition to "built-in" modules, we have the ability in python to create, distribute and most importantly *consume* community defined python modules. This is powerful because anyone who builds something useful has the ability to share with the larger python community. Creating and distributing python modules is outside the scope of this class, but we can consume any module we'd like by running the: pip install [module_name] Modules can be found in [**PyPI**](https://pypi.org/), or, the Python Package Index. Any registered module in pypi is installable via pip. However, in order to safely install modules across projects (ie: perhaps project A requires module 1 v1 but then project B, started a year later needs to use module 1 v2) we need to create what are called **virtual environments**, isolated python environments where we can safely install our pip modules and rest assured that they don't interfere with other projects / the system at lare. In order to create a virtual environment: python3 -m venv .env source .env/bin/activate The `.env` folder contains everything needed for this **"virtualenv"**. We go *inside* the env by running the `source ./env/bin/activate` command. To deactivate, (while in virtualenv): deactivate The best part about this is not only can we install our pip modules safely, we can also do this: pip freeze > requirements.txt This will collect all the installed pip modules in the virtual env and store into a file (that we are calling `requirements.txt`). This is useful because if we ever wanted to run this software from a different computer, all we would have to do is pull down the python files, create a new virtualenv and then: pip install -r requirements.txt ###Output _____no_output_____
Deploy Python bot on Messenger.ipynb
###Markdown Deploying Python bot on Facebook Messenger Let's say you have a chatbot written in Python and you want it available for many people to use. Why not deploy it on Facebook Messenger? I will show you below how you can make necessary set ups in this tutorial. Step 1. Installing neccessary modules and programYou need three main things to connect your Python code to Messenger:1. flask [link](http://flask.pocoo.org/)2. requests [link](https://2.python-requests.org//en/master/)3. ngrok [link](https://ngrok.com/product)You can install them by following the link above. If you don't have them already, you are mostly likely missing other modules required to run them. You need to install all the modules to run the Python code I will post below. Before that, let me explain the purpose of three main modules/program. `flask` is a module used to create local server in your computer that can receive messages from Messenger. `ngrok` is a program that will forward a http connection from Messenger to your computer. Lastly, `requests` is a function that will send the reply from your bot back to Messenger. Now, save the following Python code in `server.py` and put it inside a directory where your chatbot code is located. ###Code from flask import Flask, request import requests app = Flask(__name__) FB_API_URL = 'https://graph.facebook.com/v2.6/me/messages' VERIFY_TOKEN = ''# <paste your verify token here> PAGE_ACCESS_TOKEN = ''# paste your page access token here>" @app.route("/webhook",methods=['GET','POST'] ) def listen(): """This is the main function flask uses to listen at the `/webhook` endpoint""" if request.method == 'GET': return verify_webhook(request) if request.method == 'POST': payload = request.json event = payload['entry'][0]['messaging'] for x in event: if is_user_message(x): text = x['message']['text'] sender_id = x['sender']['id'] respond(sender_id, text) return "ok" def verify_webhook(req): if req.args.get("hub.verify_token") == VERIFY_TOKEN: return req.args.get("hub.challenge") else: return "incorrect" def respond(sender, message): """Formulate a response to the user and pass it on to a function that sends it.""" response = get_bot_response(message) send_message(sender, response) def get_bot_response(message): """This is just a dummy function, returning a variation of what the user said. Replace this function with one connected to chatbot.""" return "This is a dummy response to '{}'".format(message) def is_user_message(message): """Check if the message is a message from the user""" return (message.get('message') and message['message'].get('text') and not message['message'].get("is_echo")) def send_message(recipient_id, text): """Send a response to Facebook""" payload = { 'message': { 'text': text }, 'recipient': { 'id': recipient_id }, 'notification_type': 'regular' } auth = { 'access_token': PAGE_ACCESS_TOKEN } response = requests.post( FB_API_URL, params=auth, json=payload ) return response.json() ###Output _____no_output_____ ###Markdown Step 2. Understanding code to access Messenger messages and send them to your computer Now I will explain what above code does in detail. `flask` will create `app` that will listen for http requests of messages sent to `localhost:5000/webhook`, which is your localhost server. Inside `listen()` function, `request` function will use `verify_webhook(req)` function to authenticate connection between your `app` and Facebook. Once `listen()` function sees that message from Messenger is valid, it will send it to `respond(sender, message)` function, which access bot's code directly through `get_bot_response(message)` function. Finally, `send_message(recipient_id, text)` will send bot's response back to Messenger. Note that `VERIFY_TOKEN` and `PAGE_ACCESS_TOKEN` have been left out. `VERIFY_TOKEN` is any string you would be want to use as password to let Facebook know your localhost server wants to listen for messages. We will be coming back for `PAGE_ACCESS_TOKEN` later. Step 3. Running ngrok to start localhost serverOnce you installed ngrok program in your computer, start the program and don't turn it off while setting up connection. In terminal window, type `ngrok http 5000`. This will set up a http endpoint that will be forwarded to your computer on port 5000. Your http endpoint will look something like, `http://9cbec3d0.ngrok.io`. ![Imgur](https://i.imgur.com/Fs4QDvi.png) Step 4. Setting up Facebook page and getting Page Access TokenLet's head over to Facebook Developer website by clicking this [link](https://developers.facebook.com/). If you already have a Facebook account, just simply log in. Now, click on `My Apps` on the top right corner and click `Create an App`.![Imgur](https://i.imgur.com/GPlijTg.png)Create the display name of your App and click `Create App ID`. You have to do a quick security check to prove that you are not a robot. The display name can be anything you choose to be. I choose to use Bookbot for the demonstration purpose. ![Imgur](https://i.imgur.com/1Sc0QNx.png)You will be lead to your bot's dashboard. ![Imgur](https://i.imgur.com/X5ixdCv.png)First, click `Skip` for the first page you see like in the image above.Then, if you scroll down, you will see `Add a Product` section. Click on `Set Up` button for Messenger. In the next page you see, scroll down until you see `Access Tokens` section. You don't have Page Access Token yet for your chatbot because you don't have a Facebook page yet. What you want to do is create a new page. ![Imgur](https://i.imgur.com/6QJecWp.png)When you click on `Create a New Page`, you will be led to the following page:![Imgur](https://i.imgur.com/tSz1Hid.png)Click on either options to create a page. I have decided to create my BookBot page as a community page but it is up to you. ![Imgur](https://i.imgur.com/FAXRtUx.png)Once you click through a few options for profile picture, you will finally be able to see your bot's Facebook page. ![Imgur](https://i.imgur.com/E9sVpEs.png)Now, time to get back to Facebook Developer page for your bot. When you refresh the page, you can now select your bot's page to get Access Token. But you are not quite there yet. ![Imgur](https://i.imgur.com/S9pb3Kp.png)You will be told that you have to edit your permission to get access token. Just click on `Edit Permissions` and follow along the instruction until your bot is linked to the Facebook page. Now, you will see Page Access Token appear. ![Imgur](https://i.imgur.com/LqoOa5W.png)Click on it to copy to a clipboard and head over to `server.py`. Step 5. Starting chatbot serverFirst, copy and paste Page Access Token to `PAGE_ACCESS_TOKEN` in `server.py`. Save the file.Second, in a separate terminal window, `cd` to the directory where you have your chatbot code and `server.py` code. Third, in the terminal window, enter `set FLASK_APP=server.py`. Then enter `flask run`. If everything goes well, you will see something similar to following screen:![Imgur](https://i.imgur.com/LSp8HHT.png) Step 6. Setting up webhookGo to ngrok server and copy the address starting with `https`. Now go over to Facebook Development page for your bot. Underneath `Access Tokens` section, you will see `Webhooks` section. Click `Subscribe to Event`. Then copy and paste the address you copied in `Callback URL` section. Make sure to end the address with `/webhook`, because this is how your flask app will find webhook address. Ignore the callback URL in the following image. ![Imgur](https://i.imgur.com/Qu8BMF5.png)Next, fill out `Verify Token` with `VERIFY_TOKEN` from `server.py`. Finally, click `messages` and `messages_postbacks` from Subscription Fields and click `Verify and Save` button. Lastly, you want to choose the page you want to subscribe to and click `Subscribe`. ![Imgur](https://i.imgur.com/hCArdlN.jpg)Now you are all set! How do you know?![Imgur](https://i.imgur.com/EHxCGbS.jpg)On the navigation pane to the left, if you see the word `Webhook` with a green circle and a check mark, that means all the set up has been successfully completed. Step 7. Test your chatbot!Now, if you didn't get any error, everything has been set up! Once you are over to your bot's Facebook page, click on three dot icons and click `View as Page Visitor`.![Imgur](https://i.imgur.com/oFDOtGV.jpg)Now the moment of the truth! Click on `Send Message` and when the chatbox pops up start tying in question you have trained your bot for!![Imgur](https://i.imgur.com/LEgUYae.jpg)Open the chatbox in Messenger mode and here is what you will see. *Note: BookBot you see in the examples above has been changed to Testbot because of working issue.* ![Imgur](https://i.imgur.com/tDlfcRz.png)You can see the dummy chatbot code from `server.py` showing up on Messenger chatbox. If you have gotten this far, congratulations!Last but not least, you probably want your chatbot to do more than simply repeat what the user says. You can use the following simple chatbot code to simulate conversation. Copy and save the following as `chatbot.py` and make sure it is in the same directory as `server.py`. ###Code import re import random rules = {"I want (.*)":["What would it mean if you got {0}", "Why do you want {0}", "What's stopping you from getting {0}"], "if (.*)":["Do you really think it's likely that {0}", "Do you wish that {0}", "What do you think about {0}", "Really--if {0}"], "do you think (.*)":["{0} Absolutely.", "No chance"], "do you remember (.*)":["Did you think I would forget {0}", "Why haven't you been able to forget {0}", "What about {0}", "Yes .. and?"] } # Define respond() def robo_respond(message): # Call match_rule response, phrase = match_rule(rules, message) if '{0}' in response: # Replace the pronouns in the phrase phrase = replace_pronouns(phrase) # Include the phrase in the response response = response.format(phrase) return response # Define match_rule() def match_rule(rules, message): response, phrase = "default", None # Iterate over the rules dictionary for pattern, responses in rules.items(): # Create a match object match = re.search(pattern, message) if match is not None: # Choose a random response response = random.choice(responses) if '{0}' in response: phrase = match.group(1) # Return the response and phrase return response, phrase def replace_pronouns(message): message = message.lower() if 'me' in message: # Replace 'me' with 'you' return re.sub('me', 'you', message) if 'my' in message: # Replace 'my' with 'your' return re.sub('my', 'your', message) if 'your' in message: # Replace 'your' with 'my' return re.sub('your', 'my', message) if 'you' in message: # Replace 'you' with 'I' return re.sub('you', 'I', message) return message ###Output _____no_output_____
ipython notebooks/feature_sets.ipynb
###Markdown Copyright 2017 Google LLC. ###Code # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. ###Output _____no_output_____ ###Markdown Feature Sets **Learning Objective:** Create a minimal set of features that performs just as well as a more complex feature set So far, we've thrown all of our features into the model. Models with fewer features use fewer resources and are easier to maintain. Let's see if we can build a model on a minimal set of housing features that will perform equally as well as one that uses all the features in the data set. SetupAs before, let's load and prepare the California housing data. ###Code from __future__ import print_function import math from IPython import display from matplotlib import cm from matplotlib import gridspec from matplotlib import pyplot as plt import numpy as np import pandas as pd from sklearn import metrics import tensorflow as tf from tensorflow.python.data import Dataset tf.logging.set_verbosity(tf.logging.ERROR) pd.options.display.max_rows = 10 pd.options.display.float_format = '{:.1f}'.format california_housing_dataframe = pd.read_csv("https://storage.googleapis.com/mledu-datasets/california_housing_train.csv", sep=",") california_housing_dataframe = california_housing_dataframe.reindex( np.random.permutation(california_housing_dataframe.index)) def preprocess_features(california_housing_dataframe): """Prepares input features from California housing data set. Args: california_housing_dataframe: A Pandas DataFrame expected to contain data from the California housing data set. Returns: A DataFrame that contains the features to be used for the model, including synthetic features. """ selected_features = california_housing_dataframe[ ["latitude", "longitude", "housing_median_age", "total_rooms", "total_bedrooms", "population", "households", "median_income"]] processed_features = selected_features.copy() # Create a synthetic feature. processed_features["rooms_per_person"] = ( california_housing_dataframe["total_rooms"] / california_housing_dataframe["population"]) return processed_features def preprocess_targets(california_housing_dataframe): """Prepares target features (i.e., labels) from California housing data set. Args: california_housing_dataframe: A Pandas DataFrame expected to contain data from the California housing data set. Returns: A DataFrame that contains the target feature. """ output_targets = pd.DataFrame() # Scale the target to be in units of thousands of dollars. output_targets["median_house_value"] = ( california_housing_dataframe["median_house_value"] / 1000.0) return output_targets # Choose the first 12000 (out of 17000) examples for training. training_examples = preprocess_features(california_housing_dataframe.head(12000)) training_targets = preprocess_targets(california_housing_dataframe.head(12000)) # Choose the last 5000 (out of 17000) examples for validation. validation_examples = preprocess_features(california_housing_dataframe.tail(5000)) validation_targets = preprocess_targets(california_housing_dataframe.tail(5000)) # Double-check that we've done the right thing. print("Training examples summary:") display.display(training_examples.describe()) print("Validation examples summary:") display.display(validation_examples.describe()) print("Training targets summary:") display.display(training_targets.describe()) print("Validation targets summary:") display.display(validation_targets.describe()) ###Output Training examples summary: ###Markdown Task 1: Develop a Good Feature Set**What's the best performance you can get with just 2 or 3 features?**A **correlation matrix** shows pairwise correlations, both for each feature compared to the target and for each feature compared to other features.Here, correlation is defined as the [Pearson correlation coefficient](https://en.wikipedia.org/wiki/Pearson_product-moment_correlation_coefficient). You don't have to understand the mathematical details for this exercise.Correlation values have the following meanings: * `-1.0`: perfect negative correlation * `0.0`: no correlation * `1.0`: perfect positive correlation ###Code correlation_dataframe = training_examples.copy() correlation_dataframe["target"] = training_targets["median_house_value"] correlation_dataframe.corr() ###Output _____no_output_____ ###Markdown Features that have strong positive or negative correlations with the target will add information to our model. We can use the correlation matrix to find such strongly correlated features.We'd also like to have features that aren't so strongly correlated with each other, so that they add independent information.Use this information to try removing features. You can also try developing additional synthetic features, such as ratios of two raw features.For convenience, we've included the training code from the previous exercise. ###Code def construct_feature_columns(input_features): """Construct the TensorFlow Feature Columns. Args: input_features: The names of the numerical input features to use. Returns: A set of feature columns """ return set([tf.feature_column.numeric_column(my_feature) for my_feature in input_features]) def my_input_fn(features, targets, batch_size=1, shuffle=True, num_epochs=None): """Trains a linear regression model. Args: features: pandas DataFrame of features targets: pandas DataFrame of targets batch_size: Size of batches to be passed to the model shuffle: True or False. Whether to shuffle the data. num_epochs: Number of epochs for which data should be repeated. None = repeat indefinitely Returns: Tuple of (features, labels) for next data batch """ # Convert pandas data into a dict of np arrays. features = {key:np.array(value) for key,value in dict(features).items()} # Construct a dataset, and configure batching/repeating. ds = Dataset.from_tensor_slices((features,targets)) # warning: 2GB limit ds = ds.batch(batch_size).repeat(num_epochs) # Shuffle the data, if specified. if shuffle: ds = ds.shuffle(10000) # Return the next batch of data. features, labels = ds.make_one_shot_iterator().get_next() return features, labels def train_model( learning_rate, steps, batch_size, training_examples, training_targets, validation_examples, validation_targets): """Trains a linear regression model. In addition to training, this function also prints training progress information, as well as a plot of the training and validation loss over time. Args: learning_rate: A `float`, the learning rate. steps: A non-zero `int`, the total number of training steps. A training step consists of a forward and backward pass using a single batch. batch_size: A non-zero `int`, the batch size. training_examples: A `DataFrame` containing one or more columns from `california_housing_dataframe` to use as input features for training. training_targets: A `DataFrame` containing exactly one column from `california_housing_dataframe` to use as target for training. validation_examples: A `DataFrame` containing one or more columns from `california_housing_dataframe` to use as input features for validation. validation_targets: A `DataFrame` containing exactly one column from `california_housing_dataframe` to use as target for validation. Returns: A `LinearRegressor` object trained on the training data. """ periods = 10 steps_per_period = steps / periods # Create a linear regressor object. my_optimizer = tf.train.GradientDescentOptimizer(learning_rate=learning_rate) my_optimizer = tf.contrib.estimator.clip_gradients_by_norm(my_optimizer, 5.0) linear_regressor = tf.estimator.LinearRegressor( feature_columns=construct_feature_columns(training_examples), optimizer=my_optimizer ) # Create input functions. training_input_fn = lambda: my_input_fn(training_examples, training_targets["median_house_value"], batch_size=batch_size) predict_training_input_fn = lambda: my_input_fn(training_examples, training_targets["median_house_value"], num_epochs=1, shuffle=False) predict_validation_input_fn = lambda: my_input_fn(validation_examples, validation_targets["median_house_value"], num_epochs=1, shuffle=False) # Train the model, but do so inside a loop so that we can periodically assess # loss metrics. print("Training model...") print("RMSE (on training data):") training_rmse = [] validation_rmse = [] for period in range (0, periods): # Train the model, starting from the prior state. linear_regressor.train( input_fn=training_input_fn, steps=steps_per_period, ) # Take a break and compute predictions. training_predictions = linear_regressor.predict(input_fn=predict_training_input_fn) training_predictions = np.array([item['predictions'][0] for item in training_predictions]) validation_predictions = linear_regressor.predict(input_fn=predict_validation_input_fn) validation_predictions = np.array([item['predictions'][0] for item in validation_predictions]) # Compute training and validation loss. training_root_mean_squared_error = math.sqrt( metrics.mean_squared_error(training_predictions, training_targets)) validation_root_mean_squared_error = math.sqrt( metrics.mean_squared_error(validation_predictions, validation_targets)) # Occasionally print the current loss. print(" period %02d : %0.2f" % (period, training_root_mean_squared_error)) # Add the loss metrics from this period to our list. training_rmse.append(training_root_mean_squared_error) validation_rmse.append(validation_root_mean_squared_error) print("Model training finished.") # Output a graph of loss metrics over periods. plt.ylabel("RMSE") plt.xlabel("Periods") plt.title("Root Mean Squared Error vs. Periods") plt.tight_layout() plt.plot(training_rmse, label="training") plt.plot(validation_rmse, label="validation") plt.legend() return linear_regressor ###Output _____no_output_____ ###Markdown Spend 5 minutes searching for a good set of features and training parameters. Then check the solution to see what we chose. Don't forget that different features may require different learning parameters. ###Code # # Your code here: add your features of choice as a list of quoted strings. # minimal_features = [ "median_income", "rooms_per_person" ] assert minimal_features, "You must select at least one feature!" minimal_training_examples = training_examples[minimal_features] minimal_validation_examples = validation_examples[minimal_features] # # Don't forget to adjust these parameters. # train_model( learning_rate=0.1, steps=500, batch_size=5, training_examples=minimal_training_examples, training_targets=training_targets, validation_examples=minimal_validation_examples, validation_targets=validation_targets) ###Output Training model... RMSE (on training data): period 00 : 132.25 period 01 : 86.39 period 02 : 85.04 period 03 : 85.00 period 04 : 85.00 period 05 : 85.05 period 06 : 84.82 period 07 : 84.74 period 08 : 84.92 period 09 : 84.72 Model training finished. ###Markdown SolutionClick below for a solution. ###Code minimal_features = [ "median_income", "latitude", ] minimal_training_examples = training_examples[minimal_features] minimal_validation_examples = validation_examples[minimal_features] _ = train_model( learning_rate=0.01, steps=500, batch_size=5, training_examples=minimal_training_examples, training_targets=training_targets, validation_examples=minimal_validation_examples, validation_targets=validation_targets) ###Output _____no_output_____ ###Markdown Task 2: Make Better Use of LatitudePlotting `latitude` vs. `median_house_value` shows that there really isn't a linear relationship there.Instead, there are a couple of peaks, which roughly correspond to Los Angeles and San Francisco. ###Code plt.scatter(training_examples["latitude"], training_targets["median_house_value"]) ###Output _____no_output_____ ###Markdown **Try creating some synthetic features that do a better job with latitude.**For example, you could have a feature that maps `latitude` to a value of `|latitude - 38|`, and call this `distance_from_san_francisco`.Or you could break the space into 10 different buckets. `latitude_32_to_33`, `latitude_33_to_34`, etc., each showing a value of `1.0` if `latitude` is within that bucket range and a value of `0.0` otherwise.Use the correlation matrix to help guide development, and then add them to your model if you find something that looks good.What's the best validation performance you can get? ###Code # # YOUR CODE HERE: Train on a new data set that includes synthetic features based on latitude. # selected_examples = pd.DataFrame() selected_examples["distance"] = abs(training_examples["latitude"] - 38) selected_examples ###Output _____no_output_____ ###Markdown SolutionClick below for a solution. Aside from `latitude`, we'll also keep `median_income`, to compare with the previous results.We decided to bucketize the latitude. This is fairly straightforward in Pandas using `Series.apply`. ###Code LATITUDE_RANGES = zip(range(32, 44), range(33, 45)) def select_and_transform_features(source_df): selected_examples = pd.DataFrame() selected_examples["median_income"] = source_df["median_income"] for r in LATITUDE_RANGES: selected_examples["latitude_%d_to_%d" % r] = source_df["latitude"].apply( lambda l: 1.0 if l >= r[0] and l < r[1] else 0.0) return selected_examples selected_training_examples = select_and_transform_features(training_examples) selected_validation_examples = select_and_transform_features(validation_examples) _ = train_model( learning_rate=0.5, steps=500, batch_size=5, training_examples=selected_training_examples, training_targets=training_targets, validation_examples=selected_validation_examples, validation_targets=validation_targets) ###Output Training model... RMSE (on training data): period 00 : 84.00 period 01 : 85.72 period 02 : 86.31 period 03 : 81.77 period 04 : 80.64 period 05 : 80.75 period 06 : 80.63 period 07 : 79.95 period 08 : 80.15 period 09 : 80.40 Model training finished.